Jan 21 11:10:56 crc systemd[1]: Starting Kubernetes Kubelet... Jan 21 11:10:56 crc restorecon[4561]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Jan 21 11:10:56 crc restorecon[4561]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Jan 21 11:10:56 crc restorecon[4561]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Jan 21 11:10:56 crc restorecon[4561]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Jan 21 11:10:56 crc restorecon[4561]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Jan 21 11:10:56 crc restorecon[4561]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Jan 21 11:10:56 crc restorecon[4561]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Jan 21 11:10:56 crc restorecon[4561]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Jan 21 11:10:56 crc restorecon[4561]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Jan 21 11:10:56 crc restorecon[4561]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Jan 21 11:10:56 crc restorecon[4561]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Jan 21 11:10:56 crc restorecon[4561]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Jan 21 11:10:56 crc restorecon[4561]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Jan 21 11:10:56 crc restorecon[4561]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Jan 21 11:10:56 crc restorecon[4561]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Jan 21 11:10:56 crc restorecon[4561]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Jan 21 11:10:56 crc restorecon[4561]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Jan 21 11:10:56 crc restorecon[4561]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Jan 21 11:10:56 crc restorecon[4561]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 21 11:10:56 crc restorecon[4561]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 21 11:10:56 crc restorecon[4561]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 21 11:10:56 crc restorecon[4561]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 21 11:10:56 crc restorecon[4561]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 21 11:10:56 crc restorecon[4561]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 21 11:10:56 crc restorecon[4561]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 21 11:10:56 crc restorecon[4561]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 21 11:10:56 crc restorecon[4561]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 21 11:10:56 crc restorecon[4561]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 21 11:10:56 crc restorecon[4561]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 21 11:10:56 crc restorecon[4561]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 21 11:10:56 crc restorecon[4561]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 21 11:10:56 crc restorecon[4561]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 21 11:10:56 crc restorecon[4561]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Jan 21 11:10:56 crc restorecon[4561]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Jan 21 11:10:56 crc restorecon[4561]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Jan 21 11:10:56 crc restorecon[4561]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Jan 21 11:10:56 crc restorecon[4561]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Jan 21 11:10:56 crc restorecon[4561]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Jan 21 11:10:56 crc restorecon[4561]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 21 11:10:56 crc restorecon[4561]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 21 11:10:56 crc restorecon[4561]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 21 11:10:56 crc restorecon[4561]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 21 11:10:56 crc restorecon[4561]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 21 11:10:56 crc restorecon[4561]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 21 11:10:56 crc restorecon[4561]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 21 11:10:56 crc restorecon[4561]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 21 11:10:56 crc restorecon[4561]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 21 11:10:56 crc restorecon[4561]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 21 11:10:56 crc restorecon[4561]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 21 11:10:56 crc restorecon[4561]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Jan 21 11:10:56 crc restorecon[4561]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Jan 21 11:10:56 crc restorecon[4561]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 21 11:10:56 crc restorecon[4561]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Jan 21 11:10:56 crc restorecon[4561]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Jan 21 11:10:56 crc restorecon[4561]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 21 11:10:56 crc restorecon[4561]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Jan 21 11:10:56 crc restorecon[4561]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Jan 21 11:10:56 crc restorecon[4561]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 21 11:10:56 crc restorecon[4561]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Jan 21 11:10:56 crc restorecon[4561]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Jan 21 11:10:56 crc restorecon[4561]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 21 11:10:56 crc restorecon[4561]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Jan 21 11:10:56 crc restorecon[4561]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Jan 21 11:10:56 crc restorecon[4561]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 21 11:10:56 crc restorecon[4561]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Jan 21 11:10:56 crc restorecon[4561]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Jan 21 11:10:56 crc restorecon[4561]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Jan 21 11:10:57 crc restorecon[4561]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Jan 21 11:10:57 crc restorecon[4561]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Jan 21 11:10:57 crc kubenswrapper[4824]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Jan 21 11:10:57 crc kubenswrapper[4824]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Jan 21 11:10:57 crc kubenswrapper[4824]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Jan 21 11:10:57 crc kubenswrapper[4824]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Jan 21 11:10:57 crc kubenswrapper[4824]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Jan 21 11:10:57 crc kubenswrapper[4824]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.891767 4824 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.895214 4824 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.895230 4824 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.895235 4824 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.895240 4824 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.895244 4824 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.895248 4824 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.895252 4824 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.895255 4824 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.895259 4824 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.895270 4824 feature_gate.go:330] unrecognized feature gate: SignatureStores Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.895273 4824 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.895276 4824 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.895280 4824 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.895283 4824 feature_gate.go:330] unrecognized feature gate: Example Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.895286 4824 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.895289 4824 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.895292 4824 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.895296 4824 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.895299 4824 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.895303 4824 feature_gate.go:330] unrecognized feature gate: PlatformOperators Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.895306 4824 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.895310 4824 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.895313 4824 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.895317 4824 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.895320 4824 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.895324 4824 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.895328 4824 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.895332 4824 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.895336 4824 feature_gate.go:330] unrecognized feature gate: InsightsConfig Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.895339 4824 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.895343 4824 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.895346 4824 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.895349 4824 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.895352 4824 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.895357 4824 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.895362 4824 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.895367 4824 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.895371 4824 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.895376 4824 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.895380 4824 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.895384 4824 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.895389 4824 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.895393 4824 feature_gate.go:330] unrecognized feature gate: NewOLM Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.895398 4824 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.895402 4824 feature_gate.go:330] unrecognized feature gate: GatewayAPI Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.895406 4824 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.895410 4824 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.895414 4824 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.895427 4824 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.895431 4824 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.895435 4824 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.895438 4824 feature_gate.go:330] unrecognized feature gate: OVNObservability Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.895441 4824 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.895445 4824 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.895448 4824 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.895452 4824 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.895455 4824 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.895458 4824 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.895462 4824 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.895466 4824 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.895469 4824 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.895472 4824 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.895475 4824 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.895478 4824 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.895481 4824 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.895484 4824 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.895488 4824 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.895491 4824 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.895494 4824 feature_gate.go:330] unrecognized feature gate: PinnedImages Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.895498 4824 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.895504 4824 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.895869 4824 flags.go:64] FLAG: --address="0.0.0.0" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.895880 4824 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.895886 4824 flags.go:64] FLAG: --anonymous-auth="true" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.895891 4824 flags.go:64] FLAG: --application-metrics-count-limit="100" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.895895 4824 flags.go:64] FLAG: --authentication-token-webhook="false" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.895899 4824 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.895904 4824 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.895908 4824 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.895912 4824 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.895915 4824 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.895920 4824 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.895923 4824 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.895927 4824 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.895932 4824 flags.go:64] FLAG: --cgroup-root="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.895936 4824 flags.go:64] FLAG: --cgroups-per-qos="true" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.895940 4824 flags.go:64] FLAG: --client-ca-file="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.895943 4824 flags.go:64] FLAG: --cloud-config="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.895947 4824 flags.go:64] FLAG: --cloud-provider="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.895951 4824 flags.go:64] FLAG: --cluster-dns="[]" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.895971 4824 flags.go:64] FLAG: --cluster-domain="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.895975 4824 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.895979 4824 flags.go:64] FLAG: --config-dir="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.895983 4824 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.895987 4824 flags.go:64] FLAG: --container-log-max-files="5" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.895991 4824 flags.go:64] FLAG: --container-log-max-size="10Mi" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.895995 4824 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.895998 4824 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896002 4824 flags.go:64] FLAG: --containerd-namespace="k8s.io" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896006 4824 flags.go:64] FLAG: --contention-profiling="false" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896010 4824 flags.go:64] FLAG: --cpu-cfs-quota="true" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896013 4824 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896017 4824 flags.go:64] FLAG: --cpu-manager-policy="none" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896021 4824 flags.go:64] FLAG: --cpu-manager-policy-options="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896025 4824 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896029 4824 flags.go:64] FLAG: --enable-controller-attach-detach="true" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896033 4824 flags.go:64] FLAG: --enable-debugging-handlers="true" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896036 4824 flags.go:64] FLAG: --enable-load-reader="false" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896040 4824 flags.go:64] FLAG: --enable-server="true" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896043 4824 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896048 4824 flags.go:64] FLAG: --event-burst="100" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896052 4824 flags.go:64] FLAG: --event-qps="50" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896055 4824 flags.go:64] FLAG: --event-storage-age-limit="default=0" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896059 4824 flags.go:64] FLAG: --event-storage-event-limit="default=0" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896065 4824 flags.go:64] FLAG: --eviction-hard="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896071 4824 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896074 4824 flags.go:64] FLAG: --eviction-minimum-reclaim="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896078 4824 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896082 4824 flags.go:64] FLAG: --eviction-soft="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896085 4824 flags.go:64] FLAG: --eviction-soft-grace-period="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896089 4824 flags.go:64] FLAG: --exit-on-lock-contention="false" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896093 4824 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896096 4824 flags.go:64] FLAG: --experimental-mounter-path="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896100 4824 flags.go:64] FLAG: --fail-cgroupv1="false" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896104 4824 flags.go:64] FLAG: --fail-swap-on="true" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896107 4824 flags.go:64] FLAG: --feature-gates="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896112 4824 flags.go:64] FLAG: --file-check-frequency="20s" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896116 4824 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896121 4824 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896125 4824 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896129 4824 flags.go:64] FLAG: --healthz-port="10248" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896133 4824 flags.go:64] FLAG: --help="false" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896137 4824 flags.go:64] FLAG: --hostname-override="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896140 4824 flags.go:64] FLAG: --housekeeping-interval="10s" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896144 4824 flags.go:64] FLAG: --http-check-frequency="20s" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896147 4824 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896151 4824 flags.go:64] FLAG: --image-credential-provider-config="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896154 4824 flags.go:64] FLAG: --image-gc-high-threshold="85" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896158 4824 flags.go:64] FLAG: --image-gc-low-threshold="80" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896162 4824 flags.go:64] FLAG: --image-service-endpoint="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896165 4824 flags.go:64] FLAG: --kernel-memcg-notification="false" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896169 4824 flags.go:64] FLAG: --kube-api-burst="100" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896173 4824 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896176 4824 flags.go:64] FLAG: --kube-api-qps="50" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896180 4824 flags.go:64] FLAG: --kube-reserved="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896183 4824 flags.go:64] FLAG: --kube-reserved-cgroup="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896187 4824 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896191 4824 flags.go:64] FLAG: --kubelet-cgroups="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896194 4824 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896198 4824 flags.go:64] FLAG: --lock-file="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896201 4824 flags.go:64] FLAG: --log-cadvisor-usage="false" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896205 4824 flags.go:64] FLAG: --log-flush-frequency="5s" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896209 4824 flags.go:64] FLAG: --log-json-info-buffer-size="0" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896214 4824 flags.go:64] FLAG: --log-json-split-stream="false" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896218 4824 flags.go:64] FLAG: --log-text-info-buffer-size="0" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896222 4824 flags.go:64] FLAG: --log-text-split-stream="false" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896226 4824 flags.go:64] FLAG: --logging-format="text" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896230 4824 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896234 4824 flags.go:64] FLAG: --make-iptables-util-chains="true" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896237 4824 flags.go:64] FLAG: --manifest-url="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896241 4824 flags.go:64] FLAG: --manifest-url-header="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896245 4824 flags.go:64] FLAG: --max-housekeeping-interval="15s" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896249 4824 flags.go:64] FLAG: --max-open-files="1000000" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896253 4824 flags.go:64] FLAG: --max-pods="110" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896257 4824 flags.go:64] FLAG: --maximum-dead-containers="-1" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896260 4824 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896264 4824 flags.go:64] FLAG: --memory-manager-policy="None" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896267 4824 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896271 4824 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896275 4824 flags.go:64] FLAG: --node-ip="192.168.126.11" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896278 4824 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896286 4824 flags.go:64] FLAG: --node-status-max-images="50" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896290 4824 flags.go:64] FLAG: --node-status-update-frequency="10s" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896293 4824 flags.go:64] FLAG: --oom-score-adj="-999" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896297 4824 flags.go:64] FLAG: --pod-cidr="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896301 4824 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896307 4824 flags.go:64] FLAG: --pod-manifest-path="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896310 4824 flags.go:64] FLAG: --pod-max-pids="-1" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896314 4824 flags.go:64] FLAG: --pods-per-core="0" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896318 4824 flags.go:64] FLAG: --port="10250" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896321 4824 flags.go:64] FLAG: --protect-kernel-defaults="false" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896325 4824 flags.go:64] FLAG: --provider-id="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896328 4824 flags.go:64] FLAG: --qos-reserved="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896332 4824 flags.go:64] FLAG: --read-only-port="10255" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896336 4824 flags.go:64] FLAG: --register-node="true" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896340 4824 flags.go:64] FLAG: --register-schedulable="true" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896344 4824 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896350 4824 flags.go:64] FLAG: --registry-burst="10" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896354 4824 flags.go:64] FLAG: --registry-qps="5" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896358 4824 flags.go:64] FLAG: --reserved-cpus="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896361 4824 flags.go:64] FLAG: --reserved-memory="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896365 4824 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896369 4824 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896373 4824 flags.go:64] FLAG: --rotate-certificates="false" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896376 4824 flags.go:64] FLAG: --rotate-server-certificates="false" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896380 4824 flags.go:64] FLAG: --runonce="false" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896383 4824 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896387 4824 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896391 4824 flags.go:64] FLAG: --seccomp-default="false" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896394 4824 flags.go:64] FLAG: --serialize-image-pulls="true" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896398 4824 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896402 4824 flags.go:64] FLAG: --storage-driver-db="cadvisor" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896405 4824 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896409 4824 flags.go:64] FLAG: --storage-driver-password="root" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896412 4824 flags.go:64] FLAG: --storage-driver-secure="false" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896425 4824 flags.go:64] FLAG: --storage-driver-table="stats" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896430 4824 flags.go:64] FLAG: --storage-driver-user="root" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896433 4824 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896437 4824 flags.go:64] FLAG: --sync-frequency="1m0s" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896442 4824 flags.go:64] FLAG: --system-cgroups="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896446 4824 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896452 4824 flags.go:64] FLAG: --system-reserved-cgroup="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896455 4824 flags.go:64] FLAG: --tls-cert-file="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896459 4824 flags.go:64] FLAG: --tls-cipher-suites="[]" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896463 4824 flags.go:64] FLAG: --tls-min-version="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896467 4824 flags.go:64] FLAG: --tls-private-key-file="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896470 4824 flags.go:64] FLAG: --topology-manager-policy="none" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896474 4824 flags.go:64] FLAG: --topology-manager-policy-options="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896479 4824 flags.go:64] FLAG: --topology-manager-scope="container" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896484 4824 flags.go:64] FLAG: --v="2" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896489 4824 flags.go:64] FLAG: --version="false" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896494 4824 flags.go:64] FLAG: --vmodule="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896500 4824 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896504 4824 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.896593 4824 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.896597 4824 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.896601 4824 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.896605 4824 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.896609 4824 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.896613 4824 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.896617 4824 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.896621 4824 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.896624 4824 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.896627 4824 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.896631 4824 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.896635 4824 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.896638 4824 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.896642 4824 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.896645 4824 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.896648 4824 feature_gate.go:330] unrecognized feature gate: InsightsConfig Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.896651 4824 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.896654 4824 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.896658 4824 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.896661 4824 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.896664 4824 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.896668 4824 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.896671 4824 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.896675 4824 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.896678 4824 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.896681 4824 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.896684 4824 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.896688 4824 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.896691 4824 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.896694 4824 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.896698 4824 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.896701 4824 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.896704 4824 feature_gate.go:330] unrecognized feature gate: OVNObservability Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.896707 4824 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.896710 4824 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.896714 4824 feature_gate.go:330] unrecognized feature gate: PlatformOperators Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.896717 4824 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.896720 4824 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.896724 4824 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.896727 4824 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.896730 4824 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.896734 4824 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.896737 4824 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.896741 4824 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.896745 4824 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.896749 4824 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.896752 4824 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.896756 4824 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.896759 4824 feature_gate.go:330] unrecognized feature gate: NewOLM Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.896762 4824 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.896766 4824 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.896769 4824 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.896772 4824 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.896776 4824 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.896779 4824 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.896783 4824 feature_gate.go:330] unrecognized feature gate: SignatureStores Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.896786 4824 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.896790 4824 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.896794 4824 feature_gate.go:330] unrecognized feature gate: GatewayAPI Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.896797 4824 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.896801 4824 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.896804 4824 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.896807 4824 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.896810 4824 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.896813 4824 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.896816 4824 feature_gate.go:330] unrecognized feature gate: PinnedImages Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.896820 4824 feature_gate.go:330] unrecognized feature gate: Example Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.896823 4824 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.896826 4824 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.896829 4824 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.896832 4824 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.896838 4824 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.904297 4824 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.904318 4824 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904367 4824 feature_gate.go:330] unrecognized feature gate: OVNObservability Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904373 4824 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904377 4824 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904382 4824 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904388 4824 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904392 4824 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904396 4824 feature_gate.go:330] unrecognized feature gate: PinnedImages Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904400 4824 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904403 4824 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904407 4824 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904410 4824 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904414 4824 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904430 4824 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904434 4824 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904437 4824 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904441 4824 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904445 4824 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904448 4824 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904452 4824 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904457 4824 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904463 4824 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904467 4824 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904470 4824 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904473 4824 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904477 4824 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904481 4824 feature_gate.go:330] unrecognized feature gate: SignatureStores Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904485 4824 feature_gate.go:330] unrecognized feature gate: Example Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904489 4824 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904493 4824 feature_gate.go:330] unrecognized feature gate: InsightsConfig Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904505 4824 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904509 4824 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904512 4824 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904515 4824 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904519 4824 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904522 4824 feature_gate.go:330] unrecognized feature gate: PlatformOperators Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904525 4824 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904528 4824 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904532 4824 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904535 4824 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904539 4824 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904542 4824 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904545 4824 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904548 4824 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904551 4824 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904555 4824 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904558 4824 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904561 4824 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904564 4824 feature_gate.go:330] unrecognized feature gate: GatewayAPI Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904567 4824 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904570 4824 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904574 4824 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904578 4824 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904582 4824 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904586 4824 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904589 4824 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904593 4824 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904598 4824 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904601 4824 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904605 4824 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904608 4824 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904611 4824 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904615 4824 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904618 4824 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904622 4824 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904626 4824 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904629 4824 feature_gate.go:330] unrecognized feature gate: NewOLM Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904632 4824 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904636 4824 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904639 4824 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904643 4824 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904646 4824 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.904651 4824 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904745 4824 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904751 4824 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904755 4824 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904758 4824 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904762 4824 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904766 4824 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904769 4824 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904773 4824 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904776 4824 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904780 4824 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904783 4824 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904786 4824 feature_gate.go:330] unrecognized feature gate: OVNObservability Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904790 4824 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904793 4824 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904797 4824 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904800 4824 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904804 4824 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904808 4824 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904812 4824 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904817 4824 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904821 4824 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904824 4824 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904828 4824 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904832 4824 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904835 4824 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904839 4824 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904843 4824 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904846 4824 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904850 4824 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904853 4824 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904856 4824 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904860 4824 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904863 4824 feature_gate.go:330] unrecognized feature gate: Example Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904866 4824 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904869 4824 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904872 4824 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904875 4824 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904879 4824 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904883 4824 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904886 4824 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904889 4824 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904893 4824 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904896 4824 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904899 4824 feature_gate.go:330] unrecognized feature gate: PinnedImages Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904902 4824 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904905 4824 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904909 4824 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904913 4824 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904917 4824 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904921 4824 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904926 4824 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904930 4824 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904934 4824 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904937 4824 feature_gate.go:330] unrecognized feature gate: NewOLM Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904941 4824 feature_gate.go:330] unrecognized feature gate: InsightsConfig Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904944 4824 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904947 4824 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904951 4824 feature_gate.go:330] unrecognized feature gate: GatewayAPI Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904968 4824 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904971 4824 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904974 4824 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904978 4824 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904981 4824 feature_gate.go:330] unrecognized feature gate: SignatureStores Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904984 4824 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904987 4824 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904990 4824 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904993 4824 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.904997 4824 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.905000 4824 feature_gate.go:330] unrecognized feature gate: PlatformOperators Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.905004 4824 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.905007 4824 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.905011 4824 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.905109 4824 server.go:940] "Client rotation is on, will bootstrap in background" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.907688 4824 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.907750 4824 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.908759 4824 server.go:997] "Starting client certificate rotation" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.908783 4824 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.908902 4824 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2025-12-09 10:23:23.933533086 +0000 UTC Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.908949 4824 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.919532 4824 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.920955 4824 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Jan 21 11:10:57 crc kubenswrapper[4824]: E0121 11:10:57.921722 4824 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 192.168.26.120:6443: connect: connection refused" logger="UnhandledError" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.928329 4824 log.go:25] "Validated CRI v1 runtime API" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.944409 4824 log.go:25] "Validated CRI v1 image API" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.946069 4824 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.949559 4824 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2026-01-21-11-07-36-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.949588 4824 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:49 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/containers/storage/overlay-containers/75d81934760b26101869fbd8e4b5954c62b019c1cc3e5a0c9f82ed8de46b3b22/userdata/shm:{mountpoint:/var/lib/containers/storage/overlay-containers/75d81934760b26101869fbd8e4b5954c62b019c1cc3e5a0c9f82ed8de46b3b22/userdata/shm major:0 minor:42 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:50 fsType:tmpfs blockSize:0} overlay_0-43:{mountpoint:/var/lib/containers/storage/overlay/94b752e0a51c0134b00ddef6dc7a933a9d7c1d9bdc88a18dae4192a0d557d623/merged major:0 minor:43 fsType:overlay blockSize:0}] Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.963034 4824 manager.go:217] Machine: {Timestamp:2026-01-21 11:10:57.961268161 +0000 UTC m=+0.254297453 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2445404 MemoryCapacity:33654116352 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:d00624f5-786f-41dc-bdc1-a1875f167d65 BootID:695c676c-720c-4921-9e71-56b96f2fb46b Filesystems:[{Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827060224 Type:vfs Inodes:1048576 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:49 Capacity:3365408768 Type:vfs Inodes:821633 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:50 Capacity:1073741824 Type:vfs Inodes:4108168 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827056128 Type:vfs Inodes:4108168 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/var/lib/containers/storage/overlay-containers/75d81934760b26101869fbd8e4b5954c62b019c1cc3e5a0c9f82ed8de46b3b22/userdata/shm DeviceMajor:0 DeviceMinor:42 Capacity:65536000 Type:vfs Inodes:4108168 HasInodes:true} {Device:overlay_0-43 DeviceMajor:0 DeviceMinor:43 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:c5:e9:09 Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:enp3s0 MacAddress:fa:16:3e:c5:e9:09 Speed:-1 Mtu:1500} {Name:enp7s0 MacAddress:fa:16:3e:ec:64:04 Speed:-1 Mtu:1440} {Name:enp7s0.20 MacAddress:52:54:00:ec:ff:0a Speed:-1 Mtu:1436} {Name:enp7s0.21 MacAddress:52:54:00:35:1c:7c Speed:-1 Mtu:1436} {Name:enp7s0.22 MacAddress:52:54:00:53:57:a0 Speed:-1 Mtu:1436} {Name:eth10 MacAddress:8e:05:62:5e:18:a1 Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:aa:72:41:c4:00:59 Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654116352 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:65536 Type:Data Level:1} {Id:0 Size:65536 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:65536 Type:Data Level:1} {Id:1 Size:65536 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:65536 Type:Data Level:1} {Id:10 Size:65536 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:65536 Type:Data Level:1} {Id:11 Size:65536 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:65536 Type:Data Level:1} {Id:2 Size:65536 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:65536 Type:Data Level:1} {Id:3 Size:65536 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:65536 Type:Data Level:1} {Id:4 Size:65536 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:65536 Type:Data Level:1} {Id:5 Size:65536 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:65536 Type:Data Level:1} {Id:6 Size:65536 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:65536 Type:Data Level:1} {Id:7 Size:65536 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:65536 Type:Data Level:1} {Id:8 Size:65536 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:65536 Type:Data Level:1} {Id:9 Size:65536 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.963214 4824 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.963291 4824 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.963551 4824 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.963704 4824 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.963730 4824 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.963883 4824 topology_manager.go:138] "Creating topology manager with none policy" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.963891 4824 container_manager_linux.go:303] "Creating device plugin manager" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.964207 4824 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.964233 4824 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.964307 4824 state_mem.go:36] "Initialized new in-memory state store" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.964384 4824 server.go:1245] "Using root directory" path="/var/lib/kubelet" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.966217 4824 kubelet.go:418] "Attempting to sync node with API server" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.966235 4824 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.966254 4824 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.966264 4824 kubelet.go:324] "Adding apiserver pod source" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.966274 4824 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.968458 4824 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.969090 4824 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.969835 4824 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 192.168.26.120:6443: connect: connection refused Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.969844 4824 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 192.168.26.120:6443: connect: connection refused Jan 21 11:10:57 crc kubenswrapper[4824]: E0121 11:10:57.969907 4824 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 192.168.26.120:6443: connect: connection refused" logger="UnhandledError" Jan 21 11:10:57 crc kubenswrapper[4824]: E0121 11:10:57.969911 4824 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 192.168.26.120:6443: connect: connection refused" logger="UnhandledError" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.970805 4824 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.971742 4824 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.971764 4824 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.971771 4824 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.971778 4824 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.971789 4824 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.971794 4824 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.971799 4824 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.971812 4824 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.971819 4824 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.971825 4824 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.971833 4824 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.971839 4824 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.972429 4824 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.972753 4824 server.go:1280] "Started kubelet" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.973390 4824 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.973395 4824 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Jan 21 11:10:57 crc systemd[1]: Started Kubernetes Kubelet. Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.973808 4824 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 192.168.26.120:6443: connect: connection refused Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.976306 4824 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.978242 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.978288 4824 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Jan 21 11:10:57 crc kubenswrapper[4824]: E0121 11:10:57.978732 4824 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.978777 4824 volume_manager.go:287] "The desired_state_of_world populator starts" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.978879 4824 volume_manager.go:289] "Starting Kubelet Volume Manager" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.978777 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-17 23:23:38.542282267 +0000 UTC Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.978794 4824 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Jan 21 11:10:57 crc kubenswrapper[4824]: E0121 11:10:57.979765 4824 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 192.168.26.120:6443: connect: connection refused" interval="200ms" Jan 21 11:10:57 crc kubenswrapper[4824]: W0121 11:10:57.980001 4824 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 192.168.26.120:6443: connect: connection refused Jan 21 11:10:57 crc kubenswrapper[4824]: E0121 11:10:57.980151 4824 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 192.168.26.120:6443: connect: connection refused" logger="UnhandledError" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.980504 4824 server.go:460] "Adding debug handlers to kubelet server" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.980557 4824 factory.go:55] Registering systemd factory Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.980581 4824 factory.go:221] Registration of the systemd container factory successfully Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.980881 4824 factory.go:153] Registering CRI-O factory Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.980905 4824 factory.go:221] Registration of the crio container factory successfully Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.980989 4824 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.981053 4824 factory.go:103] Registering Raw factory Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.981086 4824 manager.go:1196] Started watching for new ooms in manager Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.981630 4824 manager.go:319] Starting recovery of all containers Jan 21 11:10:57 crc kubenswrapper[4824]: E0121 11:10:57.979717 4824 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 192.168.26.120:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.188cba8a9250030c default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-01-21 11:10:57.972732684 +0000 UTC m=+0.265761976,LastTimestamp:2026-01-21 11:10:57.972732684 +0000 UTC m=+0.265761976,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.985925 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.985999 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986010 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986024 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986033 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986041 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986049 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986057 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986067 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986076 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986084 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986097 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986105 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986117 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986125 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986132 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986140 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986149 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986157 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986164 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986173 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986247 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986257 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986267 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986275 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986283 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986306 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986318 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986329 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986337 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986346 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986362 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986371 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986380 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986388 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986397 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986405 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986425 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986433 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986441 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986448 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986457 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986466 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986475 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986484 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986492 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986501 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986510 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986519 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986527 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986535 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986543 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986556 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986565 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986574 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986583 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986592 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986599 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986607 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986616 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986626 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986646 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986654 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986661 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986671 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986679 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986688 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986698 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986705 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986714 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986722 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986730 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986737 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986745 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986753 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986761 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986771 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986780 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986787 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986796 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986803 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986811 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986820 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986827 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986836 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986847 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986855 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986863 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986870 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986879 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986886 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986894 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986902 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986910 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986917 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986925 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986932 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986940 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986947 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986981 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986990 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.986998 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.987006 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.987013 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.987027 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.987036 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.987046 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.987056 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.987070 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.987078 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.987086 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.987096 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.987104 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.987113 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.987121 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.987129 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.987137 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.987146 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.987159 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.987166 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.987175 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.987183 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.987190 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.987198 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.987207 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.987214 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.987222 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.987230 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.987239 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.987248 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.987257 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.987265 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.987272 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.987279 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.987289 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.987296 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.987305 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.987314 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.987323 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.987332 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.987343 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.987351 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.987359 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.987368 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.987376 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.987384 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.987397 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.987406 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.987426 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.987436 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.987446 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.989684 4824 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.989710 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.989722 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.989731 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.989739 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.989748 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.989757 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.989765 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.989773 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.989781 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.989789 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.989798 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.989806 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.989813 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.989822 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.989835 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.989851 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.989859 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.989867 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.989877 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.989886 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.989904 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.989916 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.989923 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.989934 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.989941 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.989971 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.989981 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.989989 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.989998 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.990006 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.990018 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.990027 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.990035 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.990043 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.990051 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.990061 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.990070 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.990087 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.990096 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.990104 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.990115 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.990125 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.990133 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.990142 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.990149 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.990163 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.990173 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.990184 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.990196 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.990204 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.990212 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.990224 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.990237 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.990245 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.990253 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.990261 4824 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.990273 4824 reconstruct.go:97] "Volume reconstruction finished" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.990280 4824 reconciler.go:26] "Reconciler: start to sync state" Jan 21 11:10:57 crc kubenswrapper[4824]: I0121 11:10:57.997481 4824 manager.go:324] Recovery completed Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.007670 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.009201 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.009284 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.009346 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.010734 4824 cpu_manager.go:225] "Starting CPU manager" policy="none" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.010797 4824 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.010852 4824 state_mem.go:36] "Initialized new in-memory state store" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.041482 4824 policy_none.go:49] "None policy: Start" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.042109 4824 memory_manager.go:170] "Starting memorymanager" policy="None" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.042135 4824 state_mem.go:35] "Initializing new in-memory state store" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.046489 4824 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.047706 4824 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.048054 4824 status_manager.go:217] "Starting to sync pod status with apiserver" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.048088 4824 kubelet.go:2335] "Starting kubelet main sync loop" Jan 21 11:10:58 crc kubenswrapper[4824]: E0121 11:10:58.048131 4824 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Jan 21 11:10:58 crc kubenswrapper[4824]: W0121 11:10:58.048539 4824 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 192.168.26.120:6443: connect: connection refused Jan 21 11:10:58 crc kubenswrapper[4824]: E0121 11:10:58.048576 4824 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 192.168.26.120:6443: connect: connection refused" logger="UnhandledError" Jan 21 11:10:58 crc kubenswrapper[4824]: E0121 11:10:58.079177 4824 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.080597 4824 manager.go:334] "Starting Device Plugin manager" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.080633 4824 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.080645 4824 server.go:79] "Starting device plugin registration server" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.080969 4824 eviction_manager.go:189] "Eviction manager: starting control loop" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.080990 4824 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.081147 4824 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.081216 4824 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.081227 4824 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Jan 21 11:10:58 crc kubenswrapper[4824]: E0121 11:10:58.085870 4824 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.148749 4824 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc"] Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.148826 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.149634 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.149671 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.149681 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.149811 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.150045 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.150092 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.150593 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.150681 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.150734 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.150749 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.150806 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.150817 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.150985 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.151070 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.151361 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.151836 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.151883 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.151893 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.152032 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.152127 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.152179 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.152898 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.152920 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.152929 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.153046 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.153134 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.153153 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.153161 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.153167 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.153191 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.153203 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.153203 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.153227 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.153763 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.153788 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.153796 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.153841 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.153858 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.153867 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.153889 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.153908 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.154495 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.154521 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.154530 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:10:58 crc kubenswrapper[4824]: E0121 11:10:58.180397 4824 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 192.168.26.120:6443: connect: connection refused" interval="400ms" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.181412 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.182076 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.182100 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.182110 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.182125 4824 kubelet_node_status.go:76] "Attempting to register node" node="crc" Jan 21 11:10:58 crc kubenswrapper[4824]: E0121 11:10:58.182352 4824 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 192.168.26.120:6443: connect: connection refused" node="crc" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.192602 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.192628 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.192646 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.192660 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.192672 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.192685 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.192699 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.192712 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.192801 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.192911 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.192946 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.192987 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.193016 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.193035 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.193050 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.293986 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.294057 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.294087 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.294102 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.294124 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.294140 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.294155 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.294174 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.294219 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.294233 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.294234 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.294241 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.294284 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.294294 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.294357 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.294362 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.294248 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.294386 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.294339 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.294382 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.294366 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.294445 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.294460 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.294482 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.294499 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.294530 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.294565 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.294603 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.294632 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.294684 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.382707 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.383986 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.384097 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.384186 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.384296 4824 kubelet_node_status.go:76] "Attempting to register node" node="crc" Jan 21 11:10:58 crc kubenswrapper[4824]: E0121 11:10:58.384734 4824 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 192.168.26.120:6443: connect: connection refused" node="crc" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.493041 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.499385 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.516108 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Jan 21 11:10:58 crc kubenswrapper[4824]: W0121 11:10:58.516734 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-edf769613ff88e9ab714047d66dcc4a41f0950c889ae2f5fe088499d67ab3ed3 WatchSource:0}: Error finding container edf769613ff88e9ab714047d66dcc4a41f0950c889ae2f5fe088499d67ab3ed3: Status 404 returned error can't find the container with id edf769613ff88e9ab714047d66dcc4a41f0950c889ae2f5fe088499d67ab3ed3 Jan 21 11:10:58 crc kubenswrapper[4824]: W0121 11:10:58.518298 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-a4ddd39b572f16102ce0dd8128f64021ff327707e7b38e42cd31f31d0ebdd22d WatchSource:0}: Error finding container a4ddd39b572f16102ce0dd8128f64021ff327707e7b38e42cd31f31d0ebdd22d: Status 404 returned error can't find the container with id a4ddd39b572f16102ce0dd8128f64021ff327707e7b38e42cd31f31d0ebdd22d Jan 21 11:10:58 crc kubenswrapper[4824]: W0121 11:10:58.525441 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-4ad861050920fe34134874ba9dc73c919058a01b36434d0a996bb3f687549498 WatchSource:0}: Error finding container 4ad861050920fe34134874ba9dc73c919058a01b36434d0a996bb3f687549498: Status 404 returned error can't find the container with id 4ad861050920fe34134874ba9dc73c919058a01b36434d0a996bb3f687549498 Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.530014 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.534872 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Jan 21 11:10:58 crc kubenswrapper[4824]: W0121 11:10:58.546524 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-74643405e4bd126e4c118bb553357012409644db3d7dc245d90ad17473d46cfa WatchSource:0}: Error finding container 74643405e4bd126e4c118bb553357012409644db3d7dc245d90ad17473d46cfa: Status 404 returned error can't find the container with id 74643405e4bd126e4c118bb553357012409644db3d7dc245d90ad17473d46cfa Jan 21 11:10:58 crc kubenswrapper[4824]: W0121 11:10:58.547144 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-b7b5a9d4ae92b18983fc96eac50835eb2262c50f8125af46811cfbd1c64137a4 WatchSource:0}: Error finding container b7b5a9d4ae92b18983fc96eac50835eb2262c50f8125af46811cfbd1c64137a4: Status 404 returned error can't find the container with id b7b5a9d4ae92b18983fc96eac50835eb2262c50f8125af46811cfbd1c64137a4 Jan 21 11:10:58 crc kubenswrapper[4824]: E0121 11:10:58.581443 4824 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 192.168.26.120:6443: connect: connection refused" interval="800ms" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.784857 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.785812 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.785851 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.785860 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.785881 4824 kubelet_node_status.go:76] "Attempting to register node" node="crc" Jan 21 11:10:58 crc kubenswrapper[4824]: E0121 11:10:58.786320 4824 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 192.168.26.120:6443: connect: connection refused" node="crc" Jan 21 11:10:58 crc kubenswrapper[4824]: W0121 11:10:58.883394 4824 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 192.168.26.120:6443: connect: connection refused Jan 21 11:10:58 crc kubenswrapper[4824]: E0121 11:10:58.883644 4824 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 192.168.26.120:6443: connect: connection refused" logger="UnhandledError" Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.974482 4824 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 192.168.26.120:6443: connect: connection refused Jan 21 11:10:58 crc kubenswrapper[4824]: I0121 11:10:58.979802 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-29 01:24:41.62253628 +0000 UTC Jan 21 11:10:59 crc kubenswrapper[4824]: I0121 11:10:59.051459 4824 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc" exitCode=0 Jan 21 11:10:59 crc kubenswrapper[4824]: I0121 11:10:59.051521 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc"} Jan 21 11:10:59 crc kubenswrapper[4824]: I0121 11:10:59.051595 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"edf769613ff88e9ab714047d66dcc4a41f0950c889ae2f5fe088499d67ab3ed3"} Jan 21 11:10:59 crc kubenswrapper[4824]: I0121 11:10:59.051691 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 21 11:10:59 crc kubenswrapper[4824]: I0121 11:10:59.052544 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:10:59 crc kubenswrapper[4824]: I0121 11:10:59.052570 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:10:59 crc kubenswrapper[4824]: I0121 11:10:59.052579 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:10:59 crc kubenswrapper[4824]: I0121 11:10:59.053618 4824 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="ef10a172a3279201375f7445673820033978ce1eb2bbef7a2c0ed297fd783875" exitCode=0 Jan 21 11:10:59 crc kubenswrapper[4824]: I0121 11:10:59.053718 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"ef10a172a3279201375f7445673820033978ce1eb2bbef7a2c0ed297fd783875"} Jan 21 11:10:59 crc kubenswrapper[4824]: I0121 11:10:59.053750 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"a4ddd39b572f16102ce0dd8128f64021ff327707e7b38e42cd31f31d0ebdd22d"} Jan 21 11:10:59 crc kubenswrapper[4824]: I0121 11:10:59.053838 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 21 11:10:59 crc kubenswrapper[4824]: I0121 11:10:59.054446 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 21 11:10:59 crc kubenswrapper[4824]: I0121 11:10:59.055391 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:10:59 crc kubenswrapper[4824]: I0121 11:10:59.055415 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:10:59 crc kubenswrapper[4824]: I0121 11:10:59.055427 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:10:59 crc kubenswrapper[4824]: I0121 11:10:59.055397 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:10:59 crc kubenswrapper[4824]: I0121 11:10:59.055485 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:10:59 crc kubenswrapper[4824]: I0121 11:10:59.055496 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:10:59 crc kubenswrapper[4824]: I0121 11:10:59.056473 4824 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="595783a1454735bd33f4862180b60d99a4e1b6fae1e735f2b7328391951bf422" exitCode=0 Jan 21 11:10:59 crc kubenswrapper[4824]: I0121 11:10:59.056537 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"595783a1454735bd33f4862180b60d99a4e1b6fae1e735f2b7328391951bf422"} Jan 21 11:10:59 crc kubenswrapper[4824]: I0121 11:10:59.056567 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"b7b5a9d4ae92b18983fc96eac50835eb2262c50f8125af46811cfbd1c64137a4"} Jan 21 11:10:59 crc kubenswrapper[4824]: I0121 11:10:59.056675 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 21 11:10:59 crc kubenswrapper[4824]: I0121 11:10:59.057328 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:10:59 crc kubenswrapper[4824]: I0121 11:10:59.057353 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:10:59 crc kubenswrapper[4824]: I0121 11:10:59.057363 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:10:59 crc kubenswrapper[4824]: I0121 11:10:59.057944 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"43529d1b9e1e5e22b0d56b023946119da8aacbae54b68c26ff0a1a82dc890c70"} Jan 21 11:10:59 crc kubenswrapper[4824]: I0121 11:10:59.058014 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"74643405e4bd126e4c118bb553357012409644db3d7dc245d90ad17473d46cfa"} Jan 21 11:10:59 crc kubenswrapper[4824]: I0121 11:10:59.059294 4824 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="b35ad7dd0570c191eb0080739304f3cd43f17b3d1dddf895549441ebe50deaa6" exitCode=0 Jan 21 11:10:59 crc kubenswrapper[4824]: I0121 11:10:59.059331 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"b35ad7dd0570c191eb0080739304f3cd43f17b3d1dddf895549441ebe50deaa6"} Jan 21 11:10:59 crc kubenswrapper[4824]: I0121 11:10:59.059353 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"4ad861050920fe34134874ba9dc73c919058a01b36434d0a996bb3f687549498"} Jan 21 11:10:59 crc kubenswrapper[4824]: I0121 11:10:59.059437 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 21 11:10:59 crc kubenswrapper[4824]: I0121 11:10:59.060068 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:10:59 crc kubenswrapper[4824]: I0121 11:10:59.060094 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:10:59 crc kubenswrapper[4824]: I0121 11:10:59.060103 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:10:59 crc kubenswrapper[4824]: W0121 11:10:59.167129 4824 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 192.168.26.120:6443: connect: connection refused Jan 21 11:10:59 crc kubenswrapper[4824]: E0121 11:10:59.167210 4824 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 192.168.26.120:6443: connect: connection refused" logger="UnhandledError" Jan 21 11:10:59 crc kubenswrapper[4824]: W0121 11:10:59.188181 4824 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 192.168.26.120:6443: connect: connection refused Jan 21 11:10:59 crc kubenswrapper[4824]: E0121 11:10:59.188240 4824 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 192.168.26.120:6443: connect: connection refused" logger="UnhandledError" Jan 21 11:10:59 crc kubenswrapper[4824]: E0121 11:10:59.382182 4824 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 192.168.26.120:6443: connect: connection refused" interval="1.6s" Jan 21 11:10:59 crc kubenswrapper[4824]: W0121 11:10:59.444526 4824 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 192.168.26.120:6443: connect: connection refused Jan 21 11:10:59 crc kubenswrapper[4824]: E0121 11:10:59.444584 4824 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 192.168.26.120:6443: connect: connection refused" logger="UnhandledError" Jan 21 11:10:59 crc kubenswrapper[4824]: I0121 11:10:59.586427 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 21 11:10:59 crc kubenswrapper[4824]: I0121 11:10:59.588783 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:10:59 crc kubenswrapper[4824]: I0121 11:10:59.588815 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:10:59 crc kubenswrapper[4824]: I0121 11:10:59.588825 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:10:59 crc kubenswrapper[4824]: I0121 11:10:59.588844 4824 kubelet_node_status.go:76] "Attempting to register node" node="crc" Jan 21 11:10:59 crc kubenswrapper[4824]: E0121 11:10:59.589294 4824 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 192.168.26.120:6443: connect: connection refused" node="crc" Jan 21 11:10:59 crc kubenswrapper[4824]: I0121 11:10:59.980517 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-09 20:56:54.433665752 +0000 UTC Jan 21 11:11:00 crc kubenswrapper[4824]: I0121 11:11:00.062844 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"1fd58f2438b6f9cbb2c37687e39f02cdf96a029477b1ae33f6dc9f3b70ecc926"} Jan 21 11:11:00 crc kubenswrapper[4824]: I0121 11:11:00.062937 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 21 11:11:00 crc kubenswrapper[4824]: I0121 11:11:00.064192 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:00 crc kubenswrapper[4824]: I0121 11:11:00.064216 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:00 crc kubenswrapper[4824]: I0121 11:11:00.064224 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:00 crc kubenswrapper[4824]: I0121 11:11:00.065510 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"b85db961e1e078e1f54323c79a99c9413365256ac2fce411db7932bd0580ab6a"} Jan 21 11:11:00 crc kubenswrapper[4824]: I0121 11:11:00.065532 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"a2690a1ecbc8fe1290e68bba2695a3edd833b5da933fb71e8ddc926a1340a9a3"} Jan 21 11:11:00 crc kubenswrapper[4824]: I0121 11:11:00.065542 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"e4d41492c3802a8168fffd249c15aeb9c902a5c7e01550e1aa6f9b769adfdc4f"} Jan 21 11:11:00 crc kubenswrapper[4824]: I0121 11:11:00.065593 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 21 11:11:00 crc kubenswrapper[4824]: I0121 11:11:00.066175 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:00 crc kubenswrapper[4824]: I0121 11:11:00.066192 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:00 crc kubenswrapper[4824]: I0121 11:11:00.066199 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:00 crc kubenswrapper[4824]: I0121 11:11:00.068149 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"bc6c7eff9e17b8a8ecc88bbfe703b5a01dbf1e02d342cede4f6a263a6526af43"} Jan 21 11:11:00 crc kubenswrapper[4824]: I0121 11:11:00.068185 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"9ea232ad87393685aa323ce6393ff3869f8f8d423c2406a501c69c355535adfa"} Jan 21 11:11:00 crc kubenswrapper[4824]: I0121 11:11:00.068195 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"05e076febd697ed802ce5c7ee92a8c2c636a1a2c6609ad7eb01f17dace247c4e"} Jan 21 11:11:00 crc kubenswrapper[4824]: I0121 11:11:00.068264 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 21 11:11:00 crc kubenswrapper[4824]: I0121 11:11:00.069071 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:00 crc kubenswrapper[4824]: I0121 11:11:00.069093 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:00 crc kubenswrapper[4824]: I0121 11:11:00.069100 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:00 crc kubenswrapper[4824]: I0121 11:11:00.070395 4824 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="80cd5ea5e6fc19a88c1aab5c61f827f103ab798a5c84da8dc4d45cea86d5aacb" exitCode=0 Jan 21 11:11:00 crc kubenswrapper[4824]: I0121 11:11:00.070444 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"80cd5ea5e6fc19a88c1aab5c61f827f103ab798a5c84da8dc4d45cea86d5aacb"} Jan 21 11:11:00 crc kubenswrapper[4824]: I0121 11:11:00.070527 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 21 11:11:00 crc kubenswrapper[4824]: I0121 11:11:00.071000 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:00 crc kubenswrapper[4824]: I0121 11:11:00.071026 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:00 crc kubenswrapper[4824]: I0121 11:11:00.071035 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:00 crc kubenswrapper[4824]: I0121 11:11:00.072873 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"af11ad95bcf0c6bbd5fa4ab99bb592ce1c87f40d29c5ade7832414315820a569"} Jan 21 11:11:00 crc kubenswrapper[4824]: I0121 11:11:00.072898 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"48070c09fad511108ff22e44211cf26911509b8406d66948985cbfc3009d8233"} Jan 21 11:11:00 crc kubenswrapper[4824]: I0121 11:11:00.072908 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"d7514786602746dabfa690da48d48e233fe68d6dd6674aa5ecd489c0b3185f08"} Jan 21 11:11:00 crc kubenswrapper[4824]: I0121 11:11:00.072917 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"61b14b6846f94cdeba8f1f61b3754e7663041f92c70356c4ee6cbb988a2e680f"} Jan 21 11:11:00 crc kubenswrapper[4824]: I0121 11:11:00.072925 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"a66b5c33884af00408dc35cf02f08ebd8570bd695f9f1fe558bf54fd06930431"} Jan 21 11:11:00 crc kubenswrapper[4824]: I0121 11:11:00.072999 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 21 11:11:00 crc kubenswrapper[4824]: I0121 11:11:00.073478 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:00 crc kubenswrapper[4824]: I0121 11:11:00.073512 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:00 crc kubenswrapper[4824]: I0121 11:11:00.073521 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:00 crc kubenswrapper[4824]: I0121 11:11:00.099155 4824 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Jan 21 11:11:00 crc kubenswrapper[4824]: I0121 11:11:00.860107 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 21 11:11:00 crc kubenswrapper[4824]: I0121 11:11:00.865169 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 21 11:11:00 crc kubenswrapper[4824]: I0121 11:11:00.981203 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-13 06:18:45.871212848 +0000 UTC Jan 21 11:11:01 crc kubenswrapper[4824]: I0121 11:11:01.071422 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 21 11:11:01 crc kubenswrapper[4824]: I0121 11:11:01.076445 4824 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="234d8b914e6a88480021647ae65d74e9b21c55b6788312a094d00c8a42677d48" exitCode=0 Jan 21 11:11:01 crc kubenswrapper[4824]: I0121 11:11:01.076541 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 21 11:11:01 crc kubenswrapper[4824]: I0121 11:11:01.076989 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"234d8b914e6a88480021647ae65d74e9b21c55b6788312a094d00c8a42677d48"} Jan 21 11:11:01 crc kubenswrapper[4824]: I0121 11:11:01.077029 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 21 11:11:01 crc kubenswrapper[4824]: I0121 11:11:01.077093 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 21 11:11:01 crc kubenswrapper[4824]: I0121 11:11:01.077112 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 21 11:11:01 crc kubenswrapper[4824]: I0121 11:11:01.077266 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:01 crc kubenswrapper[4824]: I0121 11:11:01.077297 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:01 crc kubenswrapper[4824]: I0121 11:11:01.077307 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:01 crc kubenswrapper[4824]: I0121 11:11:01.078211 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:01 crc kubenswrapper[4824]: I0121 11:11:01.078237 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:01 crc kubenswrapper[4824]: I0121 11:11:01.078246 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:01 crc kubenswrapper[4824]: I0121 11:11:01.078776 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:01 crc kubenswrapper[4824]: I0121 11:11:01.078805 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:01 crc kubenswrapper[4824]: I0121 11:11:01.078814 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:01 crc kubenswrapper[4824]: I0121 11:11:01.190283 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 21 11:11:01 crc kubenswrapper[4824]: I0121 11:11:01.190876 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:01 crc kubenswrapper[4824]: I0121 11:11:01.190903 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:01 crc kubenswrapper[4824]: I0121 11:11:01.190913 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:01 crc kubenswrapper[4824]: I0121 11:11:01.190928 4824 kubelet_node_status.go:76] "Attempting to register node" node="crc" Jan 21 11:11:01 crc kubenswrapper[4824]: I0121 11:11:01.981429 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-07 06:01:05.041617624 +0000 UTC Jan 21 11:11:02 crc kubenswrapper[4824]: I0121 11:11:02.083240 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"52758bcd31f2ba64edf03adfd4d1be6428015abff5b64c7c029fa787085f9c29"} Jan 21 11:11:02 crc kubenswrapper[4824]: I0121 11:11:02.083305 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"3b63e45d66bcd11ff3479cc3ec6b450017da8d21d36e967b8d496ae0f24b0ee9"} Jan 21 11:11:02 crc kubenswrapper[4824]: I0121 11:11:02.083321 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"ca06234486b59cf990c69d9501e23f935328e2ba21013cd71ed64831ea749bcb"} Jan 21 11:11:02 crc kubenswrapper[4824]: I0121 11:11:02.083333 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"d25c94f1ae0bbabb733f0e1ea1d1ba39562a92af73047a39aacc03013e239681"} Jan 21 11:11:02 crc kubenswrapper[4824]: I0121 11:11:02.083342 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"2736e73342d2adb175b0721dd4f40cc42d26fca06d5eb3334927518531c3b107"} Jan 21 11:11:02 crc kubenswrapper[4824]: I0121 11:11:02.083317 4824 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Jan 21 11:11:02 crc kubenswrapper[4824]: I0121 11:11:02.083409 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 21 11:11:02 crc kubenswrapper[4824]: I0121 11:11:02.083441 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 21 11:11:02 crc kubenswrapper[4824]: I0121 11:11:02.083349 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 21 11:11:02 crc kubenswrapper[4824]: I0121 11:11:02.084576 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:02 crc kubenswrapper[4824]: I0121 11:11:02.084611 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:02 crc kubenswrapper[4824]: I0121 11:11:02.084624 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:02 crc kubenswrapper[4824]: I0121 11:11:02.084583 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:02 crc kubenswrapper[4824]: I0121 11:11:02.084771 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:02 crc kubenswrapper[4824]: I0121 11:11:02.084783 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:02 crc kubenswrapper[4824]: I0121 11:11:02.084589 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:02 crc kubenswrapper[4824]: I0121 11:11:02.084969 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:02 crc kubenswrapper[4824]: I0121 11:11:02.084984 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:02 crc kubenswrapper[4824]: I0121 11:11:02.981900 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-17 19:34:53.295504302 +0000 UTC Jan 21 11:11:03 crc kubenswrapper[4824]: I0121 11:11:03.695045 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 21 11:11:03 crc kubenswrapper[4824]: I0121 11:11:03.695341 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 21 11:11:03 crc kubenswrapper[4824]: I0121 11:11:03.696248 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:03 crc kubenswrapper[4824]: I0121 11:11:03.696272 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:03 crc kubenswrapper[4824]: I0121 11:11:03.696281 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:03 crc kubenswrapper[4824]: I0121 11:11:03.714631 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Jan 21 11:11:03 crc kubenswrapper[4824]: I0121 11:11:03.714867 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 21 11:11:03 crc kubenswrapper[4824]: I0121 11:11:03.715531 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:03 crc kubenswrapper[4824]: I0121 11:11:03.715560 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:03 crc kubenswrapper[4824]: I0121 11:11:03.715569 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:03 crc kubenswrapper[4824]: I0121 11:11:03.896826 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 21 11:11:03 crc kubenswrapper[4824]: I0121 11:11:03.981981 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-26 02:41:02.681865632 +0000 UTC Jan 21 11:11:04 crc kubenswrapper[4824]: I0121 11:11:04.086004 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 21 11:11:04 crc kubenswrapper[4824]: I0121 11:11:04.086769 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:04 crc kubenswrapper[4824]: I0121 11:11:04.086806 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:04 crc kubenswrapper[4824]: I0121 11:11:04.086815 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:04 crc kubenswrapper[4824]: I0121 11:11:04.879469 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 21 11:11:04 crc kubenswrapper[4824]: I0121 11:11:04.879600 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 21 11:11:04 crc kubenswrapper[4824]: I0121 11:11:04.880519 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:04 crc kubenswrapper[4824]: I0121 11:11:04.880556 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:04 crc kubenswrapper[4824]: I0121 11:11:04.880565 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:04 crc kubenswrapper[4824]: I0121 11:11:04.982539 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-12 05:22:46.275144861 +0000 UTC Jan 21 11:11:05 crc kubenswrapper[4824]: I0121 11:11:05.656444 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 21 11:11:05 crc kubenswrapper[4824]: I0121 11:11:05.656558 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 21 11:11:05 crc kubenswrapper[4824]: I0121 11:11:05.657378 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:05 crc kubenswrapper[4824]: I0121 11:11:05.657400 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:05 crc kubenswrapper[4824]: I0121 11:11:05.657409 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:05 crc kubenswrapper[4824]: I0121 11:11:05.788621 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Jan 21 11:11:05 crc kubenswrapper[4824]: I0121 11:11:05.788740 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 21 11:11:05 crc kubenswrapper[4824]: I0121 11:11:05.789486 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:05 crc kubenswrapper[4824]: I0121 11:11:05.789506 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:05 crc kubenswrapper[4824]: I0121 11:11:05.789513 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:05 crc kubenswrapper[4824]: I0121 11:11:05.983156 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-20 15:24:02.024830589 +0000 UTC Jan 21 11:11:06 crc kubenswrapper[4824]: I0121 11:11:06.054634 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Jan 21 11:11:06 crc kubenswrapper[4824]: I0121 11:11:06.090161 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 21 11:11:06 crc kubenswrapper[4824]: I0121 11:11:06.092121 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:06 crc kubenswrapper[4824]: I0121 11:11:06.092165 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:06 crc kubenswrapper[4824]: I0121 11:11:06.092176 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:06 crc kubenswrapper[4824]: I0121 11:11:06.897224 4824 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Jan 21 11:11:06 crc kubenswrapper[4824]: I0121 11:11:06.897635 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Jan 21 11:11:06 crc kubenswrapper[4824]: I0121 11:11:06.984098 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-03 01:14:23.252791731 +0000 UTC Jan 21 11:11:07 crc kubenswrapper[4824]: I0121 11:11:07.984837 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-06 11:31:12.854586727 +0000 UTC Jan 21 11:11:08 crc kubenswrapper[4824]: E0121 11:11:08.085993 4824 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Jan 21 11:11:08 crc kubenswrapper[4824]: I0121 11:11:08.985429 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-18 16:11:29.852049635 +0000 UTC Jan 21 11:11:09 crc kubenswrapper[4824]: I0121 11:11:09.587345 4824 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Jan 21 11:11:09 crc kubenswrapper[4824]: I0121 11:11:09.587409 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Jan 21 11:11:09 crc kubenswrapper[4824]: I0121 11:11:09.600217 4824 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Jan 21 11:11:09 crc kubenswrapper[4824]: I0121 11:11:09.600262 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Jan 21 11:11:09 crc kubenswrapper[4824]: I0121 11:11:09.986423 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-25 17:06:48.926253659 +0000 UTC Jan 21 11:11:10 crc kubenswrapper[4824]: I0121 11:11:10.020447 4824 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Jan 21 11:11:10 crc kubenswrapper[4824]: I0121 11:11:10.020505 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Jan 21 11:11:10 crc kubenswrapper[4824]: I0121 11:11:10.083189 4824 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Liveness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Jan 21 11:11:10 crc kubenswrapper[4824]: I0121 11:11:10.083235 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Jan 21 11:11:10 crc kubenswrapper[4824]: I0121 11:11:10.986687 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-22 08:56:00.015504719 +0000 UTC Jan 21 11:11:11 crc kubenswrapper[4824]: I0121 11:11:11.075517 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 21 11:11:11 crc kubenswrapper[4824]: I0121 11:11:11.075627 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 21 11:11:11 crc kubenswrapper[4824]: I0121 11:11:11.075984 4824 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Jan 21 11:11:11 crc kubenswrapper[4824]: I0121 11:11:11.076047 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Jan 21 11:11:11 crc kubenswrapper[4824]: I0121 11:11:11.076373 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:11 crc kubenswrapper[4824]: I0121 11:11:11.076399 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:11 crc kubenswrapper[4824]: I0121 11:11:11.076408 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:11 crc kubenswrapper[4824]: I0121 11:11:11.078765 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 21 11:11:11 crc kubenswrapper[4824]: I0121 11:11:11.098364 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 21 11:11:11 crc kubenswrapper[4824]: I0121 11:11:11.098644 4824 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Jan 21 11:11:11 crc kubenswrapper[4824]: I0121 11:11:11.098678 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Jan 21 11:11:11 crc kubenswrapper[4824]: I0121 11:11:11.098908 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:11 crc kubenswrapper[4824]: I0121 11:11:11.098933 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:11 crc kubenswrapper[4824]: I0121 11:11:11.098943 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:11 crc kubenswrapper[4824]: I0121 11:11:11.987773 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-28 12:31:04.691331515 +0000 UTC Jan 21 11:11:12 crc kubenswrapper[4824]: I0121 11:11:12.988491 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-17 17:33:37.525259014 +0000 UTC Jan 21 11:11:13 crc kubenswrapper[4824]: I0121 11:11:13.698621 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 21 11:11:13 crc kubenswrapper[4824]: I0121 11:11:13.698733 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 21 11:11:13 crc kubenswrapper[4824]: I0121 11:11:13.699558 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:13 crc kubenswrapper[4824]: I0121 11:11:13.699583 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:13 crc kubenswrapper[4824]: I0121 11:11:13.699592 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:13 crc kubenswrapper[4824]: I0121 11:11:13.989017 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-27 14:45:26.821735303 +0000 UTC Jan 21 11:11:14 crc kubenswrapper[4824]: E0121 11:11:14.570662 4824 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded" interval="3.2s" Jan 21 11:11:14 crc kubenswrapper[4824]: I0121 11:11:14.572042 4824 trace.go:236] Trace[1618589615]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (21-Jan-2026 11:11:01.321) (total time: 13250ms): Jan 21 11:11:14 crc kubenswrapper[4824]: Trace[1618589615]: ---"Objects listed" error: 13249ms (11:11:14.571) Jan 21 11:11:14 crc kubenswrapper[4824]: Trace[1618589615]: [13.250012041s] [13.250012041s] END Jan 21 11:11:14 crc kubenswrapper[4824]: I0121 11:11:14.572063 4824 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Jan 21 11:11:14 crc kubenswrapper[4824]: I0121 11:11:14.572437 4824 trace.go:236] Trace[479225197]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (21-Jan-2026 11:11:01.245) (total time: 13327ms): Jan 21 11:11:14 crc kubenswrapper[4824]: Trace[479225197]: ---"Objects listed" error: 13326ms (11:11:14.572) Jan 21 11:11:14 crc kubenswrapper[4824]: Trace[479225197]: [13.327010649s] [13.327010649s] END Jan 21 11:11:14 crc kubenswrapper[4824]: I0121 11:11:14.572460 4824 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Jan 21 11:11:14 crc kubenswrapper[4824]: E0121 11:11:14.573032 4824 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Jan 21 11:11:14 crc kubenswrapper[4824]: I0121 11:11:14.573349 4824 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Jan 21 11:11:14 crc kubenswrapper[4824]: I0121 11:11:14.573575 4824 trace.go:236] Trace[577277594]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (21-Jan-2026 11:11:01.344) (total time: 13229ms): Jan 21 11:11:14 crc kubenswrapper[4824]: Trace[577277594]: ---"Objects listed" error: 13229ms (11:11:14.573) Jan 21 11:11:14 crc kubenswrapper[4824]: Trace[577277594]: [13.229062201s] [13.229062201s] END Jan 21 11:11:14 crc kubenswrapper[4824]: I0121 11:11:14.573594 4824 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Jan 21 11:11:14 crc kubenswrapper[4824]: I0121 11:11:14.574207 4824 trace.go:236] Trace[1885941545]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (21-Jan-2026 11:11:01.272) (total time: 13301ms): Jan 21 11:11:14 crc kubenswrapper[4824]: Trace[1885941545]: ---"Objects listed" error: 13301ms (11:11:14.574) Jan 21 11:11:14 crc kubenswrapper[4824]: Trace[1885941545]: [13.301572103s] [13.301572103s] END Jan 21 11:11:14 crc kubenswrapper[4824]: I0121 11:11:14.574228 4824 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Jan 21 11:11:14 crc kubenswrapper[4824]: I0121 11:11:14.606172 4824 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Jan 21 11:11:14 crc kubenswrapper[4824]: I0121 11:11:14.611110 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 21 11:11:14 crc kubenswrapper[4824]: I0121 11:11:14.616058 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 21 11:11:14 crc kubenswrapper[4824]: I0121 11:11:14.621616 4824 csr.go:261] certificate signing request csr-pz9z5 is approved, waiting to be issued Jan 21 11:11:14 crc kubenswrapper[4824]: I0121 11:11:14.626810 4824 csr.go:257] certificate signing request csr-pz9z5 is issued Jan 21 11:11:14 crc kubenswrapper[4824]: I0121 11:11:14.975684 4824 apiserver.go:52] "Watching apiserver" Jan 21 11:11:14 crc kubenswrapper[4824]: I0121 11:11:14.977825 4824 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Jan 21 11:11:14 crc kubenswrapper[4824]: I0121 11:11:14.978142 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g"] Jan 21 11:11:14 crc kubenswrapper[4824]: I0121 11:11:14.978467 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Jan 21 11:11:14 crc kubenswrapper[4824]: I0121 11:11:14.978482 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 11:11:14 crc kubenswrapper[4824]: I0121 11:11:14.978592 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 21 11:11:14 crc kubenswrapper[4824]: I0121 11:11:14.978689 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 11:11:14 crc kubenswrapper[4824]: I0121 11:11:14.978804 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Jan 21 11:11:14 crc kubenswrapper[4824]: E0121 11:11:14.978856 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 21 11:11:14 crc kubenswrapper[4824]: I0121 11:11:14.979070 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 11:11:14 crc kubenswrapper[4824]: E0121 11:11:14.979135 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 21 11:11:14 crc kubenswrapper[4824]: E0121 11:11:14.979068 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 21 11:11:14 crc kubenswrapper[4824]: I0121 11:11:14.979831 4824 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Jan 21 11:11:14 crc kubenswrapper[4824]: I0121 11:11:14.980269 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Jan 21 11:11:14 crc kubenswrapper[4824]: I0121 11:11:14.980506 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Jan 21 11:11:14 crc kubenswrapper[4824]: I0121 11:11:14.980517 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Jan 21 11:11:14 crc kubenswrapper[4824]: I0121 11:11:14.980590 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Jan 21 11:11:14 crc kubenswrapper[4824]: I0121 11:11:14.980604 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Jan 21 11:11:14 crc kubenswrapper[4824]: I0121 11:11:14.981123 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Jan 21 11:11:14 crc kubenswrapper[4824]: I0121 11:11:14.981647 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Jan 21 11:11:14 crc kubenswrapper[4824]: I0121 11:11:14.981939 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Jan 21 11:11:14 crc kubenswrapper[4824]: I0121 11:11:14.982107 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Jan 21 11:11:14 crc kubenswrapper[4824]: I0121 11:11:14.989087 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-04 18:51:07.535069084 +0000 UTC Jan 21 11:11:14 crc kubenswrapper[4824]: I0121 11:11:14.991994 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.000489 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.013572 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.022415 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.029137 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9bb99c0a-683b-44da-9c0e-e8d284e87d53\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05e076febd697ed802ce5c7ee92a8c2c636a1a2c6609ad7eb01f17dace247c4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43529d1b9e1e5e22b0d56b023946119da8aacbae54b68c26ff0a1a82dc890c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ea232ad87393685aa323ce6393ff3869f8f8d423c2406a501c69c355535adfa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc6c7eff9e17b8a8ecc88bbfe703b5a01dbf1e02d342cede4f6a263a6526af43\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.036791 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.043403 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.049719 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.056243 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.075935 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.076000 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.076022 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.076037 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.076051 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.076066 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.076083 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.076097 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.076111 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.076125 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.076139 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.076155 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.076170 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.076185 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.076202 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.076217 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.076230 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.076245 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.076259 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.076273 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.076270 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.076291 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.076350 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.076372 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.076389 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.076404 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.076419 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.076433 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.076447 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.076460 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.076475 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.076491 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.076507 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.076529 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.076544 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.076557 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.076559 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.076578 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.076596 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.076610 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.076626 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.076639 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.076653 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.076667 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.076680 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.076694 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.076708 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.076721 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.076726 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.076727 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.076792 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.076830 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.076850 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.076866 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.076883 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.076917 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.076931 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.076992 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.077010 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.077014 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.077024 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.077040 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.077056 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.077071 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.077087 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.077103 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.077117 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.077131 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.077148 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.077164 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.077178 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.077193 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.077206 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.077212 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.077219 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.077251 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.077269 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.077284 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.077299 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.077319 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.077336 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.077350 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.077366 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.077375 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.077380 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.077381 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.077425 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.077442 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.077468 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.077477 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.077488 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.077517 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.077542 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.077570 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.077576 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.077591 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.077610 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.077627 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.077645 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.077651 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.077657 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.077663 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.077694 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.077717 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.077733 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.077745 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.077749 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.077788 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.077787 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.077805 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.077817 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.077835 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.077850 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.077867 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.077882 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.077895 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.077910 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.077916 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.077925 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.077941 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.077975 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.078006 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.078018 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.078023 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.078024 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.078037 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.078055 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.078074 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.078107 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.078124 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.078139 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.078154 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.078161 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.078169 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.078186 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.078201 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.078217 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.078232 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.078247 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.078264 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.078281 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.078296 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.078312 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.078329 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.078344 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.078362 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.078377 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.078392 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.078408 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.078423 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.078439 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.078454 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.078470 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.078486 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.078501 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.078517 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.078533 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.078548 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.078563 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.078577 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.078594 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.078610 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.078625 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.078640 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.078657 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.078672 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.078688 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.078702 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.078718 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.078733 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.078748 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.078763 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.078781 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.078799 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.078816 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.078832 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.078847 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.078864 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.078879 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.078902 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.078919 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.078934 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.078950 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.078993 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.079009 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.079024 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.079040 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.079090 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.079108 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.079124 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.079140 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.079156 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.079172 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.079190 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.079207 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.079223 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.079238 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.079252 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.079271 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.079286 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.079302 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.079320 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.079335 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.079351 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.079367 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.079383 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.079399 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.079415 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.079430 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.079447 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.079462 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.079478 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.079494 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.079510 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.079526 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.079542 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.079558 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.079594 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.079616 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.079635 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.079651 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.079669 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.079688 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.079710 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.079724 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.079742 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.079757 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.079776 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.079794 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.079811 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.079828 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.079878 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.079889 4824 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.079899 4824 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.079909 4824 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.079919 4824 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.079928 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.080255 4824 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.080270 4824 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.080316 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.078214 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.078273 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.078279 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.078305 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.078340 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.078389 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.078471 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.078493 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.078525 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.078542 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.085566 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.078613 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.078626 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.078646 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.078670 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.078743 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.078765 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.078797 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.078876 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.078948 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.079099 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.079124 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.079226 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.079264 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.079254 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.079330 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.079407 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.079455 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.079500 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.079510 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.079668 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.079701 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.079866 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.080063 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.080058 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.080070 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.080171 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.080300 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.080321 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.080355 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.080512 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.080623 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.080736 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.081740 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.081758 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.081864 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: E0121 11:11:15.082020 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-21 11:11:15.581952136 +0000 UTC m=+17.874981427 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.083206 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.083262 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.083851 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.084122 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.084139 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.084206 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.084198 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.084240 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.084383 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.084419 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.084602 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.084725 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.084871 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.084887 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.085042 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.085172 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.085247 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.085260 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.085271 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.085283 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.085298 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.085188 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.085454 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.085476 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.085580 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.085623 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.085661 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.085680 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.085696 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.085945 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.086042 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.086085 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.086178 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.086234 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.086292 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.086331 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.086420 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.086849 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.086944 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.086950 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.087891 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.088164 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.088292 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.088318 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.088363 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.088647 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.088782 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.088851 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.089013 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.089113 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.089177 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.089143 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.089282 4824 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.089309 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.089372 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.089465 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.089505 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.089642 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.089732 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.089754 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: E0121 11:11:15.089834 4824 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 21 11:11:15 crc kubenswrapper[4824]: E0121 11:11:15.089884 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-21 11:11:15.589869401 +0000 UTC m=+17.882898693 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.089993 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.089998 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: E0121 11:11:15.090094 4824 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Jan 21 11:11:15 crc kubenswrapper[4824]: E0121 11:11:15.090126 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-21 11:11:15.59011691 +0000 UTC m=+17.883146203 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.090133 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.090160 4824 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.090192 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.090268 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.090314 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.090337 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.089834 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.090637 4824 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.090674 4824 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.090687 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.090697 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.090726 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.090737 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.090746 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.091086 4824 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.091107 4824 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.091120 4824 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.091135 4824 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.091144 4824 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.091209 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.091291 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.091334 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.091509 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.091541 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.092245 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.092287 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.092381 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.093311 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.093494 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.093607 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.093718 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.093906 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.093912 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.094073 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.094102 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.095508 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.095973 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.097082 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-lv4rn"] Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.097120 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.097358 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-lv4rn" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.098085 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.099814 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.100024 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.100255 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.100835 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.101918 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.102977 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.105200 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.105428 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.105689 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.105827 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.106040 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Jan 21 11:11:15 crc kubenswrapper[4824]: E0121 11:11:15.106059 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 21 11:11:15 crc kubenswrapper[4824]: E0121 11:11:15.106073 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 21 11:11:15 crc kubenswrapper[4824]: E0121 11:11:15.106084 4824 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.106127 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: E0121 11:11:15.106133 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-01-21 11:11:15.606122217 +0000 UTC m=+17.899151509 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.106215 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.106259 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.106279 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Jan 21 11:11:15 crc kubenswrapper[4824]: E0121 11:11:15.107423 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.108062 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 21 11:11:15 crc kubenswrapper[4824]: E0121 11:11:15.108151 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 21 11:11:15 crc kubenswrapper[4824]: E0121 11:11:15.108179 4824 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 21 11:11:15 crc kubenswrapper[4824]: E0121 11:11:15.108255 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-01-21 11:11:15.608236301 +0000 UTC m=+17.901265593 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.108262 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.108451 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.108629 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.108540 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.109132 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.109189 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.109220 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.109280 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.109413 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.109636 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.109430 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.109559 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.109882 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.109947 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.109654 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.109951 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.110040 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.110059 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.110101 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.110993 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.111235 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.111300 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.111511 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.111810 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.112368 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.114463 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9bb99c0a-683b-44da-9c0e-e8d284e87d53\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05e076febd697ed802ce5c7ee92a8c2c636a1a2c6609ad7eb01f17dace247c4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43529d1b9e1e5e22b0d56b023946119da8aacbae54b68c26ff0a1a82dc890c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ea232ad87393685aa323ce6393ff3869f8f8d423c2406a501c69c355535adfa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc6c7eff9e17b8a8ecc88bbfe703b5a01dbf1e02d342cede4f6a263a6526af43\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.116847 4824 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="af11ad95bcf0c6bbd5fa4ab99bb592ce1c87f40d29c5ade7832414315820a569" exitCode=255 Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.117539 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"af11ad95bcf0c6bbd5fa4ab99bb592ce1c87f40d29c5ade7832414315820a569"} Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.119595 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.119791 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.120801 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.120836 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: E0121 11:11:15.123192 4824 kubelet.go:1929] "Failed creating a mirror pod for" err="pods \"kube-controller-manager-crc\" already exists" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.124145 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.125274 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.125732 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.127897 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.129996 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.133594 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.136213 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.142126 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.142320 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.142536 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.142815 4824 scope.go:117] "RemoveContainer" containerID="af11ad95bcf0c6bbd5fa4ab99bb592ce1c87f40d29c5ade7832414315820a569" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.147562 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lv4rn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1267539-9adf-466f-8027-8ede009d19a0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbmfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lv4rn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.153267 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.159420 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.165451 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.174398 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.180378 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.186636 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.191941 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.192113 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.192152 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/c1267539-9adf-466f-8027-8ede009d19a0-hosts-file\") pod \"node-resolver-lv4rn\" (UID: \"c1267539-9adf-466f-8027-8ede009d19a0\") " pod="openshift-dns/node-resolver-lv4rn" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.192183 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dbmfl\" (UniqueName: \"kubernetes.io/projected/c1267539-9adf-466f-8027-8ede009d19a0-kube-api-access-dbmfl\") pod \"node-resolver-lv4rn\" (UID: \"c1267539-9adf-466f-8027-8ede009d19a0\") " pod="openshift-dns/node-resolver-lv4rn" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.192198 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.192276 4824 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.192293 4824 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.192303 4824 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.192312 4824 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.192320 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.192329 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.192337 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.192344 4824 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.192355 4824 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.192363 4824 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.192371 4824 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.192379 4824 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.192386 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.192394 4824 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.192402 4824 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.192410 4824 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.192418 4824 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.192426 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.192434 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.192442 4824 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.192451 4824 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.192459 4824 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.192467 4824 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.192475 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.192482 4824 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.192491 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.192498 4824 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.192505 4824 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.192513 4824 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.192520 4824 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.192528 4824 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.192534 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.192542 4824 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.192549 4824 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.192557 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.192566 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.192574 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.192582 4824 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.192590 4824 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.192598 4824 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.192606 4824 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.192614 4824 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.192622 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.192630 4824 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.192638 4824 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.192646 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.192655 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.192663 4824 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.192671 4824 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.192678 4824 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.192686 4824 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.192692 4824 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.192699 4824 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.192707 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.192718 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.192727 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.192735 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.192743 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.192750 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.192759 4824 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.192766 4824 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.192779 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.192787 4824 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.192794 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.192802 4824 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.192811 4824 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.192818 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.192825 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.192833 4824 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.192841 4824 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.192848 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.192856 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.192865 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.192872 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.192880 4824 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.192887 4824 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.192894 4824 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.192902 4824 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.192990 4824 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.193002 4824 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.193009 4824 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.193018 4824 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.193027 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.193035 4824 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.193042 4824 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.193051 4824 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.193060 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.193069 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.193077 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.193085 4824 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.193092 4824 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.193099 4824 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.193107 4824 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.193115 4824 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.193122 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.193130 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.193137 4824 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.193151 4824 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.193158 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.193166 4824 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.193174 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.193181 4824 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.193189 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.193197 4824 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.193283 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.193294 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.193302 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.193312 4824 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.193319 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.193326 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.193334 4824 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.193330 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.193344 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.193448 4824 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.193462 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.193472 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.193481 4824 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.193491 4824 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.193500 4824 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.193509 4824 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.193518 4824 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.193525 4824 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.193535 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.193544 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.193551 4824 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.193559 4824 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.193567 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.193575 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.193583 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.193591 4824 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.193599 4824 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.193606 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.193614 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.193622 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.193630 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.193637 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.193645 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.193653 4824 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.193660 4824 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.193668 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.193675 4824 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.193682 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.193692 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.193701 4824 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.193711 4824 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.193720 4824 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.193727 4824 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.193735 4824 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.193743 4824 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.193753 4824 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.193761 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.193769 4824 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.193777 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.193785 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.193792 4824 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.193801 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.193809 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.193817 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.193801 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.193824 4824 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.193932 4824 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.193943 4824 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.193952 4824 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.193977 4824 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.193999 4824 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.194007 4824 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.194016 4824 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.194023 4824 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.194031 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.194039 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.194047 4824 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.194054 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.194062 4824 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.194071 4824 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.199562 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.207541 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.212717 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lv4rn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1267539-9adf-466f-8027-8ede009d19a0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbmfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lv4rn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.225283 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d31ab61-f94c-4f49-85d0-c9e91b83069b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a66b5c33884af00408dc35cf02f08ebd8570bd695f9f1fe558bf54fd06930431\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7514786602746dabfa690da48d48e233fe68d6dd6674aa5ecd489c0b3185f08\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61b14b6846f94cdeba8f1f61b3754e7663041f92c70356c4ee6cbb988a2e680f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af11ad95bcf0c6bbd5fa4ab99bb592ce1c87f40d29c5ade7832414315820a569\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af11ad95bcf0c6bbd5fa4ab99bb592ce1c87f40d29c5ade7832414315820a569\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW0121 11:11:14.604290 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0121 11:11:14.604572 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0121 11:11:14.605875 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3284452394/tls.crt::/tmp/serving-cert-3284452394/tls.key\\\\\\\"\\\\nI0121 11:11:14.923427 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0121 11:11:14.928141 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0121 11:11:14.928162 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0121 11:11:14.928179 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0121 11:11:14.928184 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0121 11:11:14.933154 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0121 11:11:14.933230 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 11:11:14.933259 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 11:11:14.933282 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0121 11:11:14.933299 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0121 11:11:14.933315 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0121 11:11:14.933331 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0121 11:11:14.933177 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0121 11:11:14.934666 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48070c09fad511108ff22e44211cf26911509b8406d66948985cbfc3009d8233\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.234910 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9bb99c0a-683b-44da-9c0e-e8d284e87d53\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05e076febd697ed802ce5c7ee92a8c2c636a1a2c6609ad7eb01f17dace247c4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43529d1b9e1e5e22b0d56b023946119da8aacbae54b68c26ff0a1a82dc890c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ea232ad87393685aa323ce6393ff3869f8f8d423c2406a501c69c355535adfa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc6c7eff9e17b8a8ecc88bbfe703b5a01dbf1e02d342cede4f6a263a6526af43\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.289659 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.294512 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/c1267539-9adf-466f-8027-8ede009d19a0-hosts-file\") pod \"node-resolver-lv4rn\" (UID: \"c1267539-9adf-466f-8027-8ede009d19a0\") " pod="openshift-dns/node-resolver-lv4rn" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.294558 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dbmfl\" (UniqueName: \"kubernetes.io/projected/c1267539-9adf-466f-8027-8ede009d19a0-kube-api-access-dbmfl\") pod \"node-resolver-lv4rn\" (UID: \"c1267539-9adf-466f-8027-8ede009d19a0\") " pod="openshift-dns/node-resolver-lv4rn" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.294628 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/c1267539-9adf-466f-8027-8ede009d19a0-hosts-file\") pod \"node-resolver-lv4rn\" (UID: \"c1267539-9adf-466f-8027-8ede009d19a0\") " pod="openshift-dns/node-resolver-lv4rn" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.296467 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Jan 21 11:11:15 crc kubenswrapper[4824]: W0121 11:11:15.299017 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod37a5e44f_9a88_4405_be8a_b645485e7312.slice/crio-f6a34848b7ccc79f1145a419de2da50011e2e9a8aecedf718fec38ac701b9547 WatchSource:0}: Error finding container f6a34848b7ccc79f1145a419de2da50011e2e9a8aecedf718fec38ac701b9547: Status 404 returned error can't find the container with id f6a34848b7ccc79f1145a419de2da50011e2e9a8aecedf718fec38ac701b9547 Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.302243 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.307352 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dbmfl\" (UniqueName: \"kubernetes.io/projected/c1267539-9adf-466f-8027-8ede009d19a0-kube-api-access-dbmfl\") pod \"node-resolver-lv4rn\" (UID: \"c1267539-9adf-466f-8027-8ede009d19a0\") " pod="openshift-dns/node-resolver-lv4rn" Jan 21 11:11:15 crc kubenswrapper[4824]: W0121 11:11:15.310081 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd75a4c96_2883_4a0b_bab2_0fab2b6c0b49.slice/crio-5ae7529fc08fbc1bc45653348ce3eaf2bb6e0c95ad21544f7ef8f6655227058d WatchSource:0}: Error finding container 5ae7529fc08fbc1bc45653348ce3eaf2bb6e0c95ad21544f7ef8f6655227058d: Status 404 returned error can't find the container with id 5ae7529fc08fbc1bc45653348ce3eaf2bb6e0c95ad21544f7ef8f6655227058d Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.425674 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-lv4rn" Jan 21 11:11:15 crc kubenswrapper[4824]: W0121 11:11:15.442658 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc1267539_9adf_466f_8027_8ede009d19a0.slice/crio-44a63cef2a9562f6b836d0eea1fdc74e6cf08c709d90570463cdcccabafa476c WatchSource:0}: Error finding container 44a63cef2a9562f6b836d0eea1fdc74e6cf08c709d90570463cdcccabafa476c: Status 404 returned error can't find the container with id 44a63cef2a9562f6b836d0eea1fdc74e6cf08c709d90570463cdcccabafa476c Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.452278 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-zqjg9"] Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.452521 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-x94nb"] Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.452670 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-x94nb" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.452899 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" Jan 21 11:11:15 crc kubenswrapper[4824]: W0121 11:11:15.456574 4824 reflector.go:561] object-"openshift-multus"/"cni-copy-resources": failed to list *v1.ConfigMap: configmaps "cni-copy-resources" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-multus": no relationship found between node 'crc' and this object Jan 21 11:11:15 crc kubenswrapper[4824]: E0121 11:11:15.456620 4824 reflector.go:158] "Unhandled Error" err="object-\"openshift-multus\"/\"cni-copy-resources\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"cni-copy-resources\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-multus\": no relationship found between node 'crc' and this object" logger="UnhandledError" Jan 21 11:11:15 crc kubenswrapper[4824]: W0121 11:11:15.456625 4824 reflector.go:561] object-"openshift-machine-config-operator"/"openshift-service-ca.crt": failed to list *v1.ConfigMap: configmaps "openshift-service-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-machine-config-operator": no relationship found between node 'crc' and this object Jan 21 11:11:15 crc kubenswrapper[4824]: E0121 11:11:15.456651 4824 reflector.go:158] "Unhandled Error" err="object-\"openshift-machine-config-operator\"/\"openshift-service-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"openshift-service-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-machine-config-operator\": no relationship found between node 'crc' and this object" logger="UnhandledError" Jan 21 11:11:15 crc kubenswrapper[4824]: W0121 11:11:15.456750 4824 reflector.go:561] object-"openshift-multus"/"default-dockercfg-2q5b6": failed to list *v1.Secret: secrets "default-dockercfg-2q5b6" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-multus": no relationship found between node 'crc' and this object Jan 21 11:11:15 crc kubenswrapper[4824]: E0121 11:11:15.456772 4824 reflector.go:158] "Unhandled Error" err="object-\"openshift-multus\"/\"default-dockercfg-2q5b6\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"default-dockercfg-2q5b6\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-multus\": no relationship found between node 'crc' and this object" logger="UnhandledError" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.457072 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.457155 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.457356 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-gjtnh"] Jan 21 11:11:15 crc kubenswrapper[4824]: W0121 11:11:15.457383 4824 reflector.go:561] object-"openshift-multus"/"kube-root-ca.crt": failed to list *v1.ConfigMap: configmaps "kube-root-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-multus": no relationship found between node 'crc' and this object Jan 21 11:11:15 crc kubenswrapper[4824]: E0121 11:11:15.457412 4824 reflector.go:158] "Unhandled Error" err="object-\"openshift-multus\"/\"kube-root-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"kube-root-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-multus\": no relationship found between node 'crc' and this object" logger="UnhandledError" Jan 21 11:11:15 crc kubenswrapper[4824]: W0121 11:11:15.457453 4824 reflector.go:561] object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq": failed to list *v1.Secret: secrets "machine-config-daemon-dockercfg-r5tcq" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-machine-config-operator": no relationship found between node 'crc' and this object Jan 21 11:11:15 crc kubenswrapper[4824]: E0121 11:11:15.457466 4824 reflector.go:158] "Unhandled Error" err="object-\"openshift-machine-config-operator\"/\"machine-config-daemon-dockercfg-r5tcq\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"machine-config-daemon-dockercfg-r5tcq\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-machine-config-operator\": no relationship found between node 'crc' and this object" logger="UnhandledError" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.457505 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.457525 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.457531 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.457771 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-gjtnh" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.458093 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-htkvk"] Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.458727 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.462102 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.462239 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.462342 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.462506 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.462602 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.462701 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.462813 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.463126 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.464889 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.467562 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.475464 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x94nb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgr8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x94nb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.481269 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33f3d922-4ffe-409b-a49a-d88c85898260\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpwst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpwst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zqjg9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.488415 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.495082 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/6b9588d1-5876-4c2c-8de5-e5c0ad46f04e-host-var-lib-cni-bin\") pod \"multus-x94nb\" (UID: \"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\") " pod="openshift-multus/multus-x94nb" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.495110 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/6b9588d1-5876-4c2c-8de5-e5c0ad46f04e-etc-kubernetes\") pod \"multus-x94nb\" (UID: \"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\") " pod="openshift-multus/multus-x94nb" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.495127 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-host-run-ovn-kubernetes\") pod \"ovnkube-node-htkvk\" (UID: \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.495118 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.495152 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/6b9588d1-5876-4c2c-8de5-e5c0ad46f04e-cnibin\") pod \"multus-x94nb\" (UID: \"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\") " pod="openshift-multus/multus-x94nb" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.495168 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-run-systemd\") pod \"ovnkube-node-htkvk\" (UID: \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.495183 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/6b9588d1-5876-4c2c-8de5-e5c0ad46f04e-multus-conf-dir\") pod \"multus-x94nb\" (UID: \"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\") " pod="openshift-multus/multus-x94nb" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.495197 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-host-run-netns\") pod \"ovnkube-node-htkvk\" (UID: \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.495211 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/26b1108a-4b98-4e39-a3cd-e0d055089fd8-ovnkube-config\") pod \"ovnkube-node-htkvk\" (UID: \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.495310 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/6b9588d1-5876-4c2c-8de5-e5c0ad46f04e-host-run-netns\") pod \"multus-x94nb\" (UID: \"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\") " pod="openshift-multus/multus-x94nb" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.495329 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/33f3d922-4ffe-409b-a49a-d88c85898260-proxy-tls\") pod \"machine-config-daemon-zqjg9\" (UID: \"33f3d922-4ffe-409b-a49a-d88c85898260\") " pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.495342 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mpwst\" (UniqueName: \"kubernetes.io/projected/33f3d922-4ffe-409b-a49a-d88c85898260-kube-api-access-mpwst\") pod \"machine-config-daemon-zqjg9\" (UID: \"33f3d922-4ffe-409b-a49a-d88c85898260\") " pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.495359 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/6b9588d1-5876-4c2c-8de5-e5c0ad46f04e-host-run-multus-certs\") pod \"multus-x94nb\" (UID: \"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\") " pod="openshift-multus/multus-x94nb" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.495372 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-systemd-units\") pod \"ovnkube-node-htkvk\" (UID: \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.495386 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/3a8d2cac-2d0f-4894-9bc0-2abd81699c47-tuning-conf-dir\") pod \"multus-additional-cni-plugins-gjtnh\" (UID: \"3a8d2cac-2d0f-4894-9bc0-2abd81699c47\") " pod="openshift-multus/multus-additional-cni-plugins-gjtnh" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.495402 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-htkvk\" (UID: \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.495424 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/6b9588d1-5876-4c2c-8de5-e5c0ad46f04e-multus-daemon-config\") pod \"multus-x94nb\" (UID: \"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\") " pod="openshift-multus/multus-x94nb" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.495438 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/26b1108a-4b98-4e39-a3cd-e0d055089fd8-env-overrides\") pod \"ovnkube-node-htkvk\" (UID: \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.495452 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/33f3d922-4ffe-409b-a49a-d88c85898260-rootfs\") pod \"machine-config-daemon-zqjg9\" (UID: \"33f3d922-4ffe-409b-a49a-d88c85898260\") " pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.495467 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xgr8g\" (UniqueName: \"kubernetes.io/projected/6b9588d1-5876-4c2c-8de5-e5c0ad46f04e-kube-api-access-xgr8g\") pod \"multus-x94nb\" (UID: \"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\") " pod="openshift-multus/multus-x94nb" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.495490 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-run-ovn\") pod \"ovnkube-node-htkvk\" (UID: \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.495521 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/6b9588d1-5876-4c2c-8de5-e5c0ad46f04e-multus-socket-dir-parent\") pod \"multus-x94nb\" (UID: \"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\") " pod="openshift-multus/multus-x94nb" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.495534 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-log-socket\") pod \"ovnkube-node-htkvk\" (UID: \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.495546 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-host-cni-netd\") pod \"ovnkube-node-htkvk\" (UID: \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.495572 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/26b1108a-4b98-4e39-a3cd-e0d055089fd8-ovnkube-script-lib\") pod \"ovnkube-node-htkvk\" (UID: \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.495586 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-var-lib-openvswitch\") pod \"ovnkube-node-htkvk\" (UID: \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.495601 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-run-openvswitch\") pod \"ovnkube-node-htkvk\" (UID: \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.495614 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/26b1108a-4b98-4e39-a3cd-e0d055089fd8-ovn-node-metrics-cert\") pod \"ovnkube-node-htkvk\" (UID: \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.495627 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/6b9588d1-5876-4c2c-8de5-e5c0ad46f04e-system-cni-dir\") pod \"multus-x94nb\" (UID: \"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\") " pod="openshift-multus/multus-x94nb" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.495640 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/6b9588d1-5876-4c2c-8de5-e5c0ad46f04e-os-release\") pod \"multus-x94nb\" (UID: \"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\") " pod="openshift-multus/multus-x94nb" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.495673 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/6b9588d1-5876-4c2c-8de5-e5c0ad46f04e-hostroot\") pod \"multus-x94nb\" (UID: \"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\") " pod="openshift-multus/multus-x94nb" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.495686 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-host-kubelet\") pod \"ovnkube-node-htkvk\" (UID: \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.495701 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/3a8d2cac-2d0f-4894-9bc0-2abd81699c47-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-gjtnh\" (UID: \"3a8d2cac-2d0f-4894-9bc0-2abd81699c47\") " pod="openshift-multus/multus-additional-cni-plugins-gjtnh" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.495715 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/3a8d2cac-2d0f-4894-9bc0-2abd81699c47-cnibin\") pod \"multus-additional-cni-plugins-gjtnh\" (UID: \"3a8d2cac-2d0f-4894-9bc0-2abd81699c47\") " pod="openshift-multus/multus-additional-cni-plugins-gjtnh" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.495732 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/6b9588d1-5876-4c2c-8de5-e5c0ad46f04e-multus-cni-dir\") pod \"multus-x94nb\" (UID: \"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\") " pod="openshift-multus/multus-x94nb" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.495757 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/6b9588d1-5876-4c2c-8de5-e5c0ad46f04e-host-run-k8s-cni-cncf-io\") pod \"multus-x94nb\" (UID: \"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\") " pod="openshift-multus/multus-x94nb" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.495800 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-node-log\") pod \"ovnkube-node-htkvk\" (UID: \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.495818 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zcf8q\" (UniqueName: \"kubernetes.io/projected/26b1108a-4b98-4e39-a3cd-e0d055089fd8-kube-api-access-zcf8q\") pod \"ovnkube-node-htkvk\" (UID: \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.495879 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q78mk\" (UniqueName: \"kubernetes.io/projected/3a8d2cac-2d0f-4894-9bc0-2abd81699c47-kube-api-access-q78mk\") pod \"multus-additional-cni-plugins-gjtnh\" (UID: \"3a8d2cac-2d0f-4894-9bc0-2abd81699c47\") " pod="openshift-multus/multus-additional-cni-plugins-gjtnh" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.495900 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/33f3d922-4ffe-409b-a49a-d88c85898260-mcd-auth-proxy-config\") pod \"machine-config-daemon-zqjg9\" (UID: \"33f3d922-4ffe-409b-a49a-d88c85898260\") " pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.496113 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/6b9588d1-5876-4c2c-8de5-e5c0ad46f04e-host-var-lib-cni-multus\") pod \"multus-x94nb\" (UID: \"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\") " pod="openshift-multus/multus-x94nb" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.496132 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/6b9588d1-5876-4c2c-8de5-e5c0ad46f04e-host-var-lib-kubelet\") pod \"multus-x94nb\" (UID: \"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\") " pod="openshift-multus/multus-x94nb" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.496146 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-host-cni-bin\") pod \"ovnkube-node-htkvk\" (UID: \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.496159 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/6b9588d1-5876-4c2c-8de5-e5c0ad46f04e-cni-binary-copy\") pod \"multus-x94nb\" (UID: \"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\") " pod="openshift-multus/multus-x94nb" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.496173 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/3a8d2cac-2d0f-4894-9bc0-2abd81699c47-os-release\") pod \"multus-additional-cni-plugins-gjtnh\" (UID: \"3a8d2cac-2d0f-4894-9bc0-2abd81699c47\") " pod="openshift-multus/multus-additional-cni-plugins-gjtnh" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.496191 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/3a8d2cac-2d0f-4894-9bc0-2abd81699c47-cni-binary-copy\") pod \"multus-additional-cni-plugins-gjtnh\" (UID: \"3a8d2cac-2d0f-4894-9bc0-2abd81699c47\") " pod="openshift-multus/multus-additional-cni-plugins-gjtnh" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.496219 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-etc-openvswitch\") pod \"ovnkube-node-htkvk\" (UID: \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.496313 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-host-slash\") pod \"ovnkube-node-htkvk\" (UID: \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.496608 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/3a8d2cac-2d0f-4894-9bc0-2abd81699c47-system-cni-dir\") pod \"multus-additional-cni-plugins-gjtnh\" (UID: \"3a8d2cac-2d0f-4894-9bc0-2abd81699c47\") " pod="openshift-multus/multus-additional-cni-plugins-gjtnh" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.502369 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.510051 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.516736 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9bb99c0a-683b-44da-9c0e-e8d284e87d53\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05e076febd697ed802ce5c7ee92a8c2c636a1a2c6609ad7eb01f17dace247c4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43529d1b9e1e5e22b0d56b023946119da8aacbae54b68c26ff0a1a82dc890c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ea232ad87393685aa323ce6393ff3869f8f8d423c2406a501c69c355535adfa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc6c7eff9e17b8a8ecc88bbfe703b5a01dbf1e02d342cede4f6a263a6526af43\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.523188 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.528504 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lv4rn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1267539-9adf-466f-8027-8ede009d19a0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbmfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lv4rn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.535670 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d31ab61-f94c-4f49-85d0-c9e91b83069b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a66b5c33884af00408dc35cf02f08ebd8570bd695f9f1fe558bf54fd06930431\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7514786602746dabfa690da48d48e233fe68d6dd6674aa5ecd489c0b3185f08\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61b14b6846f94cdeba8f1f61b3754e7663041f92c70356c4ee6cbb988a2e680f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af11ad95bcf0c6bbd5fa4ab99bb592ce1c87f40d29c5ade7832414315820a569\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af11ad95bcf0c6bbd5fa4ab99bb592ce1c87f40d29c5ade7832414315820a569\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW0121 11:11:14.604290 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0121 11:11:14.604572 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0121 11:11:14.605875 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3284452394/tls.crt::/tmp/serving-cert-3284452394/tls.key\\\\\\\"\\\\nI0121 11:11:14.923427 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0121 11:11:14.928141 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0121 11:11:14.928162 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0121 11:11:14.928179 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0121 11:11:14.928184 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0121 11:11:14.933154 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0121 11:11:14.933230 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 11:11:14.933259 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 11:11:14.933282 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0121 11:11:14.933299 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0121 11:11:14.933315 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0121 11:11:14.933331 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0121 11:11:14.933177 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0121 11:11:14.934666 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48070c09fad511108ff22e44211cf26911509b8406d66948985cbfc3009d8233\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.541590 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.548593 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.559271 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"26b1108a-4b98-4e39-a3cd-e0d055089fd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-htkvk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.566867 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d31ab61-f94c-4f49-85d0-c9e91b83069b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a66b5c33884af00408dc35cf02f08ebd8570bd695f9f1fe558bf54fd06930431\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7514786602746dabfa690da48d48e233fe68d6dd6674aa5ecd489c0b3185f08\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61b14b6846f94cdeba8f1f61b3754e7663041f92c70356c4ee6cbb988a2e680f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af11ad95bcf0c6bbd5fa4ab99bb592ce1c87f40d29c5ade7832414315820a569\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af11ad95bcf0c6bbd5fa4ab99bb592ce1c87f40d29c5ade7832414315820a569\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW0121 11:11:14.604290 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0121 11:11:14.604572 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0121 11:11:14.605875 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3284452394/tls.crt::/tmp/serving-cert-3284452394/tls.key\\\\\\\"\\\\nI0121 11:11:14.923427 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0121 11:11:14.928141 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0121 11:11:14.928162 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0121 11:11:14.928179 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0121 11:11:14.928184 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0121 11:11:14.933154 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0121 11:11:14.933230 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 11:11:14.933259 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 11:11:14.933282 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0121 11:11:14.933299 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0121 11:11:14.933315 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0121 11:11:14.933331 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0121 11:11:14.933177 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0121 11:11:14.934666 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48070c09fad511108ff22e44211cf26911509b8406d66948985cbfc3009d8233\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.573843 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9bb99c0a-683b-44da-9c0e-e8d284e87d53\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05e076febd697ed802ce5c7ee92a8c2c636a1a2c6609ad7eb01f17dace247c4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43529d1b9e1e5e22b0d56b023946119da8aacbae54b68c26ff0a1a82dc890c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ea232ad87393685aa323ce6393ff3869f8f8d423c2406a501c69c355535adfa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc6c7eff9e17b8a8ecc88bbfe703b5a01dbf1e02d342cede4f6a263a6526af43\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.580418 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.586352 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lv4rn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1267539-9adf-466f-8027-8ede009d19a0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbmfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lv4rn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.592796 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.597667 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.597751 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xgr8g\" (UniqueName: \"kubernetes.io/projected/6b9588d1-5876-4c2c-8de5-e5c0ad46f04e-kube-api-access-xgr8g\") pod \"multus-x94nb\" (UID: \"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\") " pod="openshift-multus/multus-x94nb" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.597771 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-run-ovn\") pod \"ovnkube-node-htkvk\" (UID: \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" Jan 21 11:11:15 crc kubenswrapper[4824]: E0121 11:11:15.597784 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-21 11:11:16.597768627 +0000 UTC m=+18.890797919 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.597804 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-log-socket\") pod \"ovnkube-node-htkvk\" (UID: \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.597811 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-run-ovn\") pod \"ovnkube-node-htkvk\" (UID: \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.597825 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-host-cni-netd\") pod \"ovnkube-node-htkvk\" (UID: \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.597840 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/26b1108a-4b98-4e39-a3cd-e0d055089fd8-ovnkube-script-lib\") pod \"ovnkube-node-htkvk\" (UID: \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.597861 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/6b9588d1-5876-4c2c-8de5-e5c0ad46f04e-multus-socket-dir-parent\") pod \"multus-x94nb\" (UID: \"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\") " pod="openshift-multus/multus-x94nb" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.597866 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-log-socket\") pod \"ovnkube-node-htkvk\" (UID: \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.597883 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-host-cni-netd\") pod \"ovnkube-node-htkvk\" (UID: \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.597875 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/6b9588d1-5876-4c2c-8de5-e5c0ad46f04e-os-release\") pod \"multus-x94nb\" (UID: \"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\") " pod="openshift-multus/multus-x94nb" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.597921 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/6b9588d1-5876-4c2c-8de5-e5c0ad46f04e-os-release\") pod \"multus-x94nb\" (UID: \"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\") " pod="openshift-multus/multus-x94nb" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.597977 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/6b9588d1-5876-4c2c-8de5-e5c0ad46f04e-hostroot\") pod \"multus-x94nb\" (UID: \"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\") " pod="openshift-multus/multus-x94nb" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.598011 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/6b9588d1-5876-4c2c-8de5-e5c0ad46f04e-multus-socket-dir-parent\") pod \"multus-x94nb\" (UID: \"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\") " pod="openshift-multus/multus-x94nb" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.598016 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-host-kubelet\") pod \"ovnkube-node-htkvk\" (UID: \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.598034 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-var-lib-openvswitch\") pod \"ovnkube-node-htkvk\" (UID: \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.598051 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-run-openvswitch\") pod \"ovnkube-node-htkvk\" (UID: \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.598038 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/6b9588d1-5876-4c2c-8de5-e5c0ad46f04e-hostroot\") pod \"multus-x94nb\" (UID: \"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\") " pod="openshift-multus/multus-x94nb" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.598065 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/26b1108a-4b98-4e39-a3cd-e0d055089fd8-ovn-node-metrics-cert\") pod \"ovnkube-node-htkvk\" (UID: \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.598088 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/6b9588d1-5876-4c2c-8de5-e5c0ad46f04e-system-cni-dir\") pod \"multus-x94nb\" (UID: \"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\") " pod="openshift-multus/multus-x94nb" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.598090 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-var-lib-openvswitch\") pod \"ovnkube-node-htkvk\" (UID: \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.598103 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/3a8d2cac-2d0f-4894-9bc0-2abd81699c47-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-gjtnh\" (UID: \"3a8d2cac-2d0f-4894-9bc0-2abd81699c47\") " pod="openshift-multus/multus-additional-cni-plugins-gjtnh" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.598105 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-host-kubelet\") pod \"ovnkube-node-htkvk\" (UID: \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.598133 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/6b9588d1-5876-4c2c-8de5-e5c0ad46f04e-system-cni-dir\") pod \"multus-x94nb\" (UID: \"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\") " pod="openshift-multus/multus-x94nb" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.598137 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/6b9588d1-5876-4c2c-8de5-e5c0ad46f04e-host-run-k8s-cni-cncf-io\") pod \"multus-x94nb\" (UID: \"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\") " pod="openshift-multus/multus-x94nb" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.598118 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/6b9588d1-5876-4c2c-8de5-e5c0ad46f04e-host-run-k8s-cni-cncf-io\") pod \"multus-x94nb\" (UID: \"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\") " pod="openshift-multus/multus-x94nb" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.598158 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-run-openvswitch\") pod \"ovnkube-node-htkvk\" (UID: \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.598195 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-node-log\") pod \"ovnkube-node-htkvk\" (UID: \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.598293 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-node-log\") pod \"ovnkube-node-htkvk\" (UID: \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.598300 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zcf8q\" (UniqueName: \"kubernetes.io/projected/26b1108a-4b98-4e39-a3cd-e0d055089fd8-kube-api-access-zcf8q\") pod \"ovnkube-node-htkvk\" (UID: \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.598321 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/3a8d2cac-2d0f-4894-9bc0-2abd81699c47-cnibin\") pod \"multus-additional-cni-plugins-gjtnh\" (UID: \"3a8d2cac-2d0f-4894-9bc0-2abd81699c47\") " pod="openshift-multus/multus-additional-cni-plugins-gjtnh" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.598339 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/6b9588d1-5876-4c2c-8de5-e5c0ad46f04e-multus-cni-dir\") pod \"multus-x94nb\" (UID: \"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\") " pod="openshift-multus/multus-x94nb" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.598354 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/6b9588d1-5876-4c2c-8de5-e5c0ad46f04e-host-var-lib-cni-multus\") pod \"multus-x94nb\" (UID: \"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\") " pod="openshift-multus/multus-x94nb" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.598369 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/6b9588d1-5876-4c2c-8de5-e5c0ad46f04e-host-var-lib-kubelet\") pod \"multus-x94nb\" (UID: \"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\") " pod="openshift-multus/multus-x94nb" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.598385 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-host-cni-bin\") pod \"ovnkube-node-htkvk\" (UID: \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.598398 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/6b9588d1-5876-4c2c-8de5-e5c0ad46f04e-host-var-lib-cni-multus\") pod \"multus-x94nb\" (UID: \"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\") " pod="openshift-multus/multus-x94nb" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.598402 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q78mk\" (UniqueName: \"kubernetes.io/projected/3a8d2cac-2d0f-4894-9bc0-2abd81699c47-kube-api-access-q78mk\") pod \"multus-additional-cni-plugins-gjtnh\" (UID: \"3a8d2cac-2d0f-4894-9bc0-2abd81699c47\") " pod="openshift-multus/multus-additional-cni-plugins-gjtnh" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.598423 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/33f3d922-4ffe-409b-a49a-d88c85898260-mcd-auth-proxy-config\") pod \"machine-config-daemon-zqjg9\" (UID: \"33f3d922-4ffe-409b-a49a-d88c85898260\") " pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.598426 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/6b9588d1-5876-4c2c-8de5-e5c0ad46f04e-host-var-lib-kubelet\") pod \"multus-x94nb\" (UID: \"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\") " pod="openshift-multus/multus-x94nb" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.598440 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/6b9588d1-5876-4c2c-8de5-e5c0ad46f04e-cni-binary-copy\") pod \"multus-x94nb\" (UID: \"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\") " pod="openshift-multus/multus-x94nb" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.598455 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/3a8d2cac-2d0f-4894-9bc0-2abd81699c47-os-release\") pod \"multus-additional-cni-plugins-gjtnh\" (UID: \"3a8d2cac-2d0f-4894-9bc0-2abd81699c47\") " pod="openshift-multus/multus-additional-cni-plugins-gjtnh" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.598472 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/3a8d2cac-2d0f-4894-9bc0-2abd81699c47-cni-binary-copy\") pod \"multus-additional-cni-plugins-gjtnh\" (UID: \"3a8d2cac-2d0f-4894-9bc0-2abd81699c47\") " pod="openshift-multus/multus-additional-cni-plugins-gjtnh" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.598476 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/6b9588d1-5876-4c2c-8de5-e5c0ad46f04e-multus-cni-dir\") pod \"multus-x94nb\" (UID: \"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\") " pod="openshift-multus/multus-x94nb" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.598495 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.598510 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-host-cni-bin\") pod \"ovnkube-node-htkvk\" (UID: \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.598512 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-etc-openvswitch\") pod \"ovnkube-node-htkvk\" (UID: \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.598531 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-etc-openvswitch\") pod \"ovnkube-node-htkvk\" (UID: \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.598540 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.598552 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/3a8d2cac-2d0f-4894-9bc0-2abd81699c47-os-release\") pod \"multus-additional-cni-plugins-gjtnh\" (UID: \"3a8d2cac-2d0f-4894-9bc0-2abd81699c47\") " pod="openshift-multus/multus-additional-cni-plugins-gjtnh" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.598452 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/26b1108a-4b98-4e39-a3cd-e0d055089fd8-ovnkube-script-lib\") pod \"ovnkube-node-htkvk\" (UID: \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" Jan 21 11:11:15 crc kubenswrapper[4824]: E0121 11:11:15.598587 4824 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.598425 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/3a8d2cac-2d0f-4894-9bc0-2abd81699c47-cnibin\") pod \"multus-additional-cni-plugins-gjtnh\" (UID: \"3a8d2cac-2d0f-4894-9bc0-2abd81699c47\") " pod="openshift-multus/multus-additional-cni-plugins-gjtnh" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.598556 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-host-slash\") pod \"ovnkube-node-htkvk\" (UID: \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" Jan 21 11:11:15 crc kubenswrapper[4824]: E0121 11:11:15.598599 4824 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.598573 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-host-slash\") pod \"ovnkube-node-htkvk\" (UID: \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" Jan 21 11:11:15 crc kubenswrapper[4824]: E0121 11:11:15.598635 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-21 11:11:16.598623977 +0000 UTC m=+18.891653269 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.598651 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/3a8d2cac-2d0f-4894-9bc0-2abd81699c47-system-cni-dir\") pod \"multus-additional-cni-plugins-gjtnh\" (UID: \"3a8d2cac-2d0f-4894-9bc0-2abd81699c47\") " pod="openshift-multus/multus-additional-cni-plugins-gjtnh" Jan 21 11:11:15 crc kubenswrapper[4824]: E0121 11:11:15.598667 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-21 11:11:16.598653643 +0000 UTC m=+18.891682934 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.598673 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/3a8d2cac-2d0f-4894-9bc0-2abd81699c47-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-gjtnh\" (UID: \"3a8d2cac-2d0f-4894-9bc0-2abd81699c47\") " pod="openshift-multus/multus-additional-cni-plugins-gjtnh" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.598678 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/3a8d2cac-2d0f-4894-9bc0-2abd81699c47-system-cni-dir\") pod \"multus-additional-cni-plugins-gjtnh\" (UID: \"3a8d2cac-2d0f-4894-9bc0-2abd81699c47\") " pod="openshift-multus/multus-additional-cni-plugins-gjtnh" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.598681 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/6b9588d1-5876-4c2c-8de5-e5c0ad46f04e-etc-kubernetes\") pod \"multus-x94nb\" (UID: \"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\") " pod="openshift-multus/multus-x94nb" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.598717 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/6b9588d1-5876-4c2c-8de5-e5c0ad46f04e-etc-kubernetes\") pod \"multus-x94nb\" (UID: \"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\") " pod="openshift-multus/multus-x94nb" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.598724 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-host-run-ovn-kubernetes\") pod \"ovnkube-node-htkvk\" (UID: \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.598740 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-host-run-ovn-kubernetes\") pod \"ovnkube-node-htkvk\" (UID: \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.598740 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/6b9588d1-5876-4c2c-8de5-e5c0ad46f04e-host-var-lib-cni-bin\") pod \"multus-x94nb\" (UID: \"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\") " pod="openshift-multus/multus-x94nb" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.598757 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/6b9588d1-5876-4c2c-8de5-e5c0ad46f04e-host-var-lib-cni-bin\") pod \"multus-x94nb\" (UID: \"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\") " pod="openshift-multus/multus-x94nb" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.598788 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/6b9588d1-5876-4c2c-8de5-e5c0ad46f04e-cnibin\") pod \"multus-x94nb\" (UID: \"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\") " pod="openshift-multus/multus-x94nb" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.598802 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-run-systemd\") pod \"ovnkube-node-htkvk\" (UID: \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.598826 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/6b9588d1-5876-4c2c-8de5-e5c0ad46f04e-multus-conf-dir\") pod \"multus-x94nb\" (UID: \"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\") " pod="openshift-multus/multus-x94nb" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.598853 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/6b9588d1-5876-4c2c-8de5-e5c0ad46f04e-cnibin\") pod \"multus-x94nb\" (UID: \"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\") " pod="openshift-multus/multus-x94nb" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.598843 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-host-run-netns\") pod \"ovnkube-node-htkvk\" (UID: \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.598867 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-run-systemd\") pod \"ovnkube-node-htkvk\" (UID: \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.598875 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/26b1108a-4b98-4e39-a3cd-e0d055089fd8-ovnkube-config\") pod \"ovnkube-node-htkvk\" (UID: \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.598891 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-host-run-netns\") pod \"ovnkube-node-htkvk\" (UID: \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.598897 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/6b9588d1-5876-4c2c-8de5-e5c0ad46f04e-host-run-netns\") pod \"multus-x94nb\" (UID: \"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\") " pod="openshift-multus/multus-x94nb" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.598876 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/6b9588d1-5876-4c2c-8de5-e5c0ad46f04e-multus-conf-dir\") pod \"multus-x94nb\" (UID: \"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\") " pod="openshift-multus/multus-x94nb" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.598913 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/6b9588d1-5876-4c2c-8de5-e5c0ad46f04e-host-run-multus-certs\") pod \"multus-x94nb\" (UID: \"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\") " pod="openshift-multus/multus-x94nb" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.598914 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/33f3d922-4ffe-409b-a49a-d88c85898260-mcd-auth-proxy-config\") pod \"machine-config-daemon-zqjg9\" (UID: \"33f3d922-4ffe-409b-a49a-d88c85898260\") " pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.598931 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/6b9588d1-5876-4c2c-8de5-e5c0ad46f04e-host-run-multus-certs\") pod \"multus-x94nb\" (UID: \"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\") " pod="openshift-multus/multus-x94nb" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.598919 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/6b9588d1-5876-4c2c-8de5-e5c0ad46f04e-host-run-netns\") pod \"multus-x94nb\" (UID: \"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\") " pod="openshift-multus/multus-x94nb" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.598940 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-systemd-units\") pod \"ovnkube-node-htkvk\" (UID: \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.599007 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-systemd-units\") pod \"ovnkube-node-htkvk\" (UID: \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.599005 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/3a8d2cac-2d0f-4894-9bc0-2abd81699c47-tuning-conf-dir\") pod \"multus-additional-cni-plugins-gjtnh\" (UID: \"3a8d2cac-2d0f-4894-9bc0-2abd81699c47\") " pod="openshift-multus/multus-additional-cni-plugins-gjtnh" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.599039 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/33f3d922-4ffe-409b-a49a-d88c85898260-proxy-tls\") pod \"machine-config-daemon-zqjg9\" (UID: \"33f3d922-4ffe-409b-a49a-d88c85898260\") " pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.599054 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mpwst\" (UniqueName: \"kubernetes.io/projected/33f3d922-4ffe-409b-a49a-d88c85898260-kube-api-access-mpwst\") pod \"machine-config-daemon-zqjg9\" (UID: \"33f3d922-4ffe-409b-a49a-d88c85898260\") " pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.599089 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-htkvk\" (UID: \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.599104 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/6b9588d1-5876-4c2c-8de5-e5c0ad46f04e-multus-daemon-config\") pod \"multus-x94nb\" (UID: \"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\") " pod="openshift-multus/multus-x94nb" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.599119 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/26b1108a-4b98-4e39-a3cd-e0d055089fd8-env-overrides\") pod \"ovnkube-node-htkvk\" (UID: \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.599132 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/33f3d922-4ffe-409b-a49a-d88c85898260-rootfs\") pod \"machine-config-daemon-zqjg9\" (UID: \"33f3d922-4ffe-409b-a49a-d88c85898260\") " pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.599184 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/33f3d922-4ffe-409b-a49a-d88c85898260-rootfs\") pod \"machine-config-daemon-zqjg9\" (UID: \"33f3d922-4ffe-409b-a49a-d88c85898260\") " pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.599338 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/26b1108a-4b98-4e39-a3cd-e0d055089fd8-ovnkube-config\") pod \"ovnkube-node-htkvk\" (UID: \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.599378 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-htkvk\" (UID: \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.599453 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/3a8d2cac-2d0f-4894-9bc0-2abd81699c47-tuning-conf-dir\") pod \"multus-additional-cni-plugins-gjtnh\" (UID: \"3a8d2cac-2d0f-4894-9bc0-2abd81699c47\") " pod="openshift-multus/multus-additional-cni-plugins-gjtnh" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.599771 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/26b1108a-4b98-4e39-a3cd-e0d055089fd8-env-overrides\") pod \"ovnkube-node-htkvk\" (UID: \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.599895 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/6b9588d1-5876-4c2c-8de5-e5c0ad46f04e-multus-daemon-config\") pod \"multus-x94nb\" (UID: \"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\") " pod="openshift-multus/multus-x94nb" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.601112 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/26b1108a-4b98-4e39-a3cd-e0d055089fd8-ovn-node-metrics-cert\") pod \"ovnkube-node-htkvk\" (UID: \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.601976 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/33f3d922-4ffe-409b-a49a-d88c85898260-proxy-tls\") pod \"machine-config-daemon-zqjg9\" (UID: \"33f3d922-4ffe-409b-a49a-d88c85898260\") " pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.621564 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.627701 4824 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2027-01-21 11:06:14 +0000 UTC, rotation deadline is 2026-10-25 14:23:32.222228902 +0000 UTC Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.627745 4824 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 6651h12m16.594486s for next certificate rotation Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.689191 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zcf8q\" (UniqueName: \"kubernetes.io/projected/26b1108a-4b98-4e39-a3cd-e0d055089fd8-kube-api-access-zcf8q\") pod \"ovnkube-node-htkvk\" (UID: \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\") " pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.699500 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.699575 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 11:11:15 crc kubenswrapper[4824]: E0121 11:11:15.699657 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 21 11:11:15 crc kubenswrapper[4824]: E0121 11:11:15.699677 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 21 11:11:15 crc kubenswrapper[4824]: E0121 11:11:15.699687 4824 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 21 11:11:15 crc kubenswrapper[4824]: E0121 11:11:15.699720 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-01-21 11:11:16.699709197 +0000 UTC m=+18.992738489 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 21 11:11:15 crc kubenswrapper[4824]: E0121 11:11:15.699657 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 21 11:11:15 crc kubenswrapper[4824]: E0121 11:11:15.699768 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 21 11:11:15 crc kubenswrapper[4824]: E0121 11:11:15.699778 4824 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 21 11:11:15 crc kubenswrapper[4824]: E0121 11:11:15.699812 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-01-21 11:11:16.699803837 +0000 UTC m=+18.992833128 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.740782 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.783909 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x94nb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgr8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x94nb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:15Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.805773 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.815008 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.823721 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33f3d922-4ffe-409b-a49a-d88c85898260\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpwst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpwst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zqjg9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:15Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.844567 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.868818 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.887106 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gjtnh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a8d2cac-2d0f-4894-9bc0-2abd81699c47\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gjtnh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:15Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:15 crc kubenswrapper[4824]: W0121 11:11:15.896977 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod26b1108a_4b98_4e39_a3cd_e0d055089fd8.slice/crio-b233a1680204a0985feed02a396e93f06104160dd5b2e0d82128d39c672a1129 WatchSource:0}: Error finding container b233a1680204a0985feed02a396e93f06104160dd5b2e0d82128d39c672a1129: Status 404 returned error can't find the container with id b233a1680204a0985feed02a396e93f06104160dd5b2e0d82128d39c672a1129 Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.922793 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:15Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.963418 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:15Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:15 crc kubenswrapper[4824]: I0121 11:11:15.989995 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-15 03:25:33.709898958 +0000 UTC Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.006157 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"26b1108a-4b98-4e39-a3cd-e0d055089fd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-htkvk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:16Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.043506 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d31ab61-f94c-4f49-85d0-c9e91b83069b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a66b5c33884af00408dc35cf02f08ebd8570bd695f9f1fe558bf54fd06930431\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7514786602746dabfa690da48d48e233fe68d6dd6674aa5ecd489c0b3185f08\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61b14b6846f94cdeba8f1f61b3754e7663041f92c70356c4ee6cbb988a2e680f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af11ad95bcf0c6bbd5fa4ab99bb592ce1c87f40d29c5ade7832414315820a569\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af11ad95bcf0c6bbd5fa4ab99bb592ce1c87f40d29c5ade7832414315820a569\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW0121 11:11:14.604290 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0121 11:11:14.604572 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0121 11:11:14.605875 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3284452394/tls.crt::/tmp/serving-cert-3284452394/tls.key\\\\\\\"\\\\nI0121 11:11:14.923427 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0121 11:11:14.928141 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0121 11:11:14.928162 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0121 11:11:14.928179 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0121 11:11:14.928184 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0121 11:11:14.933154 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0121 11:11:14.933230 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 11:11:14.933259 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 11:11:14.933282 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0121 11:11:14.933299 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0121 11:11:14.933315 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0121 11:11:14.933331 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0121 11:11:14.933177 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0121 11:11:14.934666 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48070c09fad511108ff22e44211cf26911509b8406d66948985cbfc3009d8233\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:16Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.049047 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 11:11:16 crc kubenswrapper[4824]: E0121 11:11:16.049140 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.051906 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.052539 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.053558 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.054155 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.055048 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.055518 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.056033 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.056916 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.057501 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.058345 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.058767 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.059689 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.060139 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.060700 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.061652 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.062388 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.063228 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.063555 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.064342 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.065446 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.065931 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.066793 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.067219 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.068343 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.068710 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.069266 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.070193 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.070655 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.071681 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.072362 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.073132 4824 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.073232 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.074741 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.075556 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.076019 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.077319 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.077866 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.078636 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.079391 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.080462 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.080911 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.082343 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.082574 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9bb99c0a-683b-44da-9c0e-e8d284e87d53\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05e076febd697ed802ce5c7ee92a8c2c636a1a2c6609ad7eb01f17dace247c4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43529d1b9e1e5e22b0d56b023946119da8aacbae54b68c26ff0a1a82dc890c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ea232ad87393685aa323ce6393ff3869f8f8d423c2406a501c69c355535adfa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc6c7eff9e17b8a8ecc88bbfe703b5a01dbf1e02d342cede4f6a263a6526af43\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:16Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.083504 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.084842 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.085688 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.086626 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.087346 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.088531 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.088982 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.089737 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.090175 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.091122 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.091729 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.092348 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.120806 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"516c009d4720ab5b393c7482872f12fed22d4f4414bd1f3400411c0e8da88ce4"} Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.120835 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"ee18d293036971832c8d9c3f7b9eca84ab779ff6768a2db3c6f7bdcf3433cdef"} Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.120846 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"661c876c65ddef7d3211b181fa8432b45a9d8715f2d2bed40679e29188b710d9"} Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.122448 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"86d3005cd61c0973659e1c1c0d3ef1ddd6fcf95bede5465f4f2099e64a74bb13"} Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.122472 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"f6a34848b7ccc79f1145a419de2da50011e2e9a8aecedf718fec38ac701b9547"} Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.123835 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:16Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.123986 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.125124 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"3a7023cfe4c098bcdd9c0477a8187f4ed7a0d2abdbe1e143436863948f441b4f"} Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.125254 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.125901 4824 generic.go:334] "Generic (PLEG): container finished" podID="26b1108a-4b98-4e39-a3cd-e0d055089fd8" containerID="82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3" exitCode=0 Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.125950 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" event={"ID":"26b1108a-4b98-4e39-a3cd-e0d055089fd8","Type":"ContainerDied","Data":"82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3"} Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.125986 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" event={"ID":"26b1108a-4b98-4e39-a3cd-e0d055089fd8","Type":"ContainerStarted","Data":"b233a1680204a0985feed02a396e93f06104160dd5b2e0d82128d39c672a1129"} Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.127194 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"5ae7529fc08fbc1bc45653348ce3eaf2bb6e0c95ad21544f7ef8f6655227058d"} Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.128319 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-lv4rn" event={"ID":"c1267539-9adf-466f-8027-8ede009d19a0","Type":"ContainerStarted","Data":"eed1b884aa53e1ee318169a0cc2c9ad10130aa363ae4ac55e4bf065237756f53"} Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.128347 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-lv4rn" event={"ID":"c1267539-9adf-466f-8027-8ede009d19a0","Type":"ContainerStarted","Data":"44a63cef2a9562f6b836d0eea1fdc74e6cf08c709d90570463cdcccabafa476c"} Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.162840 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lv4rn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1267539-9adf-466f-8027-8ede009d19a0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbmfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lv4rn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:16Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.209570 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06f56d86-ec6f-4b57-b314-f508c0dd4d9f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d25c94f1ae0bbabb733f0e1ea1d1ba39562a92af73047a39aacc03013e239681\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca06234486b59cf990c69d9501e23f935328e2ba21013cd71ed64831ea749bcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b63e45d66bcd11ff3479cc3ec6b450017da8d21d36e967b8d496ae0f24b0ee9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52758bcd31f2ba64edf03adfd4d1be6428015abff5b64c7c029fa787085f9c29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2736e73342d2adb175b0721dd4f40cc42d26fca06d5eb3334927518531c3b107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b35ad7dd0570c191eb0080739304f3cd43f17b3d1dddf895549441ebe50deaa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b35ad7dd0570c191eb0080739304f3cd43f17b3d1dddf895549441ebe50deaa6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://80cd5ea5e6fc19a88c1aab5c61f827f103ab798a5c84da8dc4d45cea86d5aacb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80cd5ea5e6fc19a88c1aab5c61f827f103ab798a5c84da8dc4d45cea86d5aacb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://234d8b914e6a88480021647ae65d74e9b21c55b6788312a094d00c8a42677d48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://234d8b914e6a88480021647ae65d74e9b21c55b6788312a094d00c8a42677d48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:16Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.242985 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:16Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.284843 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:16Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.326431 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:16Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.365751 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x94nb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgr8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x94nb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:16Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.377733 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.423143 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33f3d922-4ffe-409b-a49a-d88c85898260\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpwst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpwst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zqjg9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:16Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.466374 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gjtnh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a8d2cac-2d0f-4894-9bc0-2abd81699c47\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gjtnh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:16Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.505255 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:16Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.517745 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.529536 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xgr8g\" (UniqueName: \"kubernetes.io/projected/6b9588d1-5876-4c2c-8de5-e5c0ad46f04e-kube-api-access-xgr8g\") pod \"multus-x94nb\" (UID: \"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\") " pod="openshift-multus/multus-x94nb" Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.530780 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q78mk\" (UniqueName: \"kubernetes.io/projected/3a8d2cac-2d0f-4894-9bc0-2abd81699c47-kube-api-access-q78mk\") pod \"multus-additional-cni-plugins-gjtnh\" (UID: \"3a8d2cac-2d0f-4894-9bc0-2abd81699c47\") " pod="openshift-multus/multus-additional-cni-plugins-gjtnh" Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.567644 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06f56d86-ec6f-4b57-b314-f508c0dd4d9f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d25c94f1ae0bbabb733f0e1ea1d1ba39562a92af73047a39aacc03013e239681\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca06234486b59cf990c69d9501e23f935328e2ba21013cd71ed64831ea749bcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b63e45d66bcd11ff3479cc3ec6b450017da8d21d36e967b8d496ae0f24b0ee9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52758bcd31f2ba64edf03adfd4d1be6428015abff5b64c7c029fa787085f9c29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2736e73342d2adb175b0721dd4f40cc42d26fca06d5eb3334927518531c3b107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b35ad7dd0570c191eb0080739304f3cd43f17b3d1dddf895549441ebe50deaa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b35ad7dd0570c191eb0080739304f3cd43f17b3d1dddf895549441ebe50deaa6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://80cd5ea5e6fc19a88c1aab5c61f827f103ab798a5c84da8dc4d45cea86d5aacb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80cd5ea5e6fc19a88c1aab5c61f827f103ab798a5c84da8dc4d45cea86d5aacb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://234d8b914e6a88480021647ae65d74e9b21c55b6788312a094d00c8a42677d48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://234d8b914e6a88480021647ae65d74e9b21c55b6788312a094d00c8a42677d48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:16Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:16 crc kubenswrapper[4824]: E0121 11:11:16.599023 4824 configmap.go:193] Couldn't get configMap openshift-multus/cni-copy-resources: failed to sync configmap cache: timed out waiting for the condition Jan 21 11:11:16 crc kubenswrapper[4824]: E0121 11:11:16.599078 4824 configmap.go:193] Couldn't get configMap openshift-multus/cni-copy-resources: failed to sync configmap cache: timed out waiting for the condition Jan 21 11:11:16 crc kubenswrapper[4824]: E0121 11:11:16.599083 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/6b9588d1-5876-4c2c-8de5-e5c0ad46f04e-cni-binary-copy podName:6b9588d1-5876-4c2c-8de5-e5c0ad46f04e nodeName:}" failed. No retries permitted until 2026-01-21 11:11:17.099068855 +0000 UTC m=+19.392098146 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cni-binary-copy" (UniqueName: "kubernetes.io/configmap/6b9588d1-5876-4c2c-8de5-e5c0ad46f04e-cni-binary-copy") pod "multus-x94nb" (UID: "6b9588d1-5876-4c2c-8de5-e5c0ad46f04e") : failed to sync configmap cache: timed out waiting for the condition Jan 21 11:11:16 crc kubenswrapper[4824]: E0121 11:11:16.599158 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/3a8d2cac-2d0f-4894-9bc0-2abd81699c47-cni-binary-copy podName:3a8d2cac-2d0f-4894-9bc0-2abd81699c47 nodeName:}" failed. No retries permitted until 2026-01-21 11:11:17.099142583 +0000 UTC m=+19.392171876 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cni-binary-copy" (UniqueName: "kubernetes.io/configmap/3a8d2cac-2d0f-4894-9bc0-2abd81699c47-cni-binary-copy") pod "multus-additional-cni-plugins-gjtnh" (UID: "3a8d2cac-2d0f-4894-9bc0-2abd81699c47") : failed to sync configmap cache: timed out waiting for the condition Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.605539 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:16Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.617111 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.617233 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 11:11:16 crc kubenswrapper[4824]: E0121 11:11:16.617249 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-21 11:11:18.617235194 +0000 UTC m=+20.910264486 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.617280 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 11:11:16 crc kubenswrapper[4824]: E0121 11:11:16.617317 4824 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 21 11:11:16 crc kubenswrapper[4824]: E0121 11:11:16.617361 4824 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Jan 21 11:11:16 crc kubenswrapper[4824]: E0121 11:11:16.617369 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-21 11:11:18.617358116 +0000 UTC m=+20.910387409 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 21 11:11:16 crc kubenswrapper[4824]: E0121 11:11:16.617388 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-21 11:11:18.617381381 +0000 UTC m=+20.910410673 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.643073 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x94nb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgr8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x94nb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:16Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.682543 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33f3d922-4ffe-409b-a49a-d88c85898260\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpwst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpwst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zqjg9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:16Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:16 crc kubenswrapper[4824]: E0121 11:11:16.707403 4824 projected.go:288] Couldn't get configMap openshift-machine-config-operator/openshift-service-ca.crt: failed to sync configmap cache: timed out waiting for the condition Jan 21 11:11:16 crc kubenswrapper[4824]: E0121 11:11:16.707436 4824 projected.go:194] Error preparing data for projected volume kube-api-access-mpwst for pod openshift-machine-config-operator/machine-config-daemon-zqjg9: failed to sync configmap cache: timed out waiting for the condition Jan 21 11:11:16 crc kubenswrapper[4824]: E0121 11:11:16.707493 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/33f3d922-4ffe-409b-a49a-d88c85898260-kube-api-access-mpwst podName:33f3d922-4ffe-409b-a49a-d88c85898260 nodeName:}" failed. No retries permitted until 2026-01-21 11:11:17.207477426 +0000 UTC m=+19.500506718 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-mpwst" (UniqueName: "kubernetes.io/projected/33f3d922-4ffe-409b-a49a-d88c85898260-kube-api-access-mpwst") pod "machine-config-daemon-zqjg9" (UID: "33f3d922-4ffe-409b-a49a-d88c85898260") : failed to sync configmap cache: timed out waiting for the condition Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.718400 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.718470 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 11:11:16 crc kubenswrapper[4824]: E0121 11:11:16.718571 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 21 11:11:16 crc kubenswrapper[4824]: E0121 11:11:16.718599 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 21 11:11:16 crc kubenswrapper[4824]: E0121 11:11:16.718610 4824 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 21 11:11:16 crc kubenswrapper[4824]: E0121 11:11:16.718646 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-01-21 11:11:18.71863493 +0000 UTC m=+21.011664221 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 21 11:11:16 crc kubenswrapper[4824]: E0121 11:11:16.718578 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 21 11:11:16 crc kubenswrapper[4824]: E0121 11:11:16.718690 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 21 11:11:16 crc kubenswrapper[4824]: E0121 11:11:16.718701 4824 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 21 11:11:16 crc kubenswrapper[4824]: E0121 11:11:16.718740 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-01-21 11:11:18.718729519 +0000 UTC m=+21.011758811 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.725693 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gjtnh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a8d2cac-2d0f-4894-9bc0-2abd81699c47\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gjtnh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:16Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.763518 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86d3005cd61c0973659e1c1c0d3ef1ddd6fcf95bede5465f4f2099e64a74bb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:16Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.777736 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.823708 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://516c009d4720ab5b393c7482872f12fed22d4f4414bd1f3400411c0e8da88ce4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee18d293036971832c8d9c3f7b9eca84ab779ff6768a2db3c6f7bdcf3433cdef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:16Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.837727 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.886161 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"26b1108a-4b98-4e39-a3cd-e0d055089fd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-htkvk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:16Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.922951 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:16Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.937005 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.983913 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d31ab61-f94c-4f49-85d0-c9e91b83069b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a66b5c33884af00408dc35cf02f08ebd8570bd695f9f1fe558bf54fd06930431\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7514786602746dabfa690da48d48e233fe68d6dd6674aa5ecd489c0b3185f08\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61b14b6846f94cdeba8f1f61b3754e7663041f92c70356c4ee6cbb988a2e680f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a7023cfe4c098bcdd9c0477a8187f4ed7a0d2abdbe1e143436863948f441b4f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af11ad95bcf0c6bbd5fa4ab99bb592ce1c87f40d29c5ade7832414315820a569\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW0121 11:11:14.604290 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0121 11:11:14.604572 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0121 11:11:14.605875 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3284452394/tls.crt::/tmp/serving-cert-3284452394/tls.key\\\\\\\"\\\\nI0121 11:11:14.923427 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0121 11:11:14.928141 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0121 11:11:14.928162 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0121 11:11:14.928179 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0121 11:11:14.928184 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0121 11:11:14.933154 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0121 11:11:14.933230 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 11:11:14.933259 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 11:11:14.933282 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0121 11:11:14.933299 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0121 11:11:14.933315 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0121 11:11:14.933331 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0121 11:11:14.933177 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0121 11:11:14.934666 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48070c09fad511108ff22e44211cf26911509b8406d66948985cbfc3009d8233\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:16Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:16 crc kubenswrapper[4824]: I0121 11:11:16.990970 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-25 09:10:21.198813534 +0000 UTC Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.023985 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9bb99c0a-683b-44da-9c0e-e8d284e87d53\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05e076febd697ed802ce5c7ee92a8c2c636a1a2c6609ad7eb01f17dace247c4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43529d1b9e1e5e22b0d56b023946119da8aacbae54b68c26ff0a1a82dc890c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ea232ad87393685aa323ce6393ff3869f8f8d423c2406a501c69c355535adfa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc6c7eff9e17b8a8ecc88bbfe703b5a01dbf1e02d342cede4f6a263a6526af43\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:17Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.048908 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.048945 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 11:11:17 crc kubenswrapper[4824]: E0121 11:11:17.049022 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 21 11:11:17 crc kubenswrapper[4824]: E0121 11:11:17.049074 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.063077 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:17Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.103465 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lv4rn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1267539-9adf-466f-8027-8ede009d19a0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eed1b884aa53e1ee318169a0cc2c9ad10130aa363ae4ac55e4bf065237756f53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbmfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lv4rn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:17Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.121835 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/6b9588d1-5876-4c2c-8de5-e5c0ad46f04e-cni-binary-copy\") pod \"multus-x94nb\" (UID: \"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\") " pod="openshift-multus/multus-x94nb" Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.121873 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/3a8d2cac-2d0f-4894-9bc0-2abd81699c47-cni-binary-copy\") pod \"multus-additional-cni-plugins-gjtnh\" (UID: \"3a8d2cac-2d0f-4894-9bc0-2abd81699c47\") " pod="openshift-multus/multus-additional-cni-plugins-gjtnh" Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.122463 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/3a8d2cac-2d0f-4894-9bc0-2abd81699c47-cni-binary-copy\") pod \"multus-additional-cni-plugins-gjtnh\" (UID: \"3a8d2cac-2d0f-4894-9bc0-2abd81699c47\") " pod="openshift-multus/multus-additional-cni-plugins-gjtnh" Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.122516 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/6b9588d1-5876-4c2c-8de5-e5c0ad46f04e-cni-binary-copy\") pod \"multus-x94nb\" (UID: \"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\") " pod="openshift-multus/multus-x94nb" Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.131709 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"6d87422473838f5218c9fb646ab2515f2f42592754ffdf07b19f7a8e093fa48a"} Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.134193 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" event={"ID":"26b1108a-4b98-4e39-a3cd-e0d055089fd8","Type":"ContainerStarted","Data":"fc4c96e33e069f7d08d33caf92cd0987668506f931fd134b246d5ce5e934f95c"} Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.134229 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" event={"ID":"26b1108a-4b98-4e39-a3cd-e0d055089fd8","Type":"ContainerStarted","Data":"ddbb0e5ae02e9d5d82a80ead7efc91f1b8e557a9fb6f39940e5530fae8de2d38"} Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.134242 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" event={"ID":"26b1108a-4b98-4e39-a3cd-e0d055089fd8","Type":"ContainerStarted","Data":"cee9cc9eec3fc815f1621a22a37805691288dbb60dd3ff64194a84be62abcddc"} Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.134252 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" event={"ID":"26b1108a-4b98-4e39-a3cd-e0d055089fd8","Type":"ContainerStarted","Data":"fc319bbc06568cc811ded2b418b04b986a31f321fed492e8be50104af79bb30b"} Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.134260 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" event={"ID":"26b1108a-4b98-4e39-a3cd-e0d055089fd8","Type":"ContainerStarted","Data":"64a9065254f889b1c43d63759283335bf78fc4fa1531a4baae472c06c075dae3"} Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.134268 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" event={"ID":"26b1108a-4b98-4e39-a3cd-e0d055089fd8","Type":"ContainerStarted","Data":"5e94ef8a8f47dc75ecfbab2e3fb8e7097c1b73eed32d7ca947f11ea198cea880"} Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.143429 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:17Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.182100 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lv4rn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1267539-9adf-466f-8027-8ede009d19a0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eed1b884aa53e1ee318169a0cc2c9ad10130aa363ae4ac55e4bf065237756f53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbmfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lv4rn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:17Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.222975 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mpwst\" (UniqueName: \"kubernetes.io/projected/33f3d922-4ffe-409b-a49a-d88c85898260-kube-api-access-mpwst\") pod \"machine-config-daemon-zqjg9\" (UID: \"33f3d922-4ffe-409b-a49a-d88c85898260\") " pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.224168 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d31ab61-f94c-4f49-85d0-c9e91b83069b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a66b5c33884af00408dc35cf02f08ebd8570bd695f9f1fe558bf54fd06930431\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7514786602746dabfa690da48d48e233fe68d6dd6674aa5ecd489c0b3185f08\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61b14b6846f94cdeba8f1f61b3754e7663041f92c70356c4ee6cbb988a2e680f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a7023cfe4c098bcdd9c0477a8187f4ed7a0d2abdbe1e143436863948f441b4f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af11ad95bcf0c6bbd5fa4ab99bb592ce1c87f40d29c5ade7832414315820a569\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW0121 11:11:14.604290 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0121 11:11:14.604572 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0121 11:11:14.605875 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3284452394/tls.crt::/tmp/serving-cert-3284452394/tls.key\\\\\\\"\\\\nI0121 11:11:14.923427 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0121 11:11:14.928141 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0121 11:11:14.928162 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0121 11:11:14.928179 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0121 11:11:14.928184 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0121 11:11:14.933154 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0121 11:11:14.933230 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 11:11:14.933259 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 11:11:14.933282 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0121 11:11:14.933299 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0121 11:11:14.933315 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0121 11:11:14.933331 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0121 11:11:14.933177 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0121 11:11:14.934666 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48070c09fad511108ff22e44211cf26911509b8406d66948985cbfc3009d8233\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:17Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.225452 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mpwst\" (UniqueName: \"kubernetes.io/projected/33f3d922-4ffe-409b-a49a-d88c85898260-kube-api-access-mpwst\") pod \"machine-config-daemon-zqjg9\" (UID: \"33f3d922-4ffe-409b-a49a-d88c85898260\") " pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.264285 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9bb99c0a-683b-44da-9c0e-e8d284e87d53\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05e076febd697ed802ce5c7ee92a8c2c636a1a2c6609ad7eb01f17dace247c4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43529d1b9e1e5e22b0d56b023946119da8aacbae54b68c26ff0a1a82dc890c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ea232ad87393685aa323ce6393ff3869f8f8d423c2406a501c69c355535adfa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc6c7eff9e17b8a8ecc88bbfe703b5a01dbf1e02d342cede4f6a263a6526af43\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:17Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.267346 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-x94nb" Jan 21 11:11:17 crc kubenswrapper[4824]: W0121 11:11:17.275255 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6b9588d1_5876_4c2c_8de5_e5c0ad46f04e.slice/crio-b30310394ccb704d21dccf9961467707c8f16173677b5f28b9969e359c6eb524 WatchSource:0}: Error finding container b30310394ccb704d21dccf9961467707c8f16173677b5f28b9969e359c6eb524: Status 404 returned error can't find the container with id b30310394ccb704d21dccf9961467707c8f16173677b5f28b9969e359c6eb524 Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.307730 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06f56d86-ec6f-4b57-b314-f508c0dd4d9f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d25c94f1ae0bbabb733f0e1ea1d1ba39562a92af73047a39aacc03013e239681\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca06234486b59cf990c69d9501e23f935328e2ba21013cd71ed64831ea749bcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b63e45d66bcd11ff3479cc3ec6b450017da8d21d36e967b8d496ae0f24b0ee9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52758bcd31f2ba64edf03adfd4d1be6428015abff5b64c7c029fa787085f9c29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2736e73342d2adb175b0721dd4f40cc42d26fca06d5eb3334927518531c3b107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b35ad7dd0570c191eb0080739304f3cd43f17b3d1dddf895549441ebe50deaa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b35ad7dd0570c191eb0080739304f3cd43f17b3d1dddf895549441ebe50deaa6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://80cd5ea5e6fc19a88c1aab5c61f827f103ab798a5c84da8dc4d45cea86d5aacb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80cd5ea5e6fc19a88c1aab5c61f827f103ab798a5c84da8dc4d45cea86d5aacb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://234d8b914e6a88480021647ae65d74e9b21c55b6788312a094d00c8a42677d48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://234d8b914e6a88480021647ae65d74e9b21c55b6788312a094d00c8a42677d48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:17Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.327205 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" Jan 21 11:11:17 crc kubenswrapper[4824]: W0121 11:11:17.335835 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod33f3d922_4ffe_409b_a49a_d88c85898260.slice/crio-da8510a8b5ef1f15b44d065bf5855b68b621b13a0308b5a3f3e3b09af89ede7f WatchSource:0}: Error finding container da8510a8b5ef1f15b44d065bf5855b68b621b13a0308b5a3f3e3b09af89ede7f: Status 404 returned error can't find the container with id da8510a8b5ef1f15b44d065bf5855b68b621b13a0308b5a3f3e3b09af89ede7f Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.343819 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:17Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.358134 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-gjtnh" Jan 21 11:11:17 crc kubenswrapper[4824]: W0121 11:11:17.369154 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3a8d2cac_2d0f_4894_9bc0_2abd81699c47.slice/crio-e52ebed3ade7f1249cea45e214ee412ca574617c3d80124ed44d587e2650f1d9 WatchSource:0}: Error finding container e52ebed3ade7f1249cea45e214ee412ca574617c3d80124ed44d587e2650f1d9: Status 404 returned error can't find the container with id e52ebed3ade7f1249cea45e214ee412ca574617c3d80124ed44d587e2650f1d9 Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.385727 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33f3d922-4ffe-409b-a49a-d88c85898260\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpwst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpwst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zqjg9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:17Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.426052 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gjtnh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a8d2cac-2d0f-4894-9bc0-2abd81699c47\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gjtnh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:17Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.464603 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86d3005cd61c0973659e1c1c0d3ef1ddd6fcf95bede5465f4f2099e64a74bb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:17Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.503847 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:17Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.548994 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x94nb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgr8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x94nb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:17Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.573055 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-jr64s"] Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.573364 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-jr64s" Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.584674 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d87422473838f5218c9fb646ab2515f2f42592754ffdf07b19f7a8e093fa48a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:17Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.598104 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.617988 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.626305 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-flrnp\" (UniqueName: \"kubernetes.io/projected/9768c60d-5bee-4286-bced-076e584b04e1-kube-api-access-flrnp\") pod \"node-ca-jr64s\" (UID: \"9768c60d-5bee-4286-bced-076e584b04e1\") " pod="openshift-image-registry/node-ca-jr64s" Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.626375 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9768c60d-5bee-4286-bced-076e584b04e1-host\") pod \"node-ca-jr64s\" (UID: \"9768c60d-5bee-4286-bced-076e584b04e1\") " pod="openshift-image-registry/node-ca-jr64s" Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.626423 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/9768c60d-5bee-4286-bced-076e584b04e1-serviceca\") pod \"node-ca-jr64s\" (UID: \"9768c60d-5bee-4286-bced-076e584b04e1\") " pod="openshift-image-registry/node-ca-jr64s" Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.637138 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.657224 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.704714 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://516c009d4720ab5b393c7482872f12fed22d4f4414bd1f3400411c0e8da88ce4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee18d293036971832c8d9c3f7b9eca84ab779ff6768a2db3c6f7bdcf3433cdef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:17Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.727580 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9768c60d-5bee-4286-bced-076e584b04e1-host\") pod \"node-ca-jr64s\" (UID: \"9768c60d-5bee-4286-bced-076e584b04e1\") " pod="openshift-image-registry/node-ca-jr64s" Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.727613 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/9768c60d-5bee-4286-bced-076e584b04e1-serviceca\") pod \"node-ca-jr64s\" (UID: \"9768c60d-5bee-4286-bced-076e584b04e1\") " pod="openshift-image-registry/node-ca-jr64s" Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.727630 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-flrnp\" (UniqueName: \"kubernetes.io/projected/9768c60d-5bee-4286-bced-076e584b04e1-kube-api-access-flrnp\") pod \"node-ca-jr64s\" (UID: \"9768c60d-5bee-4286-bced-076e584b04e1\") " pod="openshift-image-registry/node-ca-jr64s" Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.727681 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9768c60d-5bee-4286-bced-076e584b04e1-host\") pod \"node-ca-jr64s\" (UID: \"9768c60d-5bee-4286-bced-076e584b04e1\") " pod="openshift-image-registry/node-ca-jr64s" Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.728511 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/9768c60d-5bee-4286-bced-076e584b04e1-serviceca\") pod \"node-ca-jr64s\" (UID: \"9768c60d-5bee-4286-bced-076e584b04e1\") " pod="openshift-image-registry/node-ca-jr64s" Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.748262 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"26b1108a-4b98-4e39-a3cd-e0d055089fd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-htkvk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:17Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.769630 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-flrnp\" (UniqueName: \"kubernetes.io/projected/9768c60d-5bee-4286-bced-076e584b04e1-kube-api-access-flrnp\") pod \"node-ca-jr64s\" (UID: \"9768c60d-5bee-4286-bced-076e584b04e1\") " pod="openshift-image-registry/node-ca-jr64s" Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.773258 4824 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.774703 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.774725 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.774733 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.774796 4824 kubelet_node_status.go:76] "Attempting to register node" node="crc" Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.828643 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06f56d86-ec6f-4b57-b314-f508c0dd4d9f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d25c94f1ae0bbabb733f0e1ea1d1ba39562a92af73047a39aacc03013e239681\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca06234486b59cf990c69d9501e23f935328e2ba21013cd71ed64831ea749bcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b63e45d66bcd11ff3479cc3ec6b450017da8d21d36e967b8d496ae0f24b0ee9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52758bcd31f2ba64edf03adfd4d1be6428015abff5b64c7c029fa787085f9c29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2736e73342d2adb175b0721dd4f40cc42d26fca06d5eb3334927518531c3b107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b35ad7dd0570c191eb0080739304f3cd43f17b3d1dddf895549441ebe50deaa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b35ad7dd0570c191eb0080739304f3cd43f17b3d1dddf895549441ebe50deaa6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://80cd5ea5e6fc19a88c1aab5c61f827f103ab798a5c84da8dc4d45cea86d5aacb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80cd5ea5e6fc19a88c1aab5c61f827f103ab798a5c84da8dc4d45cea86d5aacb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://234d8b914e6a88480021647ae65d74e9b21c55b6788312a094d00c8a42677d48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://234d8b914e6a88480021647ae65d74e9b21c55b6788312a094d00c8a42677d48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:17Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.838384 4824 kubelet_node_status.go:115] "Node was previously registered" node="crc" Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.838606 4824 kubelet_node_status.go:79] "Successfully registered node" node="crc" Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.839449 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.839476 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.839484 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.839495 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.839504 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:17Z","lastTransitionTime":"2026-01-21T11:11:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:17 crc kubenswrapper[4824]: E0121 11:11:17.847769 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"695c676c-720c-4921-9e71-56b96f2fb46b\\\",\\\"systemUUID\\\":\\\"d00624f5-786f-41dc-bdc1-a1875f167d65\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:17Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.850255 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.850287 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.850297 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.850310 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.850317 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:17Z","lastTransitionTime":"2026-01-21T11:11:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:17 crc kubenswrapper[4824]: E0121 11:11:17.858475 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"695c676c-720c-4921-9e71-56b96f2fb46b\\\",\\\"systemUUID\\\":\\\"d00624f5-786f-41dc-bdc1-a1875f167d65\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:17Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.860882 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.860972 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.860985 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.860997 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.861019 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:17Z","lastTransitionTime":"2026-01-21T11:11:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:17 crc kubenswrapper[4824]: E0121 11:11:17.870942 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"695c676c-720c-4921-9e71-56b96f2fb46b\\\",\\\"systemUUID\\\":\\\"d00624f5-786f-41dc-bdc1-a1875f167d65\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:17Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.873515 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.873577 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.873598 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.873611 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.873619 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:17Z","lastTransitionTime":"2026-01-21T11:11:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.882135 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-jr64s" Jan 21 11:11:17 crc kubenswrapper[4824]: E0121 11:11:17.882398 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"695c676c-720c-4921-9e71-56b96f2fb46b\\\",\\\"systemUUID\\\":\\\"d00624f5-786f-41dc-bdc1-a1875f167d65\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:17Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.885168 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:17Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.885811 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.885836 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.885843 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.885860 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.885868 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:17Z","lastTransitionTime":"2026-01-21T11:11:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:17 crc kubenswrapper[4824]: W0121 11:11:17.892613 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9768c60d_5bee_4286_bced_076e584b04e1.slice/crio-b14e560c96f826d1223c405141516af2082c14ca9109faa9a67b058b78fd0c1a WatchSource:0}: Error finding container b14e560c96f826d1223c405141516af2082c14ca9109faa9a67b058b78fd0c1a: Status 404 returned error can't find the container with id b14e560c96f826d1223c405141516af2082c14ca9109faa9a67b058b78fd0c1a Jan 21 11:11:17 crc kubenswrapper[4824]: E0121 11:11:17.896813 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"695c676c-720c-4921-9e71-56b96f2fb46b\\\",\\\"systemUUID\\\":\\\"d00624f5-786f-41dc-bdc1-a1875f167d65\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:17Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:17 crc kubenswrapper[4824]: E0121 11:11:17.896914 4824 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.897809 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.897888 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.898053 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.898198 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.898332 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:17Z","lastTransitionTime":"2026-01-21T11:11:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.909070 4824 transport.go:147] "Certificate rotation detected, shutting down client connections to start using new credentials" Jan 21 11:11:17 crc kubenswrapper[4824]: W0121 11:11:17.909926 4824 reflector.go:484] object-"openshift-image-registry"/"node-ca-dockercfg-4777p": watch of *v1.Secret ended with: very short watch: object-"openshift-image-registry"/"node-ca-dockercfg-4777p": Unexpected watch close - watch lasted less than a second and no items received Jan 21 11:11:17 crc kubenswrapper[4824]: W0121 11:11:17.910365 4824 reflector.go:484] object-"openshift-multus"/"cni-copy-resources": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-multus"/"cni-copy-resources": Unexpected watch close - watch lasted less than a second and no items received Jan 21 11:11:17 crc kubenswrapper[4824]: W0121 11:11:17.913426 4824 reflector.go:484] object-"openshift-image-registry"/"openshift-service-ca.crt": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-image-registry"/"openshift-service-ca.crt": Unexpected watch close - watch lasted less than a second and no items received Jan 21 11:11:17 crc kubenswrapper[4824]: W0121 11:11:17.913509 4824 reflector.go:484] object-"openshift-image-registry"/"kube-root-ca.crt": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-image-registry"/"kube-root-ca.crt": Unexpected watch close - watch lasted less than a second and no items received Jan 21 11:11:17 crc kubenswrapper[4824]: W0121 11:11:17.913574 4824 reflector.go:484] object-"openshift-image-registry"/"image-registry-certificates": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-image-registry"/"image-registry-certificates": Unexpected watch close - watch lasted less than a second and no items received Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.931688 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gjtnh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a8d2cac-2d0f-4894-9bc0-2abd81699c47\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gjtnh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:17Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.963221 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86d3005cd61c0973659e1c1c0d3ef1ddd6fcf95bede5465f4f2099e64a74bb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:17Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:17 crc kubenswrapper[4824]: I0121 11:11:17.991400 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-08 19:25:27.25383742 +0000 UTC Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.000202 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.000233 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.000242 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.000256 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.000264 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:18Z","lastTransitionTime":"2026-01-21T11:11:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.003852 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:18Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.043334 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x94nb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgr8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x94nb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:18Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.048645 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 11:11:18 crc kubenswrapper[4824]: E0121 11:11:18.048757 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.082426 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33f3d922-4ffe-409b-a49a-d88c85898260\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpwst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpwst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zqjg9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:18Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.101864 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.101887 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.101895 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.101905 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.101914 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:18Z","lastTransitionTime":"2026-01-21T11:11:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.122362 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d87422473838f5218c9fb646ab2515f2f42592754ffdf07b19f7a8e093fa48a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:18Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.138724 4824 generic.go:334] "Generic (PLEG): container finished" podID="3a8d2cac-2d0f-4894-9bc0-2abd81699c47" containerID="b2239186f1de46481cebfcafb1cab85cd320a205b5a0dfc6dd1b321d9baedeba" exitCode=0 Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.138811 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-gjtnh" event={"ID":"3a8d2cac-2d0f-4894-9bc0-2abd81699c47","Type":"ContainerDied","Data":"b2239186f1de46481cebfcafb1cab85cd320a205b5a0dfc6dd1b321d9baedeba"} Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.138848 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-gjtnh" event={"ID":"3a8d2cac-2d0f-4894-9bc0-2abd81699c47","Type":"ContainerStarted","Data":"e52ebed3ade7f1249cea45e214ee412ca574617c3d80124ed44d587e2650f1d9"} Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.139942 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-jr64s" event={"ID":"9768c60d-5bee-4286-bced-076e584b04e1","Type":"ContainerStarted","Data":"b14e560c96f826d1223c405141516af2082c14ca9109faa9a67b058b78fd0c1a"} Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.141783 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" event={"ID":"33f3d922-4ffe-409b-a49a-d88c85898260","Type":"ContainerStarted","Data":"2a9e57b39eed8300107e944b9fcf6a2ff1e75991f8f37387a3de615b14f9bc78"} Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.141807 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" event={"ID":"33f3d922-4ffe-409b-a49a-d88c85898260","Type":"ContainerStarted","Data":"2b1b37c6d2994021cfc30f85e92c40ae35677113b17aa066d52bc7a76ae5c153"} Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.141820 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" event={"ID":"33f3d922-4ffe-409b-a49a-d88c85898260","Type":"ContainerStarted","Data":"da8510a8b5ef1f15b44d065bf5855b68b621b13a0308b5a3f3e3b09af89ede7f"} Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.147255 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-x94nb" event={"ID":"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e","Type":"ContainerStarted","Data":"75d52224135b75133fc000611cd179240b9174f6492d3263341716c7e42ccb15"} Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.147296 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-x94nb" event={"ID":"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e","Type":"ContainerStarted","Data":"b30310394ccb704d21dccf9961467707c8f16173677b5f28b9969e359c6eb524"} Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.164930 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://516c009d4720ab5b393c7482872f12fed22d4f4414bd1f3400411c0e8da88ce4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee18d293036971832c8d9c3f7b9eca84ab779ff6768a2db3c6f7bdcf3433cdef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:18Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.203984 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.204027 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.204037 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.204052 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.204064 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:18Z","lastTransitionTime":"2026-01-21T11:11:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.207646 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"26b1108a-4b98-4e39-a3cd-e0d055089fd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-htkvk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:18Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.244509 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:18Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.283264 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lv4rn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1267539-9adf-466f-8027-8ede009d19a0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eed1b884aa53e1ee318169a0cc2c9ad10130aa363ae4ac55e4bf065237756f53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbmfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lv4rn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:18Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.309092 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.309377 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.309387 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.309401 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.309410 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:18Z","lastTransitionTime":"2026-01-21T11:11:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.324709 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jr64s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9768c60d-5bee-4286-bced-076e584b04e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flrnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:17Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jr64s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:18Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.366317 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d31ab61-f94c-4f49-85d0-c9e91b83069b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a66b5c33884af00408dc35cf02f08ebd8570bd695f9f1fe558bf54fd06930431\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7514786602746dabfa690da48d48e233fe68d6dd6674aa5ecd489c0b3185f08\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61b14b6846f94cdeba8f1f61b3754e7663041f92c70356c4ee6cbb988a2e680f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a7023cfe4c098bcdd9c0477a8187f4ed7a0d2abdbe1e143436863948f441b4f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af11ad95bcf0c6bbd5fa4ab99bb592ce1c87f40d29c5ade7832414315820a569\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW0121 11:11:14.604290 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0121 11:11:14.604572 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0121 11:11:14.605875 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3284452394/tls.crt::/tmp/serving-cert-3284452394/tls.key\\\\\\\"\\\\nI0121 11:11:14.923427 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0121 11:11:14.928141 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0121 11:11:14.928162 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0121 11:11:14.928179 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0121 11:11:14.928184 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0121 11:11:14.933154 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0121 11:11:14.933230 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 11:11:14.933259 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 11:11:14.933282 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0121 11:11:14.933299 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0121 11:11:14.933315 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0121 11:11:14.933331 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0121 11:11:14.933177 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0121 11:11:14.934666 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48070c09fad511108ff22e44211cf26911509b8406d66948985cbfc3009d8233\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:18Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.404226 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9bb99c0a-683b-44da-9c0e-e8d284e87d53\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05e076febd697ed802ce5c7ee92a8c2c636a1a2c6609ad7eb01f17dace247c4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43529d1b9e1e5e22b0d56b023946119da8aacbae54b68c26ff0a1a82dc890c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ea232ad87393685aa323ce6393ff3869f8f8d423c2406a501c69c355535adfa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc6c7eff9e17b8a8ecc88bbfe703b5a01dbf1e02d342cede4f6a263a6526af43\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:18Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.411533 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.411564 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.411574 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.411588 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.411597 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:18Z","lastTransitionTime":"2026-01-21T11:11:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.448275 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06f56d86-ec6f-4b57-b314-f508c0dd4d9f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d25c94f1ae0bbabb733f0e1ea1d1ba39562a92af73047a39aacc03013e239681\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca06234486b59cf990c69d9501e23f935328e2ba21013cd71ed64831ea749bcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b63e45d66bcd11ff3479cc3ec6b450017da8d21d36e967b8d496ae0f24b0ee9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52758bcd31f2ba64edf03adfd4d1be6428015abff5b64c7c029fa787085f9c29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2736e73342d2adb175b0721dd4f40cc42d26fca06d5eb3334927518531c3b107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b35ad7dd0570c191eb0080739304f3cd43f17b3d1dddf895549441ebe50deaa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b35ad7dd0570c191eb0080739304f3cd43f17b3d1dddf895549441ebe50deaa6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://80cd5ea5e6fc19a88c1aab5c61f827f103ab798a5c84da8dc4d45cea86d5aacb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80cd5ea5e6fc19a88c1aab5c61f827f103ab798a5c84da8dc4d45cea86d5aacb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://234d8b914e6a88480021647ae65d74e9b21c55b6788312a094d00c8a42677d48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://234d8b914e6a88480021647ae65d74e9b21c55b6788312a094d00c8a42677d48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:18Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.484550 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:18Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.513483 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.513514 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.513523 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.513536 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.513544 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:18Z","lastTransitionTime":"2026-01-21T11:11:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.524203 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86d3005cd61c0973659e1c1c0d3ef1ddd6fcf95bede5465f4f2099e64a74bb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:18Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.565175 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:18Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.605909 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x94nb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75d52224135b75133fc000611cd179240b9174f6492d3263341716c7e42ccb15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgr8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x94nb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:18Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.615469 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.615497 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.615506 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.615520 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.615534 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:18Z","lastTransitionTime":"2026-01-21T11:11:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.634127 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.634219 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.634242 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 11:11:18 crc kubenswrapper[4824]: E0121 11:11:18.634271 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-21 11:11:22.634254482 +0000 UTC m=+24.927283774 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 11:11:18 crc kubenswrapper[4824]: E0121 11:11:18.634301 4824 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Jan 21 11:11:18 crc kubenswrapper[4824]: E0121 11:11:18.634336 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-21 11:11:22.634326488 +0000 UTC m=+24.927355779 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Jan 21 11:11:18 crc kubenswrapper[4824]: E0121 11:11:18.634426 4824 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 21 11:11:18 crc kubenswrapper[4824]: E0121 11:11:18.634515 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-21 11:11:22.634484828 +0000 UTC m=+24.927514130 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.643065 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33f3d922-4ffe-409b-a49a-d88c85898260\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a9e57b39eed8300107e944b9fcf6a2ff1e75991f8f37387a3de615b14f9bc78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpwst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b1b37c6d2994021cfc30f85e92c40ae35677113b17aa066d52bc7a76ae5c153\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpwst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zqjg9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:18Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.685874 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gjtnh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a8d2cac-2d0f-4894-9bc0-2abd81699c47\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2239186f1de46481cebfcafb1cab85cd320a205b5a0dfc6dd1b321d9baedeba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b2239186f1de46481cebfcafb1cab85cd320a205b5a0dfc6dd1b321d9baedeba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gjtnh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:18Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.717746 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.717785 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.717794 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.717808 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.717816 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:18Z","lastTransitionTime":"2026-01-21T11:11:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.723569 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d87422473838f5218c9fb646ab2515f2f42592754ffdf07b19f7a8e093fa48a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:18Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.735066 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.735111 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 11:11:18 crc kubenswrapper[4824]: E0121 11:11:18.735172 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 21 11:11:18 crc kubenswrapper[4824]: E0121 11:11:18.735198 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 21 11:11:18 crc kubenswrapper[4824]: E0121 11:11:18.735210 4824 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 21 11:11:18 crc kubenswrapper[4824]: E0121 11:11:18.735219 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 21 11:11:18 crc kubenswrapper[4824]: E0121 11:11:18.735239 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 21 11:11:18 crc kubenswrapper[4824]: E0121 11:11:18.735249 4824 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 21 11:11:18 crc kubenswrapper[4824]: E0121 11:11:18.735260 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-01-21 11:11:22.73524411 +0000 UTC m=+25.028273412 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 21 11:11:18 crc kubenswrapper[4824]: E0121 11:11:18.735285 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-01-21 11:11:22.735274258 +0000 UTC m=+25.028303549 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.764047 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://516c009d4720ab5b393c7482872f12fed22d4f4414bd1f3400411c0e8da88ce4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee18d293036971832c8d9c3f7b9eca84ab779ff6768a2db3c6f7bdcf3433cdef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:18Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.807515 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"26b1108a-4b98-4e39-a3cd-e0d055089fd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-htkvk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:18Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.819421 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.819461 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.819471 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.819485 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.819494 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:18Z","lastTransitionTime":"2026-01-21T11:11:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.843123 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jr64s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9768c60d-5bee-4286-bced-076e584b04e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flrnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:17Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jr64s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:18Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.877725 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.906219 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d31ab61-f94c-4f49-85d0-c9e91b83069b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a66b5c33884af00408dc35cf02f08ebd8570bd695f9f1fe558bf54fd06930431\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7514786602746dabfa690da48d48e233fe68d6dd6674aa5ecd489c0b3185f08\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61b14b6846f94cdeba8f1f61b3754e7663041f92c70356c4ee6cbb988a2e680f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a7023cfe4c098bcdd9c0477a8187f4ed7a0d2abdbe1e143436863948f441b4f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af11ad95bcf0c6bbd5fa4ab99bb592ce1c87f40d29c5ade7832414315820a569\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW0121 11:11:14.604290 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0121 11:11:14.604572 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0121 11:11:14.605875 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3284452394/tls.crt::/tmp/serving-cert-3284452394/tls.key\\\\\\\"\\\\nI0121 11:11:14.923427 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0121 11:11:14.928141 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0121 11:11:14.928162 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0121 11:11:14.928179 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0121 11:11:14.928184 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0121 11:11:14.933154 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0121 11:11:14.933230 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 11:11:14.933259 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 11:11:14.933282 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0121 11:11:14.933299 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0121 11:11:14.933315 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0121 11:11:14.933331 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0121 11:11:14.933177 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0121 11:11:14.934666 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48070c09fad511108ff22e44211cf26911509b8406d66948985cbfc3009d8233\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:18Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.921322 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.921349 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.921359 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.921373 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.921382 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:18Z","lastTransitionTime":"2026-01-21T11:11:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.943913 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9bb99c0a-683b-44da-9c0e-e8d284e87d53\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05e076febd697ed802ce5c7ee92a8c2c636a1a2c6609ad7eb01f17dace247c4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43529d1b9e1e5e22b0d56b023946119da8aacbae54b68c26ff0a1a82dc890c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ea232ad87393685aa323ce6393ff3869f8f8d423c2406a501c69c355535adfa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc6c7eff9e17b8a8ecc88bbfe703b5a01dbf1e02d342cede4f6a263a6526af43\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:18Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.983129 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:18Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:18 crc kubenswrapper[4824]: I0121 11:11:18.992322 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-25 05:27:30.994738474 +0000 UTC Jan 21 11:11:19 crc kubenswrapper[4824]: I0121 11:11:19.022900 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:19 crc kubenswrapper[4824]: I0121 11:11:19.022925 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:19 crc kubenswrapper[4824]: I0121 11:11:19.022935 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:19 crc kubenswrapper[4824]: I0121 11:11:19.022947 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:19 crc kubenswrapper[4824]: I0121 11:11:19.022968 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:19Z","lastTransitionTime":"2026-01-21T11:11:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:19 crc kubenswrapper[4824]: I0121 11:11:19.023810 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lv4rn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1267539-9adf-466f-8027-8ede009d19a0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eed1b884aa53e1ee318169a0cc2c9ad10130aa363ae4ac55e4bf065237756f53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbmfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lv4rn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:19Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:19 crc kubenswrapper[4824]: I0121 11:11:19.048747 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 11:11:19 crc kubenswrapper[4824]: E0121 11:11:19.048840 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 21 11:11:19 crc kubenswrapper[4824]: I0121 11:11:19.048751 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 11:11:19 crc kubenswrapper[4824]: E0121 11:11:19.048943 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 21 11:11:19 crc kubenswrapper[4824]: I0121 11:11:19.063665 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86d3005cd61c0973659e1c1c0d3ef1ddd6fcf95bede5465f4f2099e64a74bb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:19Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:19 crc kubenswrapper[4824]: I0121 11:11:19.077796 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Jan 21 11:11:19 crc kubenswrapper[4824]: I0121 11:11:19.122624 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:19Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:19 crc kubenswrapper[4824]: I0121 11:11:19.125005 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:19 crc kubenswrapper[4824]: I0121 11:11:19.125049 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:19 crc kubenswrapper[4824]: I0121 11:11:19.125059 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:19 crc kubenswrapper[4824]: I0121 11:11:19.125071 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:19 crc kubenswrapper[4824]: I0121 11:11:19.125079 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:19Z","lastTransitionTime":"2026-01-21T11:11:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:19 crc kubenswrapper[4824]: I0121 11:11:19.152639 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" event={"ID":"26b1108a-4b98-4e39-a3cd-e0d055089fd8","Type":"ContainerStarted","Data":"948e0497eaffa7b0d84b4085c9ed8d987eb2c6527c9a82c10e576ed9c9c1e21b"} Jan 21 11:11:19 crc kubenswrapper[4824]: I0121 11:11:19.153931 4824 generic.go:334] "Generic (PLEG): container finished" podID="3a8d2cac-2d0f-4894-9bc0-2abd81699c47" containerID="0db5ecd7759d974ae1b0d8065b3b8bd4d7957fae16eb5bab2d71c2fe69243816" exitCode=0 Jan 21 11:11:19 crc kubenswrapper[4824]: I0121 11:11:19.154005 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-gjtnh" event={"ID":"3a8d2cac-2d0f-4894-9bc0-2abd81699c47","Type":"ContainerDied","Data":"0db5ecd7759d974ae1b0d8065b3b8bd4d7957fae16eb5bab2d71c2fe69243816"} Jan 21 11:11:19 crc kubenswrapper[4824]: I0121 11:11:19.155004 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-jr64s" event={"ID":"9768c60d-5bee-4286-bced-076e584b04e1","Type":"ContainerStarted","Data":"91158f14d2e98ac9f875a5e7101df3107ef28aa5fc5d62218f460af98ff920fe"} Jan 21 11:11:19 crc kubenswrapper[4824]: I0121 11:11:19.165169 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x94nb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75d52224135b75133fc000611cd179240b9174f6492d3263341716c7e42ccb15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgr8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x94nb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:19Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:19 crc kubenswrapper[4824]: I0121 11:11:19.204264 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33f3d922-4ffe-409b-a49a-d88c85898260\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a9e57b39eed8300107e944b9fcf6a2ff1e75991f8f37387a3de615b14f9bc78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpwst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b1b37c6d2994021cfc30f85e92c40ae35677113b17aa066d52bc7a76ae5c153\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpwst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zqjg9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:19Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:19 crc kubenswrapper[4824]: I0121 11:11:19.227077 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:19 crc kubenswrapper[4824]: I0121 11:11:19.227114 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:19 crc kubenswrapper[4824]: I0121 11:11:19.227123 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:19 crc kubenswrapper[4824]: I0121 11:11:19.227136 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:19 crc kubenswrapper[4824]: I0121 11:11:19.227171 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:19Z","lastTransitionTime":"2026-01-21T11:11:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:19 crc kubenswrapper[4824]: I0121 11:11:19.245437 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gjtnh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a8d2cac-2d0f-4894-9bc0-2abd81699c47\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2239186f1de46481cebfcafb1cab85cd320a205b5a0dfc6dd1b321d9baedeba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b2239186f1de46481cebfcafb1cab85cd320a205b5a0dfc6dd1b321d9baedeba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gjtnh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:19Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:19 crc kubenswrapper[4824]: I0121 11:11:19.283808 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d87422473838f5218c9fb646ab2515f2f42592754ffdf07b19f7a8e093fa48a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:19Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:19 crc kubenswrapper[4824]: I0121 11:11:19.323598 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://516c009d4720ab5b393c7482872f12fed22d4f4414bd1f3400411c0e8da88ce4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee18d293036971832c8d9c3f7b9eca84ab779ff6768a2db3c6f7bdcf3433cdef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:19Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:19 crc kubenswrapper[4824]: I0121 11:11:19.329356 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:19 crc kubenswrapper[4824]: I0121 11:11:19.329383 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:19 crc kubenswrapper[4824]: I0121 11:11:19.329391 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:19 crc kubenswrapper[4824]: I0121 11:11:19.329404 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:19 crc kubenswrapper[4824]: I0121 11:11:19.329412 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:19Z","lastTransitionTime":"2026-01-21T11:11:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:19 crc kubenswrapper[4824]: I0121 11:11:19.357570 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Jan 21 11:11:19 crc kubenswrapper[4824]: I0121 11:11:19.387081 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"26b1108a-4b98-4e39-a3cd-e0d055089fd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-htkvk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:19Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:19 crc kubenswrapper[4824]: I0121 11:11:19.396994 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Jan 21 11:11:19 crc kubenswrapper[4824]: I0121 11:11:19.431100 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:19 crc kubenswrapper[4824]: I0121 11:11:19.431127 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:19 crc kubenswrapper[4824]: I0121 11:11:19.431136 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:19 crc kubenswrapper[4824]: I0121 11:11:19.431148 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:19 crc kubenswrapper[4824]: I0121 11:11:19.431157 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:19Z","lastTransitionTime":"2026-01-21T11:11:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:19 crc kubenswrapper[4824]: I0121 11:11:19.447561 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d31ab61-f94c-4f49-85d0-c9e91b83069b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a66b5c33884af00408dc35cf02f08ebd8570bd695f9f1fe558bf54fd06930431\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7514786602746dabfa690da48d48e233fe68d6dd6674aa5ecd489c0b3185f08\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61b14b6846f94cdeba8f1f61b3754e7663041f92c70356c4ee6cbb988a2e680f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a7023cfe4c098bcdd9c0477a8187f4ed7a0d2abdbe1e143436863948f441b4f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af11ad95bcf0c6bbd5fa4ab99bb592ce1c87f40d29c5ade7832414315820a569\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW0121 11:11:14.604290 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0121 11:11:14.604572 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0121 11:11:14.605875 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3284452394/tls.crt::/tmp/serving-cert-3284452394/tls.key\\\\\\\"\\\\nI0121 11:11:14.923427 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0121 11:11:14.928141 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0121 11:11:14.928162 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0121 11:11:14.928179 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0121 11:11:14.928184 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0121 11:11:14.933154 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0121 11:11:14.933230 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 11:11:14.933259 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 11:11:14.933282 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0121 11:11:14.933299 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0121 11:11:14.933315 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0121 11:11:14.933331 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0121 11:11:14.933177 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0121 11:11:14.934666 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48070c09fad511108ff22e44211cf26911509b8406d66948985cbfc3009d8233\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:19Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:19 crc kubenswrapper[4824]: I0121 11:11:19.483419 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9bb99c0a-683b-44da-9c0e-e8d284e87d53\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05e076febd697ed802ce5c7ee92a8c2c636a1a2c6609ad7eb01f17dace247c4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43529d1b9e1e5e22b0d56b023946119da8aacbae54b68c26ff0a1a82dc890c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ea232ad87393685aa323ce6393ff3869f8f8d423c2406a501c69c355535adfa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc6c7eff9e17b8a8ecc88bbfe703b5a01dbf1e02d342cede4f6a263a6526af43\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:19Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:19 crc kubenswrapper[4824]: I0121 11:11:19.522568 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:19Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:19 crc kubenswrapper[4824]: I0121 11:11:19.533041 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:19 crc kubenswrapper[4824]: I0121 11:11:19.533071 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:19 crc kubenswrapper[4824]: I0121 11:11:19.533080 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:19 crc kubenswrapper[4824]: I0121 11:11:19.533095 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:19 crc kubenswrapper[4824]: I0121 11:11:19.533104 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:19Z","lastTransitionTime":"2026-01-21T11:11:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:19 crc kubenswrapper[4824]: I0121 11:11:19.562676 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lv4rn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1267539-9adf-466f-8027-8ede009d19a0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eed1b884aa53e1ee318169a0cc2c9ad10130aa363ae4ac55e4bf065237756f53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbmfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lv4rn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:19Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:19 crc kubenswrapper[4824]: I0121 11:11:19.601975 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jr64s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9768c60d-5bee-4286-bced-076e584b04e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flrnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:17Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jr64s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:19Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:19 crc kubenswrapper[4824]: I0121 11:11:19.635104 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:19 crc kubenswrapper[4824]: I0121 11:11:19.635141 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:19 crc kubenswrapper[4824]: I0121 11:11:19.635152 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:19 crc kubenswrapper[4824]: I0121 11:11:19.635165 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:19 crc kubenswrapper[4824]: I0121 11:11:19.635174 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:19Z","lastTransitionTime":"2026-01-21T11:11:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:19 crc kubenswrapper[4824]: I0121 11:11:19.647634 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06f56d86-ec6f-4b57-b314-f508c0dd4d9f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d25c94f1ae0bbabb733f0e1ea1d1ba39562a92af73047a39aacc03013e239681\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca06234486b59cf990c69d9501e23f935328e2ba21013cd71ed64831ea749bcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b63e45d66bcd11ff3479cc3ec6b450017da8d21d36e967b8d496ae0f24b0ee9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52758bcd31f2ba64edf03adfd4d1be6428015abff5b64c7c029fa787085f9c29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2736e73342d2adb175b0721dd4f40cc42d26fca06d5eb3334927518531c3b107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b35ad7dd0570c191eb0080739304f3cd43f17b3d1dddf895549441ebe50deaa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b35ad7dd0570c191eb0080739304f3cd43f17b3d1dddf895549441ebe50deaa6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://80cd5ea5e6fc19a88c1aab5c61f827f103ab798a5c84da8dc4d45cea86d5aacb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80cd5ea5e6fc19a88c1aab5c61f827f103ab798a5c84da8dc4d45cea86d5aacb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://234d8b914e6a88480021647ae65d74e9b21c55b6788312a094d00c8a42677d48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://234d8b914e6a88480021647ae65d74e9b21c55b6788312a094d00c8a42677d48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:19Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:19 crc kubenswrapper[4824]: I0121 11:11:19.683534 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:19Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:19 crc kubenswrapper[4824]: I0121 11:11:19.726227 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06f56d86-ec6f-4b57-b314-f508c0dd4d9f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d25c94f1ae0bbabb733f0e1ea1d1ba39562a92af73047a39aacc03013e239681\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca06234486b59cf990c69d9501e23f935328e2ba21013cd71ed64831ea749bcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b63e45d66bcd11ff3479cc3ec6b450017da8d21d36e967b8d496ae0f24b0ee9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52758bcd31f2ba64edf03adfd4d1be6428015abff5b64c7c029fa787085f9c29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2736e73342d2adb175b0721dd4f40cc42d26fca06d5eb3334927518531c3b107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b35ad7dd0570c191eb0080739304f3cd43f17b3d1dddf895549441ebe50deaa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b35ad7dd0570c191eb0080739304f3cd43f17b3d1dddf895549441ebe50deaa6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://80cd5ea5e6fc19a88c1aab5c61f827f103ab798a5c84da8dc4d45cea86d5aacb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80cd5ea5e6fc19a88c1aab5c61f827f103ab798a5c84da8dc4d45cea86d5aacb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://234d8b914e6a88480021647ae65d74e9b21c55b6788312a094d00c8a42677d48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://234d8b914e6a88480021647ae65d74e9b21c55b6788312a094d00c8a42677d48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:19Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:19 crc kubenswrapper[4824]: I0121 11:11:19.737359 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:19 crc kubenswrapper[4824]: I0121 11:11:19.737394 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:19 crc kubenswrapper[4824]: I0121 11:11:19.737404 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:19 crc kubenswrapper[4824]: I0121 11:11:19.737417 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:19 crc kubenswrapper[4824]: I0121 11:11:19.737426 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:19Z","lastTransitionTime":"2026-01-21T11:11:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:19 crc kubenswrapper[4824]: I0121 11:11:19.762817 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:19Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:19 crc kubenswrapper[4824]: I0121 11:11:19.803275 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x94nb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75d52224135b75133fc000611cd179240b9174f6492d3263341716c7e42ccb15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgr8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x94nb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:19Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:19 crc kubenswrapper[4824]: I0121 11:11:19.839984 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:19 crc kubenswrapper[4824]: I0121 11:11:19.840016 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:19 crc kubenswrapper[4824]: I0121 11:11:19.840039 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:19 crc kubenswrapper[4824]: I0121 11:11:19.840051 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:19 crc kubenswrapper[4824]: I0121 11:11:19.840060 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:19Z","lastTransitionTime":"2026-01-21T11:11:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:19 crc kubenswrapper[4824]: I0121 11:11:19.842498 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33f3d922-4ffe-409b-a49a-d88c85898260\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a9e57b39eed8300107e944b9fcf6a2ff1e75991f8f37387a3de615b14f9bc78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpwst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b1b37c6d2994021cfc30f85e92c40ae35677113b17aa066d52bc7a76ae5c153\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpwst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zqjg9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:19Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:19 crc kubenswrapper[4824]: I0121 11:11:19.886532 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gjtnh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a8d2cac-2d0f-4894-9bc0-2abd81699c47\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2239186f1de46481cebfcafb1cab85cd320a205b5a0dfc6dd1b321d9baedeba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b2239186f1de46481cebfcafb1cab85cd320a205b5a0dfc6dd1b321d9baedeba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0db5ecd7759d974ae1b0d8065b3b8bd4d7957fae16eb5bab2d71c2fe69243816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0db5ecd7759d974ae1b0d8065b3b8bd4d7957fae16eb5bab2d71c2fe69243816\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gjtnh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:19Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:19 crc kubenswrapper[4824]: I0121 11:11:19.923822 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86d3005cd61c0973659e1c1c0d3ef1ddd6fcf95bede5465f4f2099e64a74bb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:19Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:19 crc kubenswrapper[4824]: I0121 11:11:19.942427 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:19 crc kubenswrapper[4824]: I0121 11:11:19.942453 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:19 crc kubenswrapper[4824]: I0121 11:11:19.942462 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:19 crc kubenswrapper[4824]: I0121 11:11:19.942473 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:19 crc kubenswrapper[4824]: I0121 11:11:19.942483 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:19Z","lastTransitionTime":"2026-01-21T11:11:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:19 crc kubenswrapper[4824]: I0121 11:11:19.964323 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:19Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:19 crc kubenswrapper[4824]: I0121 11:11:19.993341 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-16 03:32:15.942227212 +0000 UTC Jan 21 11:11:20 crc kubenswrapper[4824]: I0121 11:11:20.007225 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"26b1108a-4b98-4e39-a3cd-e0d055089fd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-htkvk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:20Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:20 crc kubenswrapper[4824]: I0121 11:11:20.042408 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d87422473838f5218c9fb646ab2515f2f42592754ffdf07b19f7a8e093fa48a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:20Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:20 crc kubenswrapper[4824]: I0121 11:11:20.044975 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:20 crc kubenswrapper[4824]: I0121 11:11:20.045010 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:20 crc kubenswrapper[4824]: I0121 11:11:20.045033 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:20 crc kubenswrapper[4824]: I0121 11:11:20.045047 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:20 crc kubenswrapper[4824]: I0121 11:11:20.045056 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:20Z","lastTransitionTime":"2026-01-21T11:11:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:20 crc kubenswrapper[4824]: I0121 11:11:20.048567 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 11:11:20 crc kubenswrapper[4824]: E0121 11:11:20.048658 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 21 11:11:20 crc kubenswrapper[4824]: I0121 11:11:20.084057 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://516c009d4720ab5b393c7482872f12fed22d4f4414bd1f3400411c0e8da88ce4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee18d293036971832c8d9c3f7b9eca84ab779ff6768a2db3c6f7bdcf3433cdef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:20Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:20 crc kubenswrapper[4824]: I0121 11:11:20.122851 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9bb99c0a-683b-44da-9c0e-e8d284e87d53\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05e076febd697ed802ce5c7ee92a8c2c636a1a2c6609ad7eb01f17dace247c4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43529d1b9e1e5e22b0d56b023946119da8aacbae54b68c26ff0a1a82dc890c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ea232ad87393685aa323ce6393ff3869f8f8d423c2406a501c69c355535adfa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc6c7eff9e17b8a8ecc88bbfe703b5a01dbf1e02d342cede4f6a263a6526af43\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:20Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:20 crc kubenswrapper[4824]: I0121 11:11:20.147132 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:20 crc kubenswrapper[4824]: I0121 11:11:20.147158 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:20 crc kubenswrapper[4824]: I0121 11:11:20.147166 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:20 crc kubenswrapper[4824]: I0121 11:11:20.147178 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:20 crc kubenswrapper[4824]: I0121 11:11:20.147187 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:20Z","lastTransitionTime":"2026-01-21T11:11:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:20 crc kubenswrapper[4824]: I0121 11:11:20.159324 4824 generic.go:334] "Generic (PLEG): container finished" podID="3a8d2cac-2d0f-4894-9bc0-2abd81699c47" containerID="ba9b3278c208fc7aac630ca7a0daa3de0823f85d0799d6f1f278da80ae14c862" exitCode=0 Jan 21 11:11:20 crc kubenswrapper[4824]: I0121 11:11:20.159405 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-gjtnh" event={"ID":"3a8d2cac-2d0f-4894-9bc0-2abd81699c47","Type":"ContainerDied","Data":"ba9b3278c208fc7aac630ca7a0daa3de0823f85d0799d6f1f278da80ae14c862"} Jan 21 11:11:20 crc kubenswrapper[4824]: I0121 11:11:20.163471 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:20Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:20 crc kubenswrapper[4824]: I0121 11:11:20.201922 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lv4rn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1267539-9adf-466f-8027-8ede009d19a0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eed1b884aa53e1ee318169a0cc2c9ad10130aa363ae4ac55e4bf065237756f53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbmfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lv4rn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:20Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:20 crc kubenswrapper[4824]: I0121 11:11:20.243916 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jr64s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9768c60d-5bee-4286-bced-076e584b04e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91158f14d2e98ac9f875a5e7101df3107ef28aa5fc5d62218f460af98ff920fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flrnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:17Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jr64s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:20Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:20 crc kubenswrapper[4824]: I0121 11:11:20.249765 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:20 crc kubenswrapper[4824]: I0121 11:11:20.249802 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:20 crc kubenswrapper[4824]: I0121 11:11:20.249811 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:20 crc kubenswrapper[4824]: I0121 11:11:20.249824 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:20 crc kubenswrapper[4824]: I0121 11:11:20.249832 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:20Z","lastTransitionTime":"2026-01-21T11:11:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:20 crc kubenswrapper[4824]: I0121 11:11:20.284518 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d31ab61-f94c-4f49-85d0-c9e91b83069b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a66b5c33884af00408dc35cf02f08ebd8570bd695f9f1fe558bf54fd06930431\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7514786602746dabfa690da48d48e233fe68d6dd6674aa5ecd489c0b3185f08\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61b14b6846f94cdeba8f1f61b3754e7663041f92c70356c4ee6cbb988a2e680f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a7023cfe4c098bcdd9c0477a8187f4ed7a0d2abdbe1e143436863948f441b4f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af11ad95bcf0c6bbd5fa4ab99bb592ce1c87f40d29c5ade7832414315820a569\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW0121 11:11:14.604290 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0121 11:11:14.604572 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0121 11:11:14.605875 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3284452394/tls.crt::/tmp/serving-cert-3284452394/tls.key\\\\\\\"\\\\nI0121 11:11:14.923427 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0121 11:11:14.928141 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0121 11:11:14.928162 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0121 11:11:14.928179 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0121 11:11:14.928184 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0121 11:11:14.933154 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0121 11:11:14.933230 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 11:11:14.933259 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 11:11:14.933282 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0121 11:11:14.933299 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0121 11:11:14.933315 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0121 11:11:14.933331 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0121 11:11:14.933177 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0121 11:11:14.934666 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48070c09fad511108ff22e44211cf26911509b8406d66948985cbfc3009d8233\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:20Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:20 crc kubenswrapper[4824]: I0121 11:11:20.324517 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d31ab61-f94c-4f49-85d0-c9e91b83069b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a66b5c33884af00408dc35cf02f08ebd8570bd695f9f1fe558bf54fd06930431\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7514786602746dabfa690da48d48e233fe68d6dd6674aa5ecd489c0b3185f08\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61b14b6846f94cdeba8f1f61b3754e7663041f92c70356c4ee6cbb988a2e680f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a7023cfe4c098bcdd9c0477a8187f4ed7a0d2abdbe1e143436863948f441b4f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af11ad95bcf0c6bbd5fa4ab99bb592ce1c87f40d29c5ade7832414315820a569\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW0121 11:11:14.604290 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0121 11:11:14.604572 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0121 11:11:14.605875 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3284452394/tls.crt::/tmp/serving-cert-3284452394/tls.key\\\\\\\"\\\\nI0121 11:11:14.923427 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0121 11:11:14.928141 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0121 11:11:14.928162 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0121 11:11:14.928179 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0121 11:11:14.928184 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0121 11:11:14.933154 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0121 11:11:14.933230 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 11:11:14.933259 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 11:11:14.933282 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0121 11:11:14.933299 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0121 11:11:14.933315 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0121 11:11:14.933331 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0121 11:11:14.933177 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0121 11:11:14.934666 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48070c09fad511108ff22e44211cf26911509b8406d66948985cbfc3009d8233\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:20Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:20 crc kubenswrapper[4824]: I0121 11:11:20.352109 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:20 crc kubenswrapper[4824]: I0121 11:11:20.352136 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:20 crc kubenswrapper[4824]: I0121 11:11:20.352146 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:20 crc kubenswrapper[4824]: I0121 11:11:20.352157 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:20 crc kubenswrapper[4824]: I0121 11:11:20.352166 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:20Z","lastTransitionTime":"2026-01-21T11:11:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:20 crc kubenswrapper[4824]: I0121 11:11:20.363666 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9bb99c0a-683b-44da-9c0e-e8d284e87d53\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05e076febd697ed802ce5c7ee92a8c2c636a1a2c6609ad7eb01f17dace247c4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43529d1b9e1e5e22b0d56b023946119da8aacbae54b68c26ff0a1a82dc890c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ea232ad87393685aa323ce6393ff3869f8f8d423c2406a501c69c355535adfa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc6c7eff9e17b8a8ecc88bbfe703b5a01dbf1e02d342cede4f6a263a6526af43\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:20Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:20 crc kubenswrapper[4824]: I0121 11:11:20.402828 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:20Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:20 crc kubenswrapper[4824]: I0121 11:11:20.442265 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lv4rn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1267539-9adf-466f-8027-8ede009d19a0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eed1b884aa53e1ee318169a0cc2c9ad10130aa363ae4ac55e4bf065237756f53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbmfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lv4rn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:20Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:20 crc kubenswrapper[4824]: I0121 11:11:20.453496 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:20 crc kubenswrapper[4824]: I0121 11:11:20.453527 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:20 crc kubenswrapper[4824]: I0121 11:11:20.453536 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:20 crc kubenswrapper[4824]: I0121 11:11:20.453550 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:20 crc kubenswrapper[4824]: I0121 11:11:20.453559 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:20Z","lastTransitionTime":"2026-01-21T11:11:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:20 crc kubenswrapper[4824]: I0121 11:11:20.481529 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jr64s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9768c60d-5bee-4286-bced-076e584b04e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91158f14d2e98ac9f875a5e7101df3107ef28aa5fc5d62218f460af98ff920fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flrnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:17Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jr64s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:20Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:20 crc kubenswrapper[4824]: I0121 11:11:20.529302 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06f56d86-ec6f-4b57-b314-f508c0dd4d9f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d25c94f1ae0bbabb733f0e1ea1d1ba39562a92af73047a39aacc03013e239681\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca06234486b59cf990c69d9501e23f935328e2ba21013cd71ed64831ea749bcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b63e45d66bcd11ff3479cc3ec6b450017da8d21d36e967b8d496ae0f24b0ee9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52758bcd31f2ba64edf03adfd4d1be6428015abff5b64c7c029fa787085f9c29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2736e73342d2adb175b0721dd4f40cc42d26fca06d5eb3334927518531c3b107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b35ad7dd0570c191eb0080739304f3cd43f17b3d1dddf895549441ebe50deaa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b35ad7dd0570c191eb0080739304f3cd43f17b3d1dddf895549441ebe50deaa6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://80cd5ea5e6fc19a88c1aab5c61f827f103ab798a5c84da8dc4d45cea86d5aacb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80cd5ea5e6fc19a88c1aab5c61f827f103ab798a5c84da8dc4d45cea86d5aacb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://234d8b914e6a88480021647ae65d74e9b21c55b6788312a094d00c8a42677d48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://234d8b914e6a88480021647ae65d74e9b21c55b6788312a094d00c8a42677d48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:20Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:20 crc kubenswrapper[4824]: I0121 11:11:20.555595 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:20 crc kubenswrapper[4824]: I0121 11:11:20.555622 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:20 crc kubenswrapper[4824]: I0121 11:11:20.555630 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:20 crc kubenswrapper[4824]: I0121 11:11:20.555644 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:20 crc kubenswrapper[4824]: I0121 11:11:20.555653 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:20Z","lastTransitionTime":"2026-01-21T11:11:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:20 crc kubenswrapper[4824]: I0121 11:11:20.562120 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:20Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:20 crc kubenswrapper[4824]: I0121 11:11:20.602517 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86d3005cd61c0973659e1c1c0d3ef1ddd6fcf95bede5465f4f2099e64a74bb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:20Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:20 crc kubenswrapper[4824]: I0121 11:11:20.644574 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:20Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:20 crc kubenswrapper[4824]: I0121 11:11:20.657886 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:20 crc kubenswrapper[4824]: I0121 11:11:20.657926 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:20 crc kubenswrapper[4824]: I0121 11:11:20.657964 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:20 crc kubenswrapper[4824]: I0121 11:11:20.657980 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:20 crc kubenswrapper[4824]: I0121 11:11:20.658008 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:20Z","lastTransitionTime":"2026-01-21T11:11:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:20 crc kubenswrapper[4824]: I0121 11:11:20.683779 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x94nb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75d52224135b75133fc000611cd179240b9174f6492d3263341716c7e42ccb15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgr8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x94nb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:20Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:20 crc kubenswrapper[4824]: I0121 11:11:20.724968 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33f3d922-4ffe-409b-a49a-d88c85898260\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a9e57b39eed8300107e944b9fcf6a2ff1e75991f8f37387a3de615b14f9bc78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpwst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b1b37c6d2994021cfc30f85e92c40ae35677113b17aa066d52bc7a76ae5c153\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpwst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zqjg9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:20Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:20 crc kubenswrapper[4824]: I0121 11:11:20.759980 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:20 crc kubenswrapper[4824]: I0121 11:11:20.760015 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:20 crc kubenswrapper[4824]: I0121 11:11:20.760024 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:20 crc kubenswrapper[4824]: I0121 11:11:20.760051 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:20 crc kubenswrapper[4824]: I0121 11:11:20.760059 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:20Z","lastTransitionTime":"2026-01-21T11:11:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:20 crc kubenswrapper[4824]: I0121 11:11:20.766921 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gjtnh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a8d2cac-2d0f-4894-9bc0-2abd81699c47\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2239186f1de46481cebfcafb1cab85cd320a205b5a0dfc6dd1b321d9baedeba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b2239186f1de46481cebfcafb1cab85cd320a205b5a0dfc6dd1b321d9baedeba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0db5ecd7759d974ae1b0d8065b3b8bd4d7957fae16eb5bab2d71c2fe69243816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0db5ecd7759d974ae1b0d8065b3b8bd4d7957fae16eb5bab2d71c2fe69243816\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba9b3278c208fc7aac630ca7a0daa3de0823f85d0799d6f1f278da80ae14c862\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba9b3278c208fc7aac630ca7a0daa3de0823f85d0799d6f1f278da80ae14c862\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gjtnh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:20Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:20 crc kubenswrapper[4824]: I0121 11:11:20.805642 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d87422473838f5218c9fb646ab2515f2f42592754ffdf07b19f7a8e093fa48a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:20Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:20 crc kubenswrapper[4824]: I0121 11:11:20.843531 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://516c009d4720ab5b393c7482872f12fed22d4f4414bd1f3400411c0e8da88ce4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee18d293036971832c8d9c3f7b9eca84ab779ff6768a2db3c6f7bdcf3433cdef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:20Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:20 crc kubenswrapper[4824]: I0121 11:11:20.863528 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:20 crc kubenswrapper[4824]: I0121 11:11:20.863583 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:20 crc kubenswrapper[4824]: I0121 11:11:20.863595 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:20 crc kubenswrapper[4824]: I0121 11:11:20.863611 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:20 crc kubenswrapper[4824]: I0121 11:11:20.863621 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:20Z","lastTransitionTime":"2026-01-21T11:11:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:20 crc kubenswrapper[4824]: I0121 11:11:20.887681 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"26b1108a-4b98-4e39-a3cd-e0d055089fd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-htkvk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:20Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:20 crc kubenswrapper[4824]: I0121 11:11:20.897633 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Jan 21 11:11:20 crc kubenswrapper[4824]: I0121 11:11:20.965926 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:20 crc kubenswrapper[4824]: I0121 11:11:20.965986 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:20 crc kubenswrapper[4824]: I0121 11:11:20.965997 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:20 crc kubenswrapper[4824]: I0121 11:11:20.966011 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:20 crc kubenswrapper[4824]: I0121 11:11:20.966020 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:20Z","lastTransitionTime":"2026-01-21T11:11:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:20 crc kubenswrapper[4824]: I0121 11:11:20.994328 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-17 22:16:12.690104605 +0000 UTC Jan 21 11:11:21 crc kubenswrapper[4824]: I0121 11:11:21.048816 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 11:11:21 crc kubenswrapper[4824]: E0121 11:11:21.048901 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 21 11:11:21 crc kubenswrapper[4824]: I0121 11:11:21.048944 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 11:11:21 crc kubenswrapper[4824]: E0121 11:11:21.049064 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 21 11:11:21 crc kubenswrapper[4824]: I0121 11:11:21.068189 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:21 crc kubenswrapper[4824]: I0121 11:11:21.068220 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:21 crc kubenswrapper[4824]: I0121 11:11:21.068229 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:21 crc kubenswrapper[4824]: I0121 11:11:21.068239 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:21 crc kubenswrapper[4824]: I0121 11:11:21.068247 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:21Z","lastTransitionTime":"2026-01-21T11:11:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:21 crc kubenswrapper[4824]: I0121 11:11:21.164175 4824 generic.go:334] "Generic (PLEG): container finished" podID="3a8d2cac-2d0f-4894-9bc0-2abd81699c47" containerID="79b8f6c016b43c3ac3ba7b22cca9a6b72c262cb17ddd6236c5d4970ecc09680c" exitCode=0 Jan 21 11:11:21 crc kubenswrapper[4824]: I0121 11:11:21.164365 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-gjtnh" event={"ID":"3a8d2cac-2d0f-4894-9bc0-2abd81699c47","Type":"ContainerDied","Data":"79b8f6c016b43c3ac3ba7b22cca9a6b72c262cb17ddd6236c5d4970ecc09680c"} Jan 21 11:11:21 crc kubenswrapper[4824]: I0121 11:11:21.168096 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" event={"ID":"26b1108a-4b98-4e39-a3cd-e0d055089fd8","Type":"ContainerStarted","Data":"b5e832a8a65f1a11316259d61f077c9b3a22a03e96c05118dcfeaec1d72ec973"} Jan 21 11:11:21 crc kubenswrapper[4824]: I0121 11:11:21.168302 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" Jan 21 11:11:21 crc kubenswrapper[4824]: I0121 11:11:21.169232 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:21 crc kubenswrapper[4824]: I0121 11:11:21.169264 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:21 crc kubenswrapper[4824]: I0121 11:11:21.169274 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:21 crc kubenswrapper[4824]: I0121 11:11:21.169286 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:21 crc kubenswrapper[4824]: I0121 11:11:21.169294 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:21Z","lastTransitionTime":"2026-01-21T11:11:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:21 crc kubenswrapper[4824]: I0121 11:11:21.173226 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jr64s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9768c60d-5bee-4286-bced-076e584b04e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91158f14d2e98ac9f875a5e7101df3107ef28aa5fc5d62218f460af98ff920fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flrnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:17Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jr64s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:21Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:21 crc kubenswrapper[4824]: I0121 11:11:21.185505 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" Jan 21 11:11:21 crc kubenswrapper[4824]: I0121 11:11:21.185614 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d31ab61-f94c-4f49-85d0-c9e91b83069b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a66b5c33884af00408dc35cf02f08ebd8570bd695f9f1fe558bf54fd06930431\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7514786602746dabfa690da48d48e233fe68d6dd6674aa5ecd489c0b3185f08\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61b14b6846f94cdeba8f1f61b3754e7663041f92c70356c4ee6cbb988a2e680f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a7023cfe4c098bcdd9c0477a8187f4ed7a0d2abdbe1e143436863948f441b4f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af11ad95bcf0c6bbd5fa4ab99bb592ce1c87f40d29c5ade7832414315820a569\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW0121 11:11:14.604290 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0121 11:11:14.604572 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0121 11:11:14.605875 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3284452394/tls.crt::/tmp/serving-cert-3284452394/tls.key\\\\\\\"\\\\nI0121 11:11:14.923427 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0121 11:11:14.928141 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0121 11:11:14.928162 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0121 11:11:14.928179 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0121 11:11:14.928184 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0121 11:11:14.933154 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0121 11:11:14.933230 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 11:11:14.933259 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 11:11:14.933282 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0121 11:11:14.933299 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0121 11:11:14.933315 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0121 11:11:14.933331 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0121 11:11:14.933177 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0121 11:11:14.934666 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48070c09fad511108ff22e44211cf26911509b8406d66948985cbfc3009d8233\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:21Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:21 crc kubenswrapper[4824]: I0121 11:11:21.194557 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9bb99c0a-683b-44da-9c0e-e8d284e87d53\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05e076febd697ed802ce5c7ee92a8c2c636a1a2c6609ad7eb01f17dace247c4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43529d1b9e1e5e22b0d56b023946119da8aacbae54b68c26ff0a1a82dc890c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ea232ad87393685aa323ce6393ff3869f8f8d423c2406a501c69c355535adfa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc6c7eff9e17b8a8ecc88bbfe703b5a01dbf1e02d342cede4f6a263a6526af43\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:21Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:21 crc kubenswrapper[4824]: I0121 11:11:21.204929 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:21Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:21 crc kubenswrapper[4824]: I0121 11:11:21.212148 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lv4rn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1267539-9adf-466f-8027-8ede009d19a0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eed1b884aa53e1ee318169a0cc2c9ad10130aa363ae4ac55e4bf065237756f53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbmfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lv4rn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:21Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:21 crc kubenswrapper[4824]: I0121 11:11:21.225458 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06f56d86-ec6f-4b57-b314-f508c0dd4d9f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d25c94f1ae0bbabb733f0e1ea1d1ba39562a92af73047a39aacc03013e239681\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca06234486b59cf990c69d9501e23f935328e2ba21013cd71ed64831ea749bcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b63e45d66bcd11ff3479cc3ec6b450017da8d21d36e967b8d496ae0f24b0ee9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52758bcd31f2ba64edf03adfd4d1be6428015abff5b64c7c029fa787085f9c29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2736e73342d2adb175b0721dd4f40cc42d26fca06d5eb3334927518531c3b107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b35ad7dd0570c191eb0080739304f3cd43f17b3d1dddf895549441ebe50deaa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b35ad7dd0570c191eb0080739304f3cd43f17b3d1dddf895549441ebe50deaa6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://80cd5ea5e6fc19a88c1aab5c61f827f103ab798a5c84da8dc4d45cea86d5aacb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80cd5ea5e6fc19a88c1aab5c61f827f103ab798a5c84da8dc4d45cea86d5aacb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://234d8b914e6a88480021647ae65d74e9b21c55b6788312a094d00c8a42677d48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://234d8b914e6a88480021647ae65d74e9b21c55b6788312a094d00c8a42677d48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:21Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:21 crc kubenswrapper[4824]: I0121 11:11:21.234483 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:21Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:21 crc kubenswrapper[4824]: I0121 11:11:21.244514 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86d3005cd61c0973659e1c1c0d3ef1ddd6fcf95bede5465f4f2099e64a74bb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:21Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:21 crc kubenswrapper[4824]: I0121 11:11:21.262509 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:21Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:21 crc kubenswrapper[4824]: I0121 11:11:21.270432 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:21 crc kubenswrapper[4824]: I0121 11:11:21.270459 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:21 crc kubenswrapper[4824]: I0121 11:11:21.270469 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:21 crc kubenswrapper[4824]: I0121 11:11:21.270480 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:21 crc kubenswrapper[4824]: I0121 11:11:21.270489 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:21Z","lastTransitionTime":"2026-01-21T11:11:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:21 crc kubenswrapper[4824]: I0121 11:11:21.303798 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x94nb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75d52224135b75133fc000611cd179240b9174f6492d3263341716c7e42ccb15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgr8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x94nb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:21Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:21 crc kubenswrapper[4824]: I0121 11:11:21.342711 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33f3d922-4ffe-409b-a49a-d88c85898260\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a9e57b39eed8300107e944b9fcf6a2ff1e75991f8f37387a3de615b14f9bc78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpwst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b1b37c6d2994021cfc30f85e92c40ae35677113b17aa066d52bc7a76ae5c153\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpwst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zqjg9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:21Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:21 crc kubenswrapper[4824]: I0121 11:11:21.371713 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:21 crc kubenswrapper[4824]: I0121 11:11:21.371737 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:21 crc kubenswrapper[4824]: I0121 11:11:21.371748 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:21 crc kubenswrapper[4824]: I0121 11:11:21.371765 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:21 crc kubenswrapper[4824]: I0121 11:11:21.371774 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:21Z","lastTransitionTime":"2026-01-21T11:11:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:21 crc kubenswrapper[4824]: I0121 11:11:21.384402 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gjtnh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a8d2cac-2d0f-4894-9bc0-2abd81699c47\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2239186f1de46481cebfcafb1cab85cd320a205b5a0dfc6dd1b321d9baedeba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b2239186f1de46481cebfcafb1cab85cd320a205b5a0dfc6dd1b321d9baedeba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0db5ecd7759d974ae1b0d8065b3b8bd4d7957fae16eb5bab2d71c2fe69243816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0db5ecd7759d974ae1b0d8065b3b8bd4d7957fae16eb5bab2d71c2fe69243816\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba9b3278c208fc7aac630ca7a0daa3de0823f85d0799d6f1f278da80ae14c862\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba9b3278c208fc7aac630ca7a0daa3de0823f85d0799d6f1f278da80ae14c862\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79b8f6c016b43c3ac3ba7b22cca9a6b72c262cb17ddd6236c5d4970ecc09680c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79b8f6c016b43c3ac3ba7b22cca9a6b72c262cb17ddd6236c5d4970ecc09680c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gjtnh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:21Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:21 crc kubenswrapper[4824]: I0121 11:11:21.423171 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d87422473838f5218c9fb646ab2515f2f42592754ffdf07b19f7a8e093fa48a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:21Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:21 crc kubenswrapper[4824]: I0121 11:11:21.462735 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://516c009d4720ab5b393c7482872f12fed22d4f4414bd1f3400411c0e8da88ce4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee18d293036971832c8d9c3f7b9eca84ab779ff6768a2db3c6f7bdcf3433cdef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:21Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:21 crc kubenswrapper[4824]: I0121 11:11:21.473909 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:21 crc kubenswrapper[4824]: I0121 11:11:21.473931 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:21 crc kubenswrapper[4824]: I0121 11:11:21.473940 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:21 crc kubenswrapper[4824]: I0121 11:11:21.474009 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:21 crc kubenswrapper[4824]: I0121 11:11:21.474022 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:21Z","lastTransitionTime":"2026-01-21T11:11:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:21 crc kubenswrapper[4824]: I0121 11:11:21.507370 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"26b1108a-4b98-4e39-a3cd-e0d055089fd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-htkvk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:21Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:21 crc kubenswrapper[4824]: I0121 11:11:21.546922 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06f56d86-ec6f-4b57-b314-f508c0dd4d9f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d25c94f1ae0bbabb733f0e1ea1d1ba39562a92af73047a39aacc03013e239681\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca06234486b59cf990c69d9501e23f935328e2ba21013cd71ed64831ea749bcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b63e45d66bcd11ff3479cc3ec6b450017da8d21d36e967b8d496ae0f24b0ee9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52758bcd31f2ba64edf03adfd4d1be6428015abff5b64c7c029fa787085f9c29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2736e73342d2adb175b0721dd4f40cc42d26fca06d5eb3334927518531c3b107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b35ad7dd0570c191eb0080739304f3cd43f17b3d1dddf895549441ebe50deaa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b35ad7dd0570c191eb0080739304f3cd43f17b3d1dddf895549441ebe50deaa6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://80cd5ea5e6fc19a88c1aab5c61f827f103ab798a5c84da8dc4d45cea86d5aacb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80cd5ea5e6fc19a88c1aab5c61f827f103ab798a5c84da8dc4d45cea86d5aacb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://234d8b914e6a88480021647ae65d74e9b21c55b6788312a094d00c8a42677d48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://234d8b914e6a88480021647ae65d74e9b21c55b6788312a094d00c8a42677d48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:21Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:21 crc kubenswrapper[4824]: I0121 11:11:21.576150 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:21 crc kubenswrapper[4824]: I0121 11:11:21.576172 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:21 crc kubenswrapper[4824]: I0121 11:11:21.576181 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:21 crc kubenswrapper[4824]: I0121 11:11:21.576193 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:21 crc kubenswrapper[4824]: I0121 11:11:21.576202 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:21Z","lastTransitionTime":"2026-01-21T11:11:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:21 crc kubenswrapper[4824]: I0121 11:11:21.583505 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:21Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:21 crc kubenswrapper[4824]: I0121 11:11:21.623609 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86d3005cd61c0973659e1c1c0d3ef1ddd6fcf95bede5465f4f2099e64a74bb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:21Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:21 crc kubenswrapper[4824]: I0121 11:11:21.663913 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:21Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:21 crc kubenswrapper[4824]: I0121 11:11:21.678523 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:21 crc kubenswrapper[4824]: I0121 11:11:21.678560 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:21 crc kubenswrapper[4824]: I0121 11:11:21.678569 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:21 crc kubenswrapper[4824]: I0121 11:11:21.678582 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:21 crc kubenswrapper[4824]: I0121 11:11:21.678590 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:21Z","lastTransitionTime":"2026-01-21T11:11:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:21 crc kubenswrapper[4824]: I0121 11:11:21.703252 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x94nb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75d52224135b75133fc000611cd179240b9174f6492d3263341716c7e42ccb15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgr8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x94nb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:21Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:21 crc kubenswrapper[4824]: I0121 11:11:21.742138 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33f3d922-4ffe-409b-a49a-d88c85898260\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a9e57b39eed8300107e944b9fcf6a2ff1e75991f8f37387a3de615b14f9bc78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpwst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b1b37c6d2994021cfc30f85e92c40ae35677113b17aa066d52bc7a76ae5c153\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpwst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zqjg9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:21Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:21 crc kubenswrapper[4824]: I0121 11:11:21.779915 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:21 crc kubenswrapper[4824]: I0121 11:11:21.779940 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:21 crc kubenswrapper[4824]: I0121 11:11:21.779948 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:21 crc kubenswrapper[4824]: I0121 11:11:21.779974 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:21 crc kubenswrapper[4824]: I0121 11:11:21.779982 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:21Z","lastTransitionTime":"2026-01-21T11:11:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:21 crc kubenswrapper[4824]: I0121 11:11:21.784142 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gjtnh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a8d2cac-2d0f-4894-9bc0-2abd81699c47\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2239186f1de46481cebfcafb1cab85cd320a205b5a0dfc6dd1b321d9baedeba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b2239186f1de46481cebfcafb1cab85cd320a205b5a0dfc6dd1b321d9baedeba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0db5ecd7759d974ae1b0d8065b3b8bd4d7957fae16eb5bab2d71c2fe69243816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0db5ecd7759d974ae1b0d8065b3b8bd4d7957fae16eb5bab2d71c2fe69243816\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba9b3278c208fc7aac630ca7a0daa3de0823f85d0799d6f1f278da80ae14c862\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba9b3278c208fc7aac630ca7a0daa3de0823f85d0799d6f1f278da80ae14c862\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79b8f6c016b43c3ac3ba7b22cca9a6b72c262cb17ddd6236c5d4970ecc09680c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79b8f6c016b43c3ac3ba7b22cca9a6b72c262cb17ddd6236c5d4970ecc09680c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gjtnh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:21Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:21 crc kubenswrapper[4824]: I0121 11:11:21.822821 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d87422473838f5218c9fb646ab2515f2f42592754ffdf07b19f7a8e093fa48a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:21Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:21 crc kubenswrapper[4824]: I0121 11:11:21.863025 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://516c009d4720ab5b393c7482872f12fed22d4f4414bd1f3400411c0e8da88ce4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee18d293036971832c8d9c3f7b9eca84ab779ff6768a2db3c6f7bdcf3433cdef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:21Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:21 crc kubenswrapper[4824]: I0121 11:11:21.881891 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:21 crc kubenswrapper[4824]: I0121 11:11:21.881923 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:21 crc kubenswrapper[4824]: I0121 11:11:21.881932 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:21 crc kubenswrapper[4824]: I0121 11:11:21.881944 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:21 crc kubenswrapper[4824]: I0121 11:11:21.881952 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:21Z","lastTransitionTime":"2026-01-21T11:11:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:21 crc kubenswrapper[4824]: I0121 11:11:21.905659 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"26b1108a-4b98-4e39-a3cd-e0d055089fd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc319bbc06568cc811ded2b418b04b986a31f321fed492e8be50104af79bb30b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cee9cc9eec3fc815f1621a22a37805691288dbb60dd3ff64194a84be62abcddc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc4c96e33e069f7d08d33caf92cd0987668506f931fd134b246d5ce5e934f95c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ddbb0e5ae02e9d5d82a80ead7efc91f1b8e557a9fb6f39940e5530fae8de2d38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://64a9065254f889b1c43d63759283335bf78fc4fa1531a4baae472c06c075dae3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e94ef8a8f47dc75ecfbab2e3fb8e7097c1b73eed32d7ca947f11ea198cea880\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5e832a8a65f1a11316259d61f077c9b3a22a03e96c05118dcfeaec1d72ec973\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948e0497eaffa7b0d84b4085c9ed8d987eb2c6527c9a82c10e576ed9c9c1e21b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-htkvk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:21Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:21 crc kubenswrapper[4824]: I0121 11:11:21.940663 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jr64s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9768c60d-5bee-4286-bced-076e584b04e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91158f14d2e98ac9f875a5e7101df3107ef28aa5fc5d62218f460af98ff920fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flrnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:17Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jr64s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:21Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:21 crc kubenswrapper[4824]: I0121 11:11:21.983401 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:21 crc kubenswrapper[4824]: I0121 11:11:21.983426 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:21 crc kubenswrapper[4824]: I0121 11:11:21.983435 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:21 crc kubenswrapper[4824]: I0121 11:11:21.983447 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:21 crc kubenswrapper[4824]: I0121 11:11:21.983456 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:21Z","lastTransitionTime":"2026-01-21T11:11:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:21 crc kubenswrapper[4824]: I0121 11:11:21.983688 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d31ab61-f94c-4f49-85d0-c9e91b83069b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a66b5c33884af00408dc35cf02f08ebd8570bd695f9f1fe558bf54fd06930431\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7514786602746dabfa690da48d48e233fe68d6dd6674aa5ecd489c0b3185f08\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61b14b6846f94cdeba8f1f61b3754e7663041f92c70356c4ee6cbb988a2e680f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a7023cfe4c098bcdd9c0477a8187f4ed7a0d2abdbe1e143436863948f441b4f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af11ad95bcf0c6bbd5fa4ab99bb592ce1c87f40d29c5ade7832414315820a569\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW0121 11:11:14.604290 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0121 11:11:14.604572 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0121 11:11:14.605875 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3284452394/tls.crt::/tmp/serving-cert-3284452394/tls.key\\\\\\\"\\\\nI0121 11:11:14.923427 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0121 11:11:14.928141 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0121 11:11:14.928162 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0121 11:11:14.928179 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0121 11:11:14.928184 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0121 11:11:14.933154 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0121 11:11:14.933230 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 11:11:14.933259 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 11:11:14.933282 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0121 11:11:14.933299 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0121 11:11:14.933315 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0121 11:11:14.933331 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0121 11:11:14.933177 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0121 11:11:14.934666 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48070c09fad511108ff22e44211cf26911509b8406d66948985cbfc3009d8233\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:21Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:21 crc kubenswrapper[4824]: I0121 11:11:21.994698 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-24 18:50:49.442414489 +0000 UTC Jan 21 11:11:22 crc kubenswrapper[4824]: I0121 11:11:22.023508 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9bb99c0a-683b-44da-9c0e-e8d284e87d53\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05e076febd697ed802ce5c7ee92a8c2c636a1a2c6609ad7eb01f17dace247c4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43529d1b9e1e5e22b0d56b023946119da8aacbae54b68c26ff0a1a82dc890c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ea232ad87393685aa323ce6393ff3869f8f8d423c2406a501c69c355535adfa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc6c7eff9e17b8a8ecc88bbfe703b5a01dbf1e02d342cede4f6a263a6526af43\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:22Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:22 crc kubenswrapper[4824]: I0121 11:11:22.048610 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 11:11:22 crc kubenswrapper[4824]: E0121 11:11:22.048695 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 21 11:11:22 crc kubenswrapper[4824]: I0121 11:11:22.062173 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:22Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:22 crc kubenswrapper[4824]: I0121 11:11:22.084928 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:22 crc kubenswrapper[4824]: I0121 11:11:22.084982 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:22 crc kubenswrapper[4824]: I0121 11:11:22.084993 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:22 crc kubenswrapper[4824]: I0121 11:11:22.085006 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:22 crc kubenswrapper[4824]: I0121 11:11:22.085015 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:22Z","lastTransitionTime":"2026-01-21T11:11:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:22 crc kubenswrapper[4824]: I0121 11:11:22.102374 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lv4rn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1267539-9adf-466f-8027-8ede009d19a0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eed1b884aa53e1ee318169a0cc2c9ad10130aa363ae4ac55e4bf065237756f53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbmfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lv4rn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:22Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:22 crc kubenswrapper[4824]: I0121 11:11:22.173584 4824 generic.go:334] "Generic (PLEG): container finished" podID="3a8d2cac-2d0f-4894-9bc0-2abd81699c47" containerID="38c4f218eb2ca7ea0c92bd0ae2ae9efeaa40746751608fbcf7f6c3f8d9e3c758" exitCode=0 Jan 21 11:11:22 crc kubenswrapper[4824]: I0121 11:11:22.173636 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-gjtnh" event={"ID":"3a8d2cac-2d0f-4894-9bc0-2abd81699c47","Type":"ContainerDied","Data":"38c4f218eb2ca7ea0c92bd0ae2ae9efeaa40746751608fbcf7f6c3f8d9e3c758"} Jan 21 11:11:22 crc kubenswrapper[4824]: I0121 11:11:22.173702 4824 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Jan 21 11:11:22 crc kubenswrapper[4824]: I0121 11:11:22.174031 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" Jan 21 11:11:22 crc kubenswrapper[4824]: I0121 11:11:22.186997 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:22 crc kubenswrapper[4824]: I0121 11:11:22.187019 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:22 crc kubenswrapper[4824]: I0121 11:11:22.187027 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:22 crc kubenswrapper[4824]: I0121 11:11:22.187039 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:22 crc kubenswrapper[4824]: I0121 11:11:22.187061 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:22Z","lastTransitionTime":"2026-01-21T11:11:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:22 crc kubenswrapper[4824]: I0121 11:11:22.191672 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06f56d86-ec6f-4b57-b314-f508c0dd4d9f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d25c94f1ae0bbabb733f0e1ea1d1ba39562a92af73047a39aacc03013e239681\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca06234486b59cf990c69d9501e23f935328e2ba21013cd71ed64831ea749bcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b63e45d66bcd11ff3479cc3ec6b450017da8d21d36e967b8d496ae0f24b0ee9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52758bcd31f2ba64edf03adfd4d1be6428015abff5b64c7c029fa787085f9c29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2736e73342d2adb175b0721dd4f40cc42d26fca06d5eb3334927518531c3b107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b35ad7dd0570c191eb0080739304f3cd43f17b3d1dddf895549441ebe50deaa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b35ad7dd0570c191eb0080739304f3cd43f17b3d1dddf895549441ebe50deaa6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://80cd5ea5e6fc19a88c1aab5c61f827f103ab798a5c84da8dc4d45cea86d5aacb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80cd5ea5e6fc19a88c1aab5c61f827f103ab798a5c84da8dc4d45cea86d5aacb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://234d8b914e6a88480021647ae65d74e9b21c55b6788312a094d00c8a42677d48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://234d8b914e6a88480021647ae65d74e9b21c55b6788312a094d00c8a42677d48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:22Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:22 crc kubenswrapper[4824]: I0121 11:11:22.194841 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" Jan 21 11:11:22 crc kubenswrapper[4824]: I0121 11:11:22.199930 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:22Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:22 crc kubenswrapper[4824]: I0121 11:11:22.248374 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x94nb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75d52224135b75133fc000611cd179240b9174f6492d3263341716c7e42ccb15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgr8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x94nb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:22Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:22 crc kubenswrapper[4824]: I0121 11:11:22.265498 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33f3d922-4ffe-409b-a49a-d88c85898260\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a9e57b39eed8300107e944b9fcf6a2ff1e75991f8f37387a3de615b14f9bc78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpwst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b1b37c6d2994021cfc30f85e92c40ae35677113b17aa066d52bc7a76ae5c153\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpwst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zqjg9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:22Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:22 crc kubenswrapper[4824]: I0121 11:11:22.289115 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:22 crc kubenswrapper[4824]: I0121 11:11:22.289156 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:22 crc kubenswrapper[4824]: I0121 11:11:22.289170 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:22 crc kubenswrapper[4824]: I0121 11:11:22.289186 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:22 crc kubenswrapper[4824]: I0121 11:11:22.289196 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:22Z","lastTransitionTime":"2026-01-21T11:11:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:22 crc kubenswrapper[4824]: I0121 11:11:22.305319 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gjtnh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a8d2cac-2d0f-4894-9bc0-2abd81699c47\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2239186f1de46481cebfcafb1cab85cd320a205b5a0dfc6dd1b321d9baedeba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b2239186f1de46481cebfcafb1cab85cd320a205b5a0dfc6dd1b321d9baedeba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0db5ecd7759d974ae1b0d8065b3b8bd4d7957fae16eb5bab2d71c2fe69243816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0db5ecd7759d974ae1b0d8065b3b8bd4d7957fae16eb5bab2d71c2fe69243816\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba9b3278c208fc7aac630ca7a0daa3de0823f85d0799d6f1f278da80ae14c862\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba9b3278c208fc7aac630ca7a0daa3de0823f85d0799d6f1f278da80ae14c862\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79b8f6c016b43c3ac3ba7b22cca9a6b72c262cb17ddd6236c5d4970ecc09680c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79b8f6c016b43c3ac3ba7b22cca9a6b72c262cb17ddd6236c5d4970ecc09680c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38c4f218eb2ca7ea0c92bd0ae2ae9efeaa40746751608fbcf7f6c3f8d9e3c758\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38c4f218eb2ca7ea0c92bd0ae2ae9efeaa40746751608fbcf7f6c3f8d9e3c758\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gjtnh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:22Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:22 crc kubenswrapper[4824]: I0121 11:11:22.343535 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86d3005cd61c0973659e1c1c0d3ef1ddd6fcf95bede5465f4f2099e64a74bb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:22Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:22 crc kubenswrapper[4824]: I0121 11:11:22.383166 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:22Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:22 crc kubenswrapper[4824]: I0121 11:11:22.391494 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:22 crc kubenswrapper[4824]: I0121 11:11:22.391522 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:22 crc kubenswrapper[4824]: I0121 11:11:22.391530 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:22 crc kubenswrapper[4824]: I0121 11:11:22.391542 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:22 crc kubenswrapper[4824]: I0121 11:11:22.391552 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:22Z","lastTransitionTime":"2026-01-21T11:11:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:22 crc kubenswrapper[4824]: I0121 11:11:22.426979 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"26b1108a-4b98-4e39-a3cd-e0d055089fd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc319bbc06568cc811ded2b418b04b986a31f321fed492e8be50104af79bb30b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cee9cc9eec3fc815f1621a22a37805691288dbb60dd3ff64194a84be62abcddc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc4c96e33e069f7d08d33caf92cd0987668506f931fd134b246d5ce5e934f95c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ddbb0e5ae02e9d5d82a80ead7efc91f1b8e557a9fb6f39940e5530fae8de2d38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://64a9065254f889b1c43d63759283335bf78fc4fa1531a4baae472c06c075dae3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e94ef8a8f47dc75ecfbab2e3fb8e7097c1b73eed32d7ca947f11ea198cea880\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5e832a8a65f1a11316259d61f077c9b3a22a03e96c05118dcfeaec1d72ec973\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948e0497eaffa7b0d84b4085c9ed8d987eb2c6527c9a82c10e576ed9c9c1e21b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-htkvk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:22Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:22 crc kubenswrapper[4824]: I0121 11:11:22.463415 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d87422473838f5218c9fb646ab2515f2f42592754ffdf07b19f7a8e093fa48a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:22Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:22 crc kubenswrapper[4824]: I0121 11:11:22.493985 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:22 crc kubenswrapper[4824]: I0121 11:11:22.494014 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:22 crc kubenswrapper[4824]: I0121 11:11:22.494022 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:22 crc kubenswrapper[4824]: I0121 11:11:22.494036 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:22 crc kubenswrapper[4824]: I0121 11:11:22.494056 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:22Z","lastTransitionTime":"2026-01-21T11:11:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:22 crc kubenswrapper[4824]: I0121 11:11:22.503884 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://516c009d4720ab5b393c7482872f12fed22d4f4414bd1f3400411c0e8da88ce4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee18d293036971832c8d9c3f7b9eca84ab779ff6768a2db3c6f7bdcf3433cdef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:22Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:22 crc kubenswrapper[4824]: I0121 11:11:22.544553 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9bb99c0a-683b-44da-9c0e-e8d284e87d53\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05e076febd697ed802ce5c7ee92a8c2c636a1a2c6609ad7eb01f17dace247c4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43529d1b9e1e5e22b0d56b023946119da8aacbae54b68c26ff0a1a82dc890c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ea232ad87393685aa323ce6393ff3869f8f8d423c2406a501c69c355535adfa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc6c7eff9e17b8a8ecc88bbfe703b5a01dbf1e02d342cede4f6a263a6526af43\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:22Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:22 crc kubenswrapper[4824]: I0121 11:11:22.583564 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:22Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:22 crc kubenswrapper[4824]: I0121 11:11:22.596225 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:22 crc kubenswrapper[4824]: I0121 11:11:22.596259 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:22 crc kubenswrapper[4824]: I0121 11:11:22.596268 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:22 crc kubenswrapper[4824]: I0121 11:11:22.596281 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:22 crc kubenswrapper[4824]: I0121 11:11:22.596292 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:22Z","lastTransitionTime":"2026-01-21T11:11:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:22 crc kubenswrapper[4824]: I0121 11:11:22.622231 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lv4rn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1267539-9adf-466f-8027-8ede009d19a0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eed1b884aa53e1ee318169a0cc2c9ad10130aa363ae4ac55e4bf065237756f53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbmfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lv4rn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:22Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:22 crc kubenswrapper[4824]: I0121 11:11:22.662535 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jr64s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9768c60d-5bee-4286-bced-076e584b04e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91158f14d2e98ac9f875a5e7101df3107ef28aa5fc5d62218f460af98ff920fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flrnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:17Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jr64s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:22Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:22 crc kubenswrapper[4824]: I0121 11:11:22.667776 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 21 11:11:22 crc kubenswrapper[4824]: I0121 11:11:22.667884 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 11:11:22 crc kubenswrapper[4824]: I0121 11:11:22.667906 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 11:11:22 crc kubenswrapper[4824]: E0121 11:11:22.667929 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-21 11:11:30.667909826 +0000 UTC m=+32.960939119 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 11:11:22 crc kubenswrapper[4824]: E0121 11:11:22.667982 4824 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Jan 21 11:11:22 crc kubenswrapper[4824]: E0121 11:11:22.668026 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-21 11:11:30.668014034 +0000 UTC m=+32.961043326 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Jan 21 11:11:22 crc kubenswrapper[4824]: E0121 11:11:22.668092 4824 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 21 11:11:22 crc kubenswrapper[4824]: E0121 11:11:22.668143 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-21 11:11:30.668130364 +0000 UTC m=+32.961159656 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 21 11:11:22 crc kubenswrapper[4824]: I0121 11:11:22.698040 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:22 crc kubenswrapper[4824]: I0121 11:11:22.698090 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:22 crc kubenswrapper[4824]: I0121 11:11:22.698100 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:22 crc kubenswrapper[4824]: I0121 11:11:22.698113 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:22 crc kubenswrapper[4824]: I0121 11:11:22.698124 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:22Z","lastTransitionTime":"2026-01-21T11:11:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:22 crc kubenswrapper[4824]: I0121 11:11:22.704712 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d31ab61-f94c-4f49-85d0-c9e91b83069b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a66b5c33884af00408dc35cf02f08ebd8570bd695f9f1fe558bf54fd06930431\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7514786602746dabfa690da48d48e233fe68d6dd6674aa5ecd489c0b3185f08\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61b14b6846f94cdeba8f1f61b3754e7663041f92c70356c4ee6cbb988a2e680f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a7023cfe4c098bcdd9c0477a8187f4ed7a0d2abdbe1e143436863948f441b4f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af11ad95bcf0c6bbd5fa4ab99bb592ce1c87f40d29c5ade7832414315820a569\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW0121 11:11:14.604290 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0121 11:11:14.604572 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0121 11:11:14.605875 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3284452394/tls.crt::/tmp/serving-cert-3284452394/tls.key\\\\\\\"\\\\nI0121 11:11:14.923427 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0121 11:11:14.928141 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0121 11:11:14.928162 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0121 11:11:14.928179 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0121 11:11:14.928184 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0121 11:11:14.933154 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0121 11:11:14.933230 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 11:11:14.933259 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 11:11:14.933282 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0121 11:11:14.933299 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0121 11:11:14.933315 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0121 11:11:14.933331 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0121 11:11:14.933177 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0121 11:11:14.934666 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48070c09fad511108ff22e44211cf26911509b8406d66948985cbfc3009d8233\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:22Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:22 crc kubenswrapper[4824]: I0121 11:11:22.743548 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:22Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:22 crc kubenswrapper[4824]: I0121 11:11:22.768695 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 11:11:22 crc kubenswrapper[4824]: I0121 11:11:22.768769 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 11:11:22 crc kubenswrapper[4824]: E0121 11:11:22.768870 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 21 11:11:22 crc kubenswrapper[4824]: E0121 11:11:22.768874 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 21 11:11:22 crc kubenswrapper[4824]: E0121 11:11:22.768920 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 21 11:11:22 crc kubenswrapper[4824]: E0121 11:11:22.768934 4824 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 21 11:11:22 crc kubenswrapper[4824]: E0121 11:11:22.768890 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 21 11:11:22 crc kubenswrapper[4824]: E0121 11:11:22.768992 4824 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 21 11:11:22 crc kubenswrapper[4824]: E0121 11:11:22.769003 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-01-21 11:11:30.76898634 +0000 UTC m=+33.062015642 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 21 11:11:22 crc kubenswrapper[4824]: E0121 11:11:22.769032 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-01-21 11:11:30.769019003 +0000 UTC m=+33.062048295 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 21 11:11:22 crc kubenswrapper[4824]: I0121 11:11:22.781399 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lv4rn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1267539-9adf-466f-8027-8ede009d19a0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eed1b884aa53e1ee318169a0cc2c9ad10130aa363ae4ac55e4bf065237756f53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbmfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lv4rn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:22Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:22 crc kubenswrapper[4824]: I0121 11:11:22.799596 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:22 crc kubenswrapper[4824]: I0121 11:11:22.799624 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:22 crc kubenswrapper[4824]: I0121 11:11:22.799632 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:22 crc kubenswrapper[4824]: I0121 11:11:22.799644 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:22 crc kubenswrapper[4824]: I0121 11:11:22.799651 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:22Z","lastTransitionTime":"2026-01-21T11:11:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:22 crc kubenswrapper[4824]: I0121 11:11:22.822081 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jr64s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9768c60d-5bee-4286-bced-076e584b04e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91158f14d2e98ac9f875a5e7101df3107ef28aa5fc5d62218f460af98ff920fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flrnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:17Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jr64s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:22Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:22 crc kubenswrapper[4824]: I0121 11:11:22.868305 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d31ab61-f94c-4f49-85d0-c9e91b83069b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a66b5c33884af00408dc35cf02f08ebd8570bd695f9f1fe558bf54fd06930431\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7514786602746dabfa690da48d48e233fe68d6dd6674aa5ecd489c0b3185f08\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61b14b6846f94cdeba8f1f61b3754e7663041f92c70356c4ee6cbb988a2e680f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a7023cfe4c098bcdd9c0477a8187f4ed7a0d2abdbe1e143436863948f441b4f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af11ad95bcf0c6bbd5fa4ab99bb592ce1c87f40d29c5ade7832414315820a569\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW0121 11:11:14.604290 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0121 11:11:14.604572 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0121 11:11:14.605875 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3284452394/tls.crt::/tmp/serving-cert-3284452394/tls.key\\\\\\\"\\\\nI0121 11:11:14.923427 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0121 11:11:14.928141 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0121 11:11:14.928162 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0121 11:11:14.928179 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0121 11:11:14.928184 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0121 11:11:14.933154 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0121 11:11:14.933230 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 11:11:14.933259 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 11:11:14.933282 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0121 11:11:14.933299 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0121 11:11:14.933315 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0121 11:11:14.933331 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0121 11:11:14.933177 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0121 11:11:14.934666 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48070c09fad511108ff22e44211cf26911509b8406d66948985cbfc3009d8233\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:22Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:22 crc kubenswrapper[4824]: I0121 11:11:22.900922 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:22 crc kubenswrapper[4824]: I0121 11:11:22.900943 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:22 crc kubenswrapper[4824]: I0121 11:11:22.900953 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:22 crc kubenswrapper[4824]: I0121 11:11:22.900984 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:22 crc kubenswrapper[4824]: I0121 11:11:22.900992 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:22Z","lastTransitionTime":"2026-01-21T11:11:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:22 crc kubenswrapper[4824]: I0121 11:11:22.904417 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9bb99c0a-683b-44da-9c0e-e8d284e87d53\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05e076febd697ed802ce5c7ee92a8c2c636a1a2c6609ad7eb01f17dace247c4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43529d1b9e1e5e22b0d56b023946119da8aacbae54b68c26ff0a1a82dc890c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ea232ad87393685aa323ce6393ff3869f8f8d423c2406a501c69c355535adfa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc6c7eff9e17b8a8ecc88bbfe703b5a01dbf1e02d342cede4f6a263a6526af43\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:22Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:22 crc kubenswrapper[4824]: I0121 11:11:22.946517 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06f56d86-ec6f-4b57-b314-f508c0dd4d9f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d25c94f1ae0bbabb733f0e1ea1d1ba39562a92af73047a39aacc03013e239681\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca06234486b59cf990c69d9501e23f935328e2ba21013cd71ed64831ea749bcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b63e45d66bcd11ff3479cc3ec6b450017da8d21d36e967b8d496ae0f24b0ee9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52758bcd31f2ba64edf03adfd4d1be6428015abff5b64c7c029fa787085f9c29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2736e73342d2adb175b0721dd4f40cc42d26fca06d5eb3334927518531c3b107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b35ad7dd0570c191eb0080739304f3cd43f17b3d1dddf895549441ebe50deaa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b35ad7dd0570c191eb0080739304f3cd43f17b3d1dddf895549441ebe50deaa6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://80cd5ea5e6fc19a88c1aab5c61f827f103ab798a5c84da8dc4d45cea86d5aacb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80cd5ea5e6fc19a88c1aab5c61f827f103ab798a5c84da8dc4d45cea86d5aacb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://234d8b914e6a88480021647ae65d74e9b21c55b6788312a094d00c8a42677d48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://234d8b914e6a88480021647ae65d74e9b21c55b6788312a094d00c8a42677d48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:22Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:22 crc kubenswrapper[4824]: I0121 11:11:22.982425 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:22Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:22 crc kubenswrapper[4824]: I0121 11:11:22.995542 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-16 09:40:14.721699282 +0000 UTC Jan 21 11:11:23 crc kubenswrapper[4824]: I0121 11:11:23.003094 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:23 crc kubenswrapper[4824]: I0121 11:11:23.003127 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:23 crc kubenswrapper[4824]: I0121 11:11:23.003136 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:23 crc kubenswrapper[4824]: I0121 11:11:23.003149 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:23 crc kubenswrapper[4824]: I0121 11:11:23.003156 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:23Z","lastTransitionTime":"2026-01-21T11:11:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:23 crc kubenswrapper[4824]: I0121 11:11:23.022100 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33f3d922-4ffe-409b-a49a-d88c85898260\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a9e57b39eed8300107e944b9fcf6a2ff1e75991f8f37387a3de615b14f9bc78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpwst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b1b37c6d2994021cfc30f85e92c40ae35677113b17aa066d52bc7a76ae5c153\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpwst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zqjg9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:23Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:23 crc kubenswrapper[4824]: I0121 11:11:23.049029 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 11:11:23 crc kubenswrapper[4824]: I0121 11:11:23.049131 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 11:11:23 crc kubenswrapper[4824]: E0121 11:11:23.049249 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 21 11:11:23 crc kubenswrapper[4824]: E0121 11:11:23.049401 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 21 11:11:23 crc kubenswrapper[4824]: I0121 11:11:23.064453 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gjtnh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a8d2cac-2d0f-4894-9bc0-2abd81699c47\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2239186f1de46481cebfcafb1cab85cd320a205b5a0dfc6dd1b321d9baedeba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b2239186f1de46481cebfcafb1cab85cd320a205b5a0dfc6dd1b321d9baedeba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0db5ecd7759d974ae1b0d8065b3b8bd4d7957fae16eb5bab2d71c2fe69243816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0db5ecd7759d974ae1b0d8065b3b8bd4d7957fae16eb5bab2d71c2fe69243816\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba9b3278c208fc7aac630ca7a0daa3de0823f85d0799d6f1f278da80ae14c862\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba9b3278c208fc7aac630ca7a0daa3de0823f85d0799d6f1f278da80ae14c862\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79b8f6c016b43c3ac3ba7b22cca9a6b72c262cb17ddd6236c5d4970ecc09680c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79b8f6c016b43c3ac3ba7b22cca9a6b72c262cb17ddd6236c5d4970ecc09680c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38c4f218eb2ca7ea0c92bd0ae2ae9efeaa40746751608fbcf7f6c3f8d9e3c758\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38c4f218eb2ca7ea0c92bd0ae2ae9efeaa40746751608fbcf7f6c3f8d9e3c758\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gjtnh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:23Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:23 crc kubenswrapper[4824]: I0121 11:11:23.103407 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86d3005cd61c0973659e1c1c0d3ef1ddd6fcf95bede5465f4f2099e64a74bb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:23Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:23 crc kubenswrapper[4824]: I0121 11:11:23.104597 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:23 crc kubenswrapper[4824]: I0121 11:11:23.104626 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:23 crc kubenswrapper[4824]: I0121 11:11:23.104635 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:23 crc kubenswrapper[4824]: I0121 11:11:23.104646 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:23 crc kubenswrapper[4824]: I0121 11:11:23.104654 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:23Z","lastTransitionTime":"2026-01-21T11:11:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:23 crc kubenswrapper[4824]: I0121 11:11:23.143864 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:23Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:23 crc kubenswrapper[4824]: I0121 11:11:23.177488 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-htkvk_26b1108a-4b98-4e39-a3cd-e0d055089fd8/ovnkube-controller/0.log" Jan 21 11:11:23 crc kubenswrapper[4824]: I0121 11:11:23.179555 4824 generic.go:334] "Generic (PLEG): container finished" podID="26b1108a-4b98-4e39-a3cd-e0d055089fd8" containerID="b5e832a8a65f1a11316259d61f077c9b3a22a03e96c05118dcfeaec1d72ec973" exitCode=1 Jan 21 11:11:23 crc kubenswrapper[4824]: I0121 11:11:23.179618 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" event={"ID":"26b1108a-4b98-4e39-a3cd-e0d055089fd8","Type":"ContainerDied","Data":"b5e832a8a65f1a11316259d61f077c9b3a22a03e96c05118dcfeaec1d72ec973"} Jan 21 11:11:23 crc kubenswrapper[4824]: I0121 11:11:23.180079 4824 scope.go:117] "RemoveContainer" containerID="b5e832a8a65f1a11316259d61f077c9b3a22a03e96c05118dcfeaec1d72ec973" Jan 21 11:11:23 crc kubenswrapper[4824]: I0121 11:11:23.181889 4824 generic.go:334] "Generic (PLEG): container finished" podID="3a8d2cac-2d0f-4894-9bc0-2abd81699c47" containerID="d66ad9831d6029c8c350054525a00069e869d39c1b5638e06e9b772666bd8630" exitCode=0 Jan 21 11:11:23 crc kubenswrapper[4824]: I0121 11:11:23.181929 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-gjtnh" event={"ID":"3a8d2cac-2d0f-4894-9bc0-2abd81699c47","Type":"ContainerDied","Data":"d66ad9831d6029c8c350054525a00069e869d39c1b5638e06e9b772666bd8630"} Jan 21 11:11:23 crc kubenswrapper[4824]: I0121 11:11:23.188232 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x94nb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75d52224135b75133fc000611cd179240b9174f6492d3263341716c7e42ccb15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgr8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x94nb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:23Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:23 crc kubenswrapper[4824]: I0121 11:11:23.205908 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:23 crc kubenswrapper[4824]: I0121 11:11:23.205942 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:23 crc kubenswrapper[4824]: I0121 11:11:23.205968 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:23 crc kubenswrapper[4824]: I0121 11:11:23.205981 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:23 crc kubenswrapper[4824]: I0121 11:11:23.205990 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:23Z","lastTransitionTime":"2026-01-21T11:11:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:23 crc kubenswrapper[4824]: I0121 11:11:23.223047 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d87422473838f5218c9fb646ab2515f2f42592754ffdf07b19f7a8e093fa48a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:23Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:23 crc kubenswrapper[4824]: I0121 11:11:23.263649 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://516c009d4720ab5b393c7482872f12fed22d4f4414bd1f3400411c0e8da88ce4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee18d293036971832c8d9c3f7b9eca84ab779ff6768a2db3c6f7bdcf3433cdef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:23Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:23 crc kubenswrapper[4824]: I0121 11:11:23.306859 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"26b1108a-4b98-4e39-a3cd-e0d055089fd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc319bbc06568cc811ded2b418b04b986a31f321fed492e8be50104af79bb30b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cee9cc9eec3fc815f1621a22a37805691288dbb60dd3ff64194a84be62abcddc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc4c96e33e069f7d08d33caf92cd0987668506f931fd134b246d5ce5e934f95c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ddbb0e5ae02e9d5d82a80ead7efc91f1b8e557a9fb6f39940e5530fae8de2d38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://64a9065254f889b1c43d63759283335bf78fc4fa1531a4baae472c06c075dae3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e94ef8a8f47dc75ecfbab2e3fb8e7097c1b73eed32d7ca947f11ea198cea880\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5e832a8a65f1a11316259d61f077c9b3a22a03e96c05118dcfeaec1d72ec973\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948e0497eaffa7b0d84b4085c9ed8d987eb2c6527c9a82c10e576ed9c9c1e21b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-htkvk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:23Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:23 crc kubenswrapper[4824]: I0121 11:11:23.307159 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:23 crc kubenswrapper[4824]: I0121 11:11:23.307185 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:23 crc kubenswrapper[4824]: I0121 11:11:23.307192 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:23 crc kubenswrapper[4824]: I0121 11:11:23.307206 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:23 crc kubenswrapper[4824]: I0121 11:11:23.307214 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:23Z","lastTransitionTime":"2026-01-21T11:11:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:23 crc kubenswrapper[4824]: I0121 11:11:23.343294 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d87422473838f5218c9fb646ab2515f2f42592754ffdf07b19f7a8e093fa48a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:23Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:23 crc kubenswrapper[4824]: I0121 11:11:23.384745 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://516c009d4720ab5b393c7482872f12fed22d4f4414bd1f3400411c0e8da88ce4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee18d293036971832c8d9c3f7b9eca84ab779ff6768a2db3c6f7bdcf3433cdef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:23Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:23 crc kubenswrapper[4824]: I0121 11:11:23.409191 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:23 crc kubenswrapper[4824]: I0121 11:11:23.409240 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:23 crc kubenswrapper[4824]: I0121 11:11:23.409250 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:23 crc kubenswrapper[4824]: I0121 11:11:23.409264 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:23 crc kubenswrapper[4824]: I0121 11:11:23.409273 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:23Z","lastTransitionTime":"2026-01-21T11:11:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:23 crc kubenswrapper[4824]: I0121 11:11:23.428799 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"26b1108a-4b98-4e39-a3cd-e0d055089fd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc319bbc06568cc811ded2b418b04b986a31f321fed492e8be50104af79bb30b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cee9cc9eec3fc815f1621a22a37805691288dbb60dd3ff64194a84be62abcddc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc4c96e33e069f7d08d33caf92cd0987668506f931fd134b246d5ce5e934f95c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ddbb0e5ae02e9d5d82a80ead7efc91f1b8e557a9fb6f39940e5530fae8de2d38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://64a9065254f889b1c43d63759283335bf78fc4fa1531a4baae472c06c075dae3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e94ef8a8f47dc75ecfbab2e3fb8e7097c1b73eed32d7ca947f11ea198cea880\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5e832a8a65f1a11316259d61f077c9b3a22a03e96c05118dcfeaec1d72ec973\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5e832a8a65f1a11316259d61f077c9b3a22a03e96c05118dcfeaec1d72ec973\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-21T11:11:22Z\\\",\\\"message\\\":\\\"ices_controller.go:444] Built service openshift-operator-lifecycle-manager/olm-operator-metrics LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI0121 11:11:22.643711 6069 services_controller.go:445] Built service openshift-operator-lifecycle-manager/olm-operator-metrics LB template configs for network=default: []services.lbConfig(nil)\\\\nI0121 11:11:22.643033 6069 default_network_controller.go:776] Recording success event on pod openshift-image-registry/node-ca-jr64s\\\\nI0121 11:11:22.642992 6069 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI0121 11:11:22.643725 6069 services_controller.go:451] Built service openshift-operator-lifecycle-manager/olm-operator-metrics cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-operator-lifecycle-manager/olm-operator-metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-operator-lifecycle-manager/olm-operator-metrics\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.168\\\\\\\", Port:8443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templa\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948e0497eaffa7b0d84b4085c9ed8d987eb2c6527c9a82c10e576ed9c9c1e21b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-htkvk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:23Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:23 crc kubenswrapper[4824]: I0121 11:11:23.462465 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jr64s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9768c60d-5bee-4286-bced-076e584b04e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91158f14d2e98ac9f875a5e7101df3107ef28aa5fc5d62218f460af98ff920fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flrnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:17Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jr64s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:23Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:23 crc kubenswrapper[4824]: I0121 11:11:23.504136 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d31ab61-f94c-4f49-85d0-c9e91b83069b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a66b5c33884af00408dc35cf02f08ebd8570bd695f9f1fe558bf54fd06930431\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7514786602746dabfa690da48d48e233fe68d6dd6674aa5ecd489c0b3185f08\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61b14b6846f94cdeba8f1f61b3754e7663041f92c70356c4ee6cbb988a2e680f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a7023cfe4c098bcdd9c0477a8187f4ed7a0d2abdbe1e143436863948f441b4f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af11ad95bcf0c6bbd5fa4ab99bb592ce1c87f40d29c5ade7832414315820a569\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW0121 11:11:14.604290 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0121 11:11:14.604572 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0121 11:11:14.605875 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3284452394/tls.crt::/tmp/serving-cert-3284452394/tls.key\\\\\\\"\\\\nI0121 11:11:14.923427 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0121 11:11:14.928141 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0121 11:11:14.928162 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0121 11:11:14.928179 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0121 11:11:14.928184 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0121 11:11:14.933154 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0121 11:11:14.933230 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 11:11:14.933259 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 11:11:14.933282 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0121 11:11:14.933299 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0121 11:11:14.933315 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0121 11:11:14.933331 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0121 11:11:14.933177 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0121 11:11:14.934666 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48070c09fad511108ff22e44211cf26911509b8406d66948985cbfc3009d8233\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:23Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:23 crc kubenswrapper[4824]: I0121 11:11:23.511607 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:23 crc kubenswrapper[4824]: I0121 11:11:23.511644 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:23 crc kubenswrapper[4824]: I0121 11:11:23.511654 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:23 crc kubenswrapper[4824]: I0121 11:11:23.511668 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:23 crc kubenswrapper[4824]: I0121 11:11:23.511683 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:23Z","lastTransitionTime":"2026-01-21T11:11:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:23 crc kubenswrapper[4824]: I0121 11:11:23.543310 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9bb99c0a-683b-44da-9c0e-e8d284e87d53\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05e076febd697ed802ce5c7ee92a8c2c636a1a2c6609ad7eb01f17dace247c4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43529d1b9e1e5e22b0d56b023946119da8aacbae54b68c26ff0a1a82dc890c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ea232ad87393685aa323ce6393ff3869f8f8d423c2406a501c69c355535adfa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc6c7eff9e17b8a8ecc88bbfe703b5a01dbf1e02d342cede4f6a263a6526af43\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:23Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:23 crc kubenswrapper[4824]: I0121 11:11:23.583687 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:23Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:23 crc kubenswrapper[4824]: I0121 11:11:23.614347 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:23 crc kubenswrapper[4824]: I0121 11:11:23.614391 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:23 crc kubenswrapper[4824]: I0121 11:11:23.614406 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:23 crc kubenswrapper[4824]: I0121 11:11:23.614421 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:23 crc kubenswrapper[4824]: I0121 11:11:23.614430 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:23Z","lastTransitionTime":"2026-01-21T11:11:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:23 crc kubenswrapper[4824]: I0121 11:11:23.623509 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lv4rn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1267539-9adf-466f-8027-8ede009d19a0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eed1b884aa53e1ee318169a0cc2c9ad10130aa363ae4ac55e4bf065237756f53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbmfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lv4rn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:23Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:23 crc kubenswrapper[4824]: I0121 11:11:23.669085 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06f56d86-ec6f-4b57-b314-f508c0dd4d9f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d25c94f1ae0bbabb733f0e1ea1d1ba39562a92af73047a39aacc03013e239681\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca06234486b59cf990c69d9501e23f935328e2ba21013cd71ed64831ea749bcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b63e45d66bcd11ff3479cc3ec6b450017da8d21d36e967b8d496ae0f24b0ee9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52758bcd31f2ba64edf03adfd4d1be6428015abff5b64c7c029fa787085f9c29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2736e73342d2adb175b0721dd4f40cc42d26fca06d5eb3334927518531c3b107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b35ad7dd0570c191eb0080739304f3cd43f17b3d1dddf895549441ebe50deaa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b35ad7dd0570c191eb0080739304f3cd43f17b3d1dddf895549441ebe50deaa6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://80cd5ea5e6fc19a88c1aab5c61f827f103ab798a5c84da8dc4d45cea86d5aacb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80cd5ea5e6fc19a88c1aab5c61f827f103ab798a5c84da8dc4d45cea86d5aacb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://234d8b914e6a88480021647ae65d74e9b21c55b6788312a094d00c8a42677d48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://234d8b914e6a88480021647ae65d74e9b21c55b6788312a094d00c8a42677d48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:23Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:23 crc kubenswrapper[4824]: I0121 11:11:23.703818 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:23Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:23 crc kubenswrapper[4824]: I0121 11:11:23.716356 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:23 crc kubenswrapper[4824]: I0121 11:11:23.716390 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:23 crc kubenswrapper[4824]: I0121 11:11:23.716400 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:23 crc kubenswrapper[4824]: I0121 11:11:23.716414 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:23 crc kubenswrapper[4824]: I0121 11:11:23.716422 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:23Z","lastTransitionTime":"2026-01-21T11:11:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:23 crc kubenswrapper[4824]: I0121 11:11:23.744770 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86d3005cd61c0973659e1c1c0d3ef1ddd6fcf95bede5465f4f2099e64a74bb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:23Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:23 crc kubenswrapper[4824]: I0121 11:11:23.782858 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:23Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:23 crc kubenswrapper[4824]: I0121 11:11:23.818683 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:23 crc kubenswrapper[4824]: I0121 11:11:23.818715 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:23 crc kubenswrapper[4824]: I0121 11:11:23.818725 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:23 crc kubenswrapper[4824]: I0121 11:11:23.818738 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:23 crc kubenswrapper[4824]: I0121 11:11:23.818746 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:23Z","lastTransitionTime":"2026-01-21T11:11:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:23 crc kubenswrapper[4824]: I0121 11:11:23.823553 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x94nb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75d52224135b75133fc000611cd179240b9174f6492d3263341716c7e42ccb15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgr8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x94nb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:23Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:23 crc kubenswrapper[4824]: I0121 11:11:23.863514 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33f3d922-4ffe-409b-a49a-d88c85898260\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a9e57b39eed8300107e944b9fcf6a2ff1e75991f8f37387a3de615b14f9bc78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpwst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b1b37c6d2994021cfc30f85e92c40ae35677113b17aa066d52bc7a76ae5c153\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpwst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zqjg9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:23Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:23 crc kubenswrapper[4824]: I0121 11:11:23.904912 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gjtnh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a8d2cac-2d0f-4894-9bc0-2abd81699c47\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2239186f1de46481cebfcafb1cab85cd320a205b5a0dfc6dd1b321d9baedeba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b2239186f1de46481cebfcafb1cab85cd320a205b5a0dfc6dd1b321d9baedeba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0db5ecd7759d974ae1b0d8065b3b8bd4d7957fae16eb5bab2d71c2fe69243816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0db5ecd7759d974ae1b0d8065b3b8bd4d7957fae16eb5bab2d71c2fe69243816\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba9b3278c208fc7aac630ca7a0daa3de0823f85d0799d6f1f278da80ae14c862\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba9b3278c208fc7aac630ca7a0daa3de0823f85d0799d6f1f278da80ae14c862\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79b8f6c016b43c3ac3ba7b22cca9a6b72c262cb17ddd6236c5d4970ecc09680c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79b8f6c016b43c3ac3ba7b22cca9a6b72c262cb17ddd6236c5d4970ecc09680c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38c4f218eb2ca7ea0c92bd0ae2ae9efeaa40746751608fbcf7f6c3f8d9e3c758\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38c4f218eb2ca7ea0c92bd0ae2ae9efeaa40746751608fbcf7f6c3f8d9e3c758\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d66ad9831d6029c8c350054525a00069e869d39c1b5638e06e9b772666bd8630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d66ad9831d6029c8c350054525a00069e869d39c1b5638e06e9b772666bd8630\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gjtnh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:23Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:23 crc kubenswrapper[4824]: I0121 11:11:23.920412 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:23 crc kubenswrapper[4824]: I0121 11:11:23.920454 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:23 crc kubenswrapper[4824]: I0121 11:11:23.920463 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:23 crc kubenswrapper[4824]: I0121 11:11:23.920478 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:23 crc kubenswrapper[4824]: I0121 11:11:23.920486 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:23Z","lastTransitionTime":"2026-01-21T11:11:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:23 crc kubenswrapper[4824]: I0121 11:11:23.996101 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-09 05:26:17.249976943 +0000 UTC Jan 21 11:11:24 crc kubenswrapper[4824]: I0121 11:11:24.022868 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:24 crc kubenswrapper[4824]: I0121 11:11:24.022911 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:24 crc kubenswrapper[4824]: I0121 11:11:24.022920 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:24 crc kubenswrapper[4824]: I0121 11:11:24.022934 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:24 crc kubenswrapper[4824]: I0121 11:11:24.022941 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:24Z","lastTransitionTime":"2026-01-21T11:11:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:24 crc kubenswrapper[4824]: I0121 11:11:24.049310 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 11:11:24 crc kubenswrapper[4824]: E0121 11:11:24.049437 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 21 11:11:24 crc kubenswrapper[4824]: I0121 11:11:24.124550 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:24 crc kubenswrapper[4824]: I0121 11:11:24.124582 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:24 crc kubenswrapper[4824]: I0121 11:11:24.124590 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:24 crc kubenswrapper[4824]: I0121 11:11:24.124602 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:24 crc kubenswrapper[4824]: I0121 11:11:24.124610 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:24Z","lastTransitionTime":"2026-01-21T11:11:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:24 crc kubenswrapper[4824]: I0121 11:11:24.185109 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-htkvk_26b1108a-4b98-4e39-a3cd-e0d055089fd8/ovnkube-controller/1.log" Jan 21 11:11:24 crc kubenswrapper[4824]: I0121 11:11:24.185543 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-htkvk_26b1108a-4b98-4e39-a3cd-e0d055089fd8/ovnkube-controller/0.log" Jan 21 11:11:24 crc kubenswrapper[4824]: I0121 11:11:24.187709 4824 generic.go:334] "Generic (PLEG): container finished" podID="26b1108a-4b98-4e39-a3cd-e0d055089fd8" containerID="8e688b12ddb605e11511f11a6b41f84cd31b6065797f6b7332b1255595bd27cb" exitCode=1 Jan 21 11:11:24 crc kubenswrapper[4824]: I0121 11:11:24.187768 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" event={"ID":"26b1108a-4b98-4e39-a3cd-e0d055089fd8","Type":"ContainerDied","Data":"8e688b12ddb605e11511f11a6b41f84cd31b6065797f6b7332b1255595bd27cb"} Jan 21 11:11:24 crc kubenswrapper[4824]: I0121 11:11:24.187806 4824 scope.go:117] "RemoveContainer" containerID="b5e832a8a65f1a11316259d61f077c9b3a22a03e96c05118dcfeaec1d72ec973" Jan 21 11:11:24 crc kubenswrapper[4824]: I0121 11:11:24.188263 4824 scope.go:117] "RemoveContainer" containerID="8e688b12ddb605e11511f11a6b41f84cd31b6065797f6b7332b1255595bd27cb" Jan 21 11:11:24 crc kubenswrapper[4824]: E0121 11:11:24.188404 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-htkvk_openshift-ovn-kubernetes(26b1108a-4b98-4e39-a3cd-e0d055089fd8)\"" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" podUID="26b1108a-4b98-4e39-a3cd-e0d055089fd8" Jan 21 11:11:24 crc kubenswrapper[4824]: I0121 11:11:24.190572 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-gjtnh" event={"ID":"3a8d2cac-2d0f-4894-9bc0-2abd81699c47","Type":"ContainerStarted","Data":"67c122a26811deaa9b70d19a1a65daee318ca3af2c626712482179d72fe67875"} Jan 21 11:11:24 crc kubenswrapper[4824]: I0121 11:11:24.199391 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9bb99c0a-683b-44da-9c0e-e8d284e87d53\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05e076febd697ed802ce5c7ee92a8c2c636a1a2c6609ad7eb01f17dace247c4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43529d1b9e1e5e22b0d56b023946119da8aacbae54b68c26ff0a1a82dc890c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ea232ad87393685aa323ce6393ff3869f8f8d423c2406a501c69c355535adfa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc6c7eff9e17b8a8ecc88bbfe703b5a01dbf1e02d342cede4f6a263a6526af43\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:24Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:24 crc kubenswrapper[4824]: I0121 11:11:24.208704 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:24Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:24 crc kubenswrapper[4824]: I0121 11:11:24.216419 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lv4rn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1267539-9adf-466f-8027-8ede009d19a0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eed1b884aa53e1ee318169a0cc2c9ad10130aa363ae4ac55e4bf065237756f53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbmfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lv4rn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:24Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:24 crc kubenswrapper[4824]: I0121 11:11:24.223536 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jr64s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9768c60d-5bee-4286-bced-076e584b04e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91158f14d2e98ac9f875a5e7101df3107ef28aa5fc5d62218f460af98ff920fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flrnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:17Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jr64s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:24Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:24 crc kubenswrapper[4824]: I0121 11:11:24.225693 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:24 crc kubenswrapper[4824]: I0121 11:11:24.225717 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:24 crc kubenswrapper[4824]: I0121 11:11:24.225725 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:24 crc kubenswrapper[4824]: I0121 11:11:24.225736 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:24 crc kubenswrapper[4824]: I0121 11:11:24.225744 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:24Z","lastTransitionTime":"2026-01-21T11:11:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:24 crc kubenswrapper[4824]: I0121 11:11:24.232920 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d31ab61-f94c-4f49-85d0-c9e91b83069b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a66b5c33884af00408dc35cf02f08ebd8570bd695f9f1fe558bf54fd06930431\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7514786602746dabfa690da48d48e233fe68d6dd6674aa5ecd489c0b3185f08\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61b14b6846f94cdeba8f1f61b3754e7663041f92c70356c4ee6cbb988a2e680f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a7023cfe4c098bcdd9c0477a8187f4ed7a0d2abdbe1e143436863948f441b4f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af11ad95bcf0c6bbd5fa4ab99bb592ce1c87f40d29c5ade7832414315820a569\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW0121 11:11:14.604290 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0121 11:11:14.604572 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0121 11:11:14.605875 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3284452394/tls.crt::/tmp/serving-cert-3284452394/tls.key\\\\\\\"\\\\nI0121 11:11:14.923427 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0121 11:11:14.928141 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0121 11:11:14.928162 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0121 11:11:14.928179 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0121 11:11:14.928184 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0121 11:11:14.933154 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0121 11:11:14.933230 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 11:11:14.933259 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 11:11:14.933282 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0121 11:11:14.933299 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0121 11:11:14.933315 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0121 11:11:14.933331 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0121 11:11:14.933177 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0121 11:11:14.934666 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48070c09fad511108ff22e44211cf26911509b8406d66948985cbfc3009d8233\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:24Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:24 crc kubenswrapper[4824]: I0121 11:11:24.245620 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06f56d86-ec6f-4b57-b314-f508c0dd4d9f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d25c94f1ae0bbabb733f0e1ea1d1ba39562a92af73047a39aacc03013e239681\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca06234486b59cf990c69d9501e23f935328e2ba21013cd71ed64831ea749bcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b63e45d66bcd11ff3479cc3ec6b450017da8d21d36e967b8d496ae0f24b0ee9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52758bcd31f2ba64edf03adfd4d1be6428015abff5b64c7c029fa787085f9c29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2736e73342d2adb175b0721dd4f40cc42d26fca06d5eb3334927518531c3b107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b35ad7dd0570c191eb0080739304f3cd43f17b3d1dddf895549441ebe50deaa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b35ad7dd0570c191eb0080739304f3cd43f17b3d1dddf895549441ebe50deaa6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://80cd5ea5e6fc19a88c1aab5c61f827f103ab798a5c84da8dc4d45cea86d5aacb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80cd5ea5e6fc19a88c1aab5c61f827f103ab798a5c84da8dc4d45cea86d5aacb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://234d8b914e6a88480021647ae65d74e9b21c55b6788312a094d00c8a42677d48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://234d8b914e6a88480021647ae65d74e9b21c55b6788312a094d00c8a42677d48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:24Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:24 crc kubenswrapper[4824]: I0121 11:11:24.253617 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:24Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:24 crc kubenswrapper[4824]: I0121 11:11:24.262660 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x94nb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75d52224135b75133fc000611cd179240b9174f6492d3263341716c7e42ccb15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgr8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x94nb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:24Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:24 crc kubenswrapper[4824]: I0121 11:11:24.270445 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33f3d922-4ffe-409b-a49a-d88c85898260\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a9e57b39eed8300107e944b9fcf6a2ff1e75991f8f37387a3de615b14f9bc78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpwst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b1b37c6d2994021cfc30f85e92c40ae35677113b17aa066d52bc7a76ae5c153\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpwst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zqjg9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:24Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:24 crc kubenswrapper[4824]: I0121 11:11:24.304839 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gjtnh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a8d2cac-2d0f-4894-9bc0-2abd81699c47\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2239186f1de46481cebfcafb1cab85cd320a205b5a0dfc6dd1b321d9baedeba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b2239186f1de46481cebfcafb1cab85cd320a205b5a0dfc6dd1b321d9baedeba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0db5ecd7759d974ae1b0d8065b3b8bd4d7957fae16eb5bab2d71c2fe69243816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0db5ecd7759d974ae1b0d8065b3b8bd4d7957fae16eb5bab2d71c2fe69243816\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba9b3278c208fc7aac630ca7a0daa3de0823f85d0799d6f1f278da80ae14c862\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba9b3278c208fc7aac630ca7a0daa3de0823f85d0799d6f1f278da80ae14c862\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79b8f6c016b43c3ac3ba7b22cca9a6b72c262cb17ddd6236c5d4970ecc09680c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79b8f6c016b43c3ac3ba7b22cca9a6b72c262cb17ddd6236c5d4970ecc09680c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38c4f218eb2ca7ea0c92bd0ae2ae9efeaa40746751608fbcf7f6c3f8d9e3c758\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38c4f218eb2ca7ea0c92bd0ae2ae9efeaa40746751608fbcf7f6c3f8d9e3c758\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d66ad9831d6029c8c350054525a00069e869d39c1b5638e06e9b772666bd8630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d66ad9831d6029c8c350054525a00069e869d39c1b5638e06e9b772666bd8630\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gjtnh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:24Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:24 crc kubenswrapper[4824]: I0121 11:11:24.327532 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:24 crc kubenswrapper[4824]: I0121 11:11:24.327566 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:24 crc kubenswrapper[4824]: I0121 11:11:24.327577 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:24 crc kubenswrapper[4824]: I0121 11:11:24.327592 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:24 crc kubenswrapper[4824]: I0121 11:11:24.327601 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:24Z","lastTransitionTime":"2026-01-21T11:11:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:24 crc kubenswrapper[4824]: I0121 11:11:24.344611 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86d3005cd61c0973659e1c1c0d3ef1ddd6fcf95bede5465f4f2099e64a74bb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:24Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:24 crc kubenswrapper[4824]: I0121 11:11:24.383265 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:24Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:24 crc kubenswrapper[4824]: I0121 11:11:24.427250 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"26b1108a-4b98-4e39-a3cd-e0d055089fd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc319bbc06568cc811ded2b418b04b986a31f321fed492e8be50104af79bb30b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cee9cc9eec3fc815f1621a22a37805691288dbb60dd3ff64194a84be62abcddc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc4c96e33e069f7d08d33caf92cd0987668506f931fd134b246d5ce5e934f95c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ddbb0e5ae02e9d5d82a80ead7efc91f1b8e557a9fb6f39940e5530fae8de2d38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://64a9065254f889b1c43d63759283335bf78fc4fa1531a4baae472c06c075dae3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e94ef8a8f47dc75ecfbab2e3fb8e7097c1b73eed32d7ca947f11ea198cea880\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e688b12ddb605e11511f11a6b41f84cd31b6065797f6b7332b1255595bd27cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5e832a8a65f1a11316259d61f077c9b3a22a03e96c05118dcfeaec1d72ec973\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-21T11:11:22Z\\\",\\\"message\\\":\\\"ices_controller.go:444] Built service openshift-operator-lifecycle-manager/olm-operator-metrics LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI0121 11:11:22.643711 6069 services_controller.go:445] Built service openshift-operator-lifecycle-manager/olm-operator-metrics LB template configs for network=default: []services.lbConfig(nil)\\\\nI0121 11:11:22.643033 6069 default_network_controller.go:776] Recording success event on pod openshift-image-registry/node-ca-jr64s\\\\nI0121 11:11:22.642992 6069 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI0121 11:11:22.643725 6069 services_controller.go:451] Built service openshift-operator-lifecycle-manager/olm-operator-metrics cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-operator-lifecycle-manager/olm-operator-metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-operator-lifecycle-manager/olm-operator-metrics\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.168\\\\\\\", Port:8443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templa\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:20Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e688b12ddb605e11511f11a6b41f84cd31b6065797f6b7332b1255595bd27cb\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-21T11:11:23Z\\\",\\\"message\\\":\\\" crc\\\\nF0121 11:11:23.774847 6243 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:23Z is after 2025-08-24T17:21:41Z]\\\\nI0121 11:11:23.774842 6243 services_controller.go:451] Built service openshift-kube-storage-version-migrator-operator/metrics cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-kube-storage-version-migrator-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-storage-version-migrator-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeO\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948e0497eaffa7b0d84b4085c9ed8d987eb2c6527c9a82c10e576ed9c9c1e21b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-htkvk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:24Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:24 crc kubenswrapper[4824]: I0121 11:11:24.429942 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:24 crc kubenswrapper[4824]: I0121 11:11:24.429990 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:24 crc kubenswrapper[4824]: I0121 11:11:24.429999 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:24 crc kubenswrapper[4824]: I0121 11:11:24.430015 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:24 crc kubenswrapper[4824]: I0121 11:11:24.430024 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:24Z","lastTransitionTime":"2026-01-21T11:11:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:24 crc kubenswrapper[4824]: I0121 11:11:24.463098 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d87422473838f5218c9fb646ab2515f2f42592754ffdf07b19f7a8e093fa48a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:24Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:24 crc kubenswrapper[4824]: I0121 11:11:24.503767 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://516c009d4720ab5b393c7482872f12fed22d4f4414bd1f3400411c0e8da88ce4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee18d293036971832c8d9c3f7b9eca84ab779ff6768a2db3c6f7bdcf3433cdef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:24Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:24 crc kubenswrapper[4824]: I0121 11:11:24.531695 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:24 crc kubenswrapper[4824]: I0121 11:11:24.531727 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:24 crc kubenswrapper[4824]: I0121 11:11:24.531736 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:24 crc kubenswrapper[4824]: I0121 11:11:24.531748 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:24 crc kubenswrapper[4824]: I0121 11:11:24.531758 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:24Z","lastTransitionTime":"2026-01-21T11:11:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:24 crc kubenswrapper[4824]: I0121 11:11:24.542016 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jr64s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9768c60d-5bee-4286-bced-076e584b04e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91158f14d2e98ac9f875a5e7101df3107ef28aa5fc5d62218f460af98ff920fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flrnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:17Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jr64s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:24Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:24 crc kubenswrapper[4824]: I0121 11:11:24.585043 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d31ab61-f94c-4f49-85d0-c9e91b83069b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a66b5c33884af00408dc35cf02f08ebd8570bd695f9f1fe558bf54fd06930431\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7514786602746dabfa690da48d48e233fe68d6dd6674aa5ecd489c0b3185f08\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61b14b6846f94cdeba8f1f61b3754e7663041f92c70356c4ee6cbb988a2e680f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a7023cfe4c098bcdd9c0477a8187f4ed7a0d2abdbe1e143436863948f441b4f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af11ad95bcf0c6bbd5fa4ab99bb592ce1c87f40d29c5ade7832414315820a569\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW0121 11:11:14.604290 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0121 11:11:14.604572 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0121 11:11:14.605875 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3284452394/tls.crt::/tmp/serving-cert-3284452394/tls.key\\\\\\\"\\\\nI0121 11:11:14.923427 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0121 11:11:14.928141 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0121 11:11:14.928162 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0121 11:11:14.928179 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0121 11:11:14.928184 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0121 11:11:14.933154 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0121 11:11:14.933230 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 11:11:14.933259 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 11:11:14.933282 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0121 11:11:14.933299 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0121 11:11:14.933315 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0121 11:11:14.933331 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0121 11:11:14.933177 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0121 11:11:14.934666 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48070c09fad511108ff22e44211cf26911509b8406d66948985cbfc3009d8233\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:24Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:24 crc kubenswrapper[4824]: I0121 11:11:24.623999 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9bb99c0a-683b-44da-9c0e-e8d284e87d53\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05e076febd697ed802ce5c7ee92a8c2c636a1a2c6609ad7eb01f17dace247c4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43529d1b9e1e5e22b0d56b023946119da8aacbae54b68c26ff0a1a82dc890c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ea232ad87393685aa323ce6393ff3869f8f8d423c2406a501c69c355535adfa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc6c7eff9e17b8a8ecc88bbfe703b5a01dbf1e02d342cede4f6a263a6526af43\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:24Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:24 crc kubenswrapper[4824]: I0121 11:11:24.633496 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:24 crc kubenswrapper[4824]: I0121 11:11:24.633531 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:24 crc kubenswrapper[4824]: I0121 11:11:24.633543 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:24 crc kubenswrapper[4824]: I0121 11:11:24.633561 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:24 crc kubenswrapper[4824]: I0121 11:11:24.633572 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:24Z","lastTransitionTime":"2026-01-21T11:11:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:24 crc kubenswrapper[4824]: I0121 11:11:24.665253 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:24Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:24 crc kubenswrapper[4824]: I0121 11:11:24.703207 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lv4rn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1267539-9adf-466f-8027-8ede009d19a0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eed1b884aa53e1ee318169a0cc2c9ad10130aa363ae4ac55e4bf065237756f53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbmfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lv4rn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:24Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:24 crc kubenswrapper[4824]: I0121 11:11:24.735019 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:24 crc kubenswrapper[4824]: I0121 11:11:24.735047 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:24 crc kubenswrapper[4824]: I0121 11:11:24.735057 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:24 crc kubenswrapper[4824]: I0121 11:11:24.735080 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:24 crc kubenswrapper[4824]: I0121 11:11:24.735089 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:24Z","lastTransitionTime":"2026-01-21T11:11:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:24 crc kubenswrapper[4824]: I0121 11:11:24.747549 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06f56d86-ec6f-4b57-b314-f508c0dd4d9f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d25c94f1ae0bbabb733f0e1ea1d1ba39562a92af73047a39aacc03013e239681\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca06234486b59cf990c69d9501e23f935328e2ba21013cd71ed64831ea749bcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b63e45d66bcd11ff3479cc3ec6b450017da8d21d36e967b8d496ae0f24b0ee9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52758bcd31f2ba64edf03adfd4d1be6428015abff5b64c7c029fa787085f9c29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2736e73342d2adb175b0721dd4f40cc42d26fca06d5eb3334927518531c3b107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b35ad7dd0570c191eb0080739304f3cd43f17b3d1dddf895549441ebe50deaa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b35ad7dd0570c191eb0080739304f3cd43f17b3d1dddf895549441ebe50deaa6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://80cd5ea5e6fc19a88c1aab5c61f827f103ab798a5c84da8dc4d45cea86d5aacb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80cd5ea5e6fc19a88c1aab5c61f827f103ab798a5c84da8dc4d45cea86d5aacb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://234d8b914e6a88480021647ae65d74e9b21c55b6788312a094d00c8a42677d48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://234d8b914e6a88480021647ae65d74e9b21c55b6788312a094d00c8a42677d48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:24Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:24 crc kubenswrapper[4824]: I0121 11:11:24.783457 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:24Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:24 crc kubenswrapper[4824]: I0121 11:11:24.825627 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86d3005cd61c0973659e1c1c0d3ef1ddd6fcf95bede5465f4f2099e64a74bb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:24Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:24 crc kubenswrapper[4824]: I0121 11:11:24.836662 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:24 crc kubenswrapper[4824]: I0121 11:11:24.836692 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:24 crc kubenswrapper[4824]: I0121 11:11:24.836703 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:24 crc kubenswrapper[4824]: I0121 11:11:24.836717 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:24 crc kubenswrapper[4824]: I0121 11:11:24.836726 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:24Z","lastTransitionTime":"2026-01-21T11:11:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:24 crc kubenswrapper[4824]: I0121 11:11:24.866222 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:24Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:24 crc kubenswrapper[4824]: I0121 11:11:24.904202 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x94nb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75d52224135b75133fc000611cd179240b9174f6492d3263341716c7e42ccb15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgr8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x94nb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:24Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:24 crc kubenswrapper[4824]: I0121 11:11:24.939328 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:24 crc kubenswrapper[4824]: I0121 11:11:24.939359 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:24 crc kubenswrapper[4824]: I0121 11:11:24.939369 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:24 crc kubenswrapper[4824]: I0121 11:11:24.939381 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:24 crc kubenswrapper[4824]: I0121 11:11:24.939390 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:24Z","lastTransitionTime":"2026-01-21T11:11:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:24 crc kubenswrapper[4824]: I0121 11:11:24.943630 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33f3d922-4ffe-409b-a49a-d88c85898260\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a9e57b39eed8300107e944b9fcf6a2ff1e75991f8f37387a3de615b14f9bc78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpwst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b1b37c6d2994021cfc30f85e92c40ae35677113b17aa066d52bc7a76ae5c153\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpwst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zqjg9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:24Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:24 crc kubenswrapper[4824]: I0121 11:11:24.985166 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gjtnh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a8d2cac-2d0f-4894-9bc0-2abd81699c47\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67c122a26811deaa9b70d19a1a65daee318ca3af2c626712482179d72fe67875\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2239186f1de46481cebfcafb1cab85cd320a205b5a0dfc6dd1b321d9baedeba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b2239186f1de46481cebfcafb1cab85cd320a205b5a0dfc6dd1b321d9baedeba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0db5ecd7759d974ae1b0d8065b3b8bd4d7957fae16eb5bab2d71c2fe69243816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0db5ecd7759d974ae1b0d8065b3b8bd4d7957fae16eb5bab2d71c2fe69243816\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba9b3278c208fc7aac630ca7a0daa3de0823f85d0799d6f1f278da80ae14c862\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba9b3278c208fc7aac630ca7a0daa3de0823f85d0799d6f1f278da80ae14c862\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79b8f6c016b43c3ac3ba7b22cca9a6b72c262cb17ddd6236c5d4970ecc09680c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79b8f6c016b43c3ac3ba7b22cca9a6b72c262cb17ddd6236c5d4970ecc09680c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38c4f218eb2ca7ea0c92bd0ae2ae9efeaa40746751608fbcf7f6c3f8d9e3c758\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38c4f218eb2ca7ea0c92bd0ae2ae9efeaa40746751608fbcf7f6c3f8d9e3c758\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d66ad9831d6029c8c350054525a00069e869d39c1b5638e06e9b772666bd8630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d66ad9831d6029c8c350054525a00069e869d39c1b5638e06e9b772666bd8630\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gjtnh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:24Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:24 crc kubenswrapper[4824]: I0121 11:11:24.996308 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-11 13:03:07.313307472 +0000 UTC Jan 21 11:11:25 crc kubenswrapper[4824]: I0121 11:11:25.022325 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d87422473838f5218c9fb646ab2515f2f42592754ffdf07b19f7a8e093fa48a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:25Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:25 crc kubenswrapper[4824]: I0121 11:11:25.041309 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:25 crc kubenswrapper[4824]: I0121 11:11:25.041342 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:25 crc kubenswrapper[4824]: I0121 11:11:25.041350 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:25 crc kubenswrapper[4824]: I0121 11:11:25.041364 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:25 crc kubenswrapper[4824]: I0121 11:11:25.041373 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:25Z","lastTransitionTime":"2026-01-21T11:11:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:25 crc kubenswrapper[4824]: I0121 11:11:25.048589 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 11:11:25 crc kubenswrapper[4824]: E0121 11:11:25.048678 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 21 11:11:25 crc kubenswrapper[4824]: I0121 11:11:25.048588 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 11:11:25 crc kubenswrapper[4824]: E0121 11:11:25.048787 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 21 11:11:25 crc kubenswrapper[4824]: I0121 11:11:25.062738 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://516c009d4720ab5b393c7482872f12fed22d4f4414bd1f3400411c0e8da88ce4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee18d293036971832c8d9c3f7b9eca84ab779ff6768a2db3c6f7bdcf3433cdef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:25Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:25 crc kubenswrapper[4824]: I0121 11:11:25.105751 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"26b1108a-4b98-4e39-a3cd-e0d055089fd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc319bbc06568cc811ded2b418b04b986a31f321fed492e8be50104af79bb30b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cee9cc9eec3fc815f1621a22a37805691288dbb60dd3ff64194a84be62abcddc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc4c96e33e069f7d08d33caf92cd0987668506f931fd134b246d5ce5e934f95c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ddbb0e5ae02e9d5d82a80ead7efc91f1b8e557a9fb6f39940e5530fae8de2d38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://64a9065254f889b1c43d63759283335bf78fc4fa1531a4baae472c06c075dae3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e94ef8a8f47dc75ecfbab2e3fb8e7097c1b73eed32d7ca947f11ea198cea880\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e688b12ddb605e11511f11a6b41f84cd31b6065797f6b7332b1255595bd27cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5e832a8a65f1a11316259d61f077c9b3a22a03e96c05118dcfeaec1d72ec973\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-21T11:11:22Z\\\",\\\"message\\\":\\\"ices_controller.go:444] Built service openshift-operator-lifecycle-manager/olm-operator-metrics LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI0121 11:11:22.643711 6069 services_controller.go:445] Built service openshift-operator-lifecycle-manager/olm-operator-metrics LB template configs for network=default: []services.lbConfig(nil)\\\\nI0121 11:11:22.643033 6069 default_network_controller.go:776] Recording success event on pod openshift-image-registry/node-ca-jr64s\\\\nI0121 11:11:22.642992 6069 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI0121 11:11:22.643725 6069 services_controller.go:451] Built service openshift-operator-lifecycle-manager/olm-operator-metrics cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-operator-lifecycle-manager/olm-operator-metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-operator-lifecycle-manager/olm-operator-metrics\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.168\\\\\\\", Port:8443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templa\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:20Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e688b12ddb605e11511f11a6b41f84cd31b6065797f6b7332b1255595bd27cb\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-21T11:11:23Z\\\",\\\"message\\\":\\\" crc\\\\nF0121 11:11:23.774847 6243 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:23Z is after 2025-08-24T17:21:41Z]\\\\nI0121 11:11:23.774842 6243 services_controller.go:451] Built service openshift-kube-storage-version-migrator-operator/metrics cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-kube-storage-version-migrator-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-storage-version-migrator-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeO\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948e0497eaffa7b0d84b4085c9ed8d987eb2c6527c9a82c10e576ed9c9c1e21b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-htkvk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:25Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:25 crc kubenswrapper[4824]: I0121 11:11:25.143330 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:25 crc kubenswrapper[4824]: I0121 11:11:25.143420 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:25 crc kubenswrapper[4824]: I0121 11:11:25.143441 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:25 crc kubenswrapper[4824]: I0121 11:11:25.143468 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:25 crc kubenswrapper[4824]: I0121 11:11:25.143483 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:25Z","lastTransitionTime":"2026-01-21T11:11:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:25 crc kubenswrapper[4824]: I0121 11:11:25.194697 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-htkvk_26b1108a-4b98-4e39-a3cd-e0d055089fd8/ovnkube-controller/1.log" Jan 21 11:11:25 crc kubenswrapper[4824]: I0121 11:11:25.198278 4824 scope.go:117] "RemoveContainer" containerID="8e688b12ddb605e11511f11a6b41f84cd31b6065797f6b7332b1255595bd27cb" Jan 21 11:11:25 crc kubenswrapper[4824]: E0121 11:11:25.198411 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-htkvk_openshift-ovn-kubernetes(26b1108a-4b98-4e39-a3cd-e0d055089fd8)\"" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" podUID="26b1108a-4b98-4e39-a3cd-e0d055089fd8" Jan 21 11:11:25 crc kubenswrapper[4824]: I0121 11:11:25.207826 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9bb99c0a-683b-44da-9c0e-e8d284e87d53\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05e076febd697ed802ce5c7ee92a8c2c636a1a2c6609ad7eb01f17dace247c4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43529d1b9e1e5e22b0d56b023946119da8aacbae54b68c26ff0a1a82dc890c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ea232ad87393685aa323ce6393ff3869f8f8d423c2406a501c69c355535adfa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc6c7eff9e17b8a8ecc88bbfe703b5a01dbf1e02d342cede4f6a263a6526af43\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:25Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:25 crc kubenswrapper[4824]: I0121 11:11:25.216837 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:25Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:25 crc kubenswrapper[4824]: I0121 11:11:25.223654 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lv4rn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1267539-9adf-466f-8027-8ede009d19a0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eed1b884aa53e1ee318169a0cc2c9ad10130aa363ae4ac55e4bf065237756f53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbmfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lv4rn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:25Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:25 crc kubenswrapper[4824]: I0121 11:11:25.245334 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:25 crc kubenswrapper[4824]: I0121 11:11:25.245369 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:25 crc kubenswrapper[4824]: I0121 11:11:25.245378 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:25 crc kubenswrapper[4824]: I0121 11:11:25.245390 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:25 crc kubenswrapper[4824]: I0121 11:11:25.245400 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:25Z","lastTransitionTime":"2026-01-21T11:11:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:25 crc kubenswrapper[4824]: I0121 11:11:25.262241 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jr64s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9768c60d-5bee-4286-bced-076e584b04e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91158f14d2e98ac9f875a5e7101df3107ef28aa5fc5d62218f460af98ff920fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flrnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:17Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jr64s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:25Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:25 crc kubenswrapper[4824]: I0121 11:11:25.302953 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d31ab61-f94c-4f49-85d0-c9e91b83069b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a66b5c33884af00408dc35cf02f08ebd8570bd695f9f1fe558bf54fd06930431\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7514786602746dabfa690da48d48e233fe68d6dd6674aa5ecd489c0b3185f08\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61b14b6846f94cdeba8f1f61b3754e7663041f92c70356c4ee6cbb988a2e680f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a7023cfe4c098bcdd9c0477a8187f4ed7a0d2abdbe1e143436863948f441b4f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af11ad95bcf0c6bbd5fa4ab99bb592ce1c87f40d29c5ade7832414315820a569\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW0121 11:11:14.604290 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0121 11:11:14.604572 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0121 11:11:14.605875 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3284452394/tls.crt::/tmp/serving-cert-3284452394/tls.key\\\\\\\"\\\\nI0121 11:11:14.923427 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0121 11:11:14.928141 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0121 11:11:14.928162 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0121 11:11:14.928179 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0121 11:11:14.928184 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0121 11:11:14.933154 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0121 11:11:14.933230 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 11:11:14.933259 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 11:11:14.933282 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0121 11:11:14.933299 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0121 11:11:14.933315 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0121 11:11:14.933331 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0121 11:11:14.933177 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0121 11:11:14.934666 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48070c09fad511108ff22e44211cf26911509b8406d66948985cbfc3009d8233\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:25Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:25 crc kubenswrapper[4824]: I0121 11:11:25.347124 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06f56d86-ec6f-4b57-b314-f508c0dd4d9f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d25c94f1ae0bbabb733f0e1ea1d1ba39562a92af73047a39aacc03013e239681\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca06234486b59cf990c69d9501e23f935328e2ba21013cd71ed64831ea749bcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b63e45d66bcd11ff3479cc3ec6b450017da8d21d36e967b8d496ae0f24b0ee9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52758bcd31f2ba64edf03adfd4d1be6428015abff5b64c7c029fa787085f9c29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2736e73342d2adb175b0721dd4f40cc42d26fca06d5eb3334927518531c3b107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b35ad7dd0570c191eb0080739304f3cd43f17b3d1dddf895549441ebe50deaa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b35ad7dd0570c191eb0080739304f3cd43f17b3d1dddf895549441ebe50deaa6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://80cd5ea5e6fc19a88c1aab5c61f827f103ab798a5c84da8dc4d45cea86d5aacb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80cd5ea5e6fc19a88c1aab5c61f827f103ab798a5c84da8dc4d45cea86d5aacb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://234d8b914e6a88480021647ae65d74e9b21c55b6788312a094d00c8a42677d48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://234d8b914e6a88480021647ae65d74e9b21c55b6788312a094d00c8a42677d48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:25Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:25 crc kubenswrapper[4824]: I0121 11:11:25.347620 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:25 crc kubenswrapper[4824]: I0121 11:11:25.347657 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:25 crc kubenswrapper[4824]: I0121 11:11:25.347670 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:25 crc kubenswrapper[4824]: I0121 11:11:25.347684 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:25 crc kubenswrapper[4824]: I0121 11:11:25.347695 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:25Z","lastTransitionTime":"2026-01-21T11:11:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:25 crc kubenswrapper[4824]: I0121 11:11:25.384235 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:25Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:25 crc kubenswrapper[4824]: I0121 11:11:25.422528 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x94nb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75d52224135b75133fc000611cd179240b9174f6492d3263341716c7e42ccb15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgr8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x94nb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:25Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:25 crc kubenswrapper[4824]: I0121 11:11:25.449392 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:25 crc kubenswrapper[4824]: I0121 11:11:25.449436 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:25 crc kubenswrapper[4824]: I0121 11:11:25.449447 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:25 crc kubenswrapper[4824]: I0121 11:11:25.449465 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:25 crc kubenswrapper[4824]: I0121 11:11:25.449475 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:25Z","lastTransitionTime":"2026-01-21T11:11:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:25 crc kubenswrapper[4824]: I0121 11:11:25.463093 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33f3d922-4ffe-409b-a49a-d88c85898260\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a9e57b39eed8300107e944b9fcf6a2ff1e75991f8f37387a3de615b14f9bc78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpwst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b1b37c6d2994021cfc30f85e92c40ae35677113b17aa066d52bc7a76ae5c153\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpwst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zqjg9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:25Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:25 crc kubenswrapper[4824]: I0121 11:11:25.504364 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gjtnh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a8d2cac-2d0f-4894-9bc0-2abd81699c47\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67c122a26811deaa9b70d19a1a65daee318ca3af2c626712482179d72fe67875\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2239186f1de46481cebfcafb1cab85cd320a205b5a0dfc6dd1b321d9baedeba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b2239186f1de46481cebfcafb1cab85cd320a205b5a0dfc6dd1b321d9baedeba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0db5ecd7759d974ae1b0d8065b3b8bd4d7957fae16eb5bab2d71c2fe69243816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0db5ecd7759d974ae1b0d8065b3b8bd4d7957fae16eb5bab2d71c2fe69243816\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba9b3278c208fc7aac630ca7a0daa3de0823f85d0799d6f1f278da80ae14c862\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba9b3278c208fc7aac630ca7a0daa3de0823f85d0799d6f1f278da80ae14c862\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79b8f6c016b43c3ac3ba7b22cca9a6b72c262cb17ddd6236c5d4970ecc09680c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79b8f6c016b43c3ac3ba7b22cca9a6b72c262cb17ddd6236c5d4970ecc09680c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38c4f218eb2ca7ea0c92bd0ae2ae9efeaa40746751608fbcf7f6c3f8d9e3c758\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38c4f218eb2ca7ea0c92bd0ae2ae9efeaa40746751608fbcf7f6c3f8d9e3c758\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d66ad9831d6029c8c350054525a00069e869d39c1b5638e06e9b772666bd8630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d66ad9831d6029c8c350054525a00069e869d39c1b5638e06e9b772666bd8630\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gjtnh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:25Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:25 crc kubenswrapper[4824]: I0121 11:11:25.543869 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86d3005cd61c0973659e1c1c0d3ef1ddd6fcf95bede5465f4f2099e64a74bb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:25Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:25 crc kubenswrapper[4824]: I0121 11:11:25.551235 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:25 crc kubenswrapper[4824]: I0121 11:11:25.551280 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:25 crc kubenswrapper[4824]: I0121 11:11:25.551290 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:25 crc kubenswrapper[4824]: I0121 11:11:25.551307 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:25 crc kubenswrapper[4824]: I0121 11:11:25.551317 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:25Z","lastTransitionTime":"2026-01-21T11:11:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:25 crc kubenswrapper[4824]: I0121 11:11:25.582519 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:25Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:25 crc kubenswrapper[4824]: I0121 11:11:25.626050 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"26b1108a-4b98-4e39-a3cd-e0d055089fd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc319bbc06568cc811ded2b418b04b986a31f321fed492e8be50104af79bb30b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cee9cc9eec3fc815f1621a22a37805691288dbb60dd3ff64194a84be62abcddc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc4c96e33e069f7d08d33caf92cd0987668506f931fd134b246d5ce5e934f95c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ddbb0e5ae02e9d5d82a80ead7efc91f1b8e557a9fb6f39940e5530fae8de2d38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://64a9065254f889b1c43d63759283335bf78fc4fa1531a4baae472c06c075dae3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e94ef8a8f47dc75ecfbab2e3fb8e7097c1b73eed32d7ca947f11ea198cea880\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e688b12ddb605e11511f11a6b41f84cd31b6065797f6b7332b1255595bd27cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e688b12ddb605e11511f11a6b41f84cd31b6065797f6b7332b1255595bd27cb\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-21T11:11:23Z\\\",\\\"message\\\":\\\" crc\\\\nF0121 11:11:23.774847 6243 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:23Z is after 2025-08-24T17:21:41Z]\\\\nI0121 11:11:23.774842 6243 services_controller.go:451] Built service openshift-kube-storage-version-migrator-operator/metrics cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-kube-storage-version-migrator-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-storage-version-migrator-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeO\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:23Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-htkvk_openshift-ovn-kubernetes(26b1108a-4b98-4e39-a3cd-e0d055089fd8)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948e0497eaffa7b0d84b4085c9ed8d987eb2c6527c9a82c10e576ed9c9c1e21b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-htkvk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:25Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:25 crc kubenswrapper[4824]: I0121 11:11:25.652835 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:25 crc kubenswrapper[4824]: I0121 11:11:25.652866 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:25 crc kubenswrapper[4824]: I0121 11:11:25.652875 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:25 crc kubenswrapper[4824]: I0121 11:11:25.652894 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:25 crc kubenswrapper[4824]: I0121 11:11:25.652903 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:25Z","lastTransitionTime":"2026-01-21T11:11:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:25 crc kubenswrapper[4824]: I0121 11:11:25.662841 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d87422473838f5218c9fb646ab2515f2f42592754ffdf07b19f7a8e093fa48a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:25Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:25 crc kubenswrapper[4824]: I0121 11:11:25.702938 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://516c009d4720ab5b393c7482872f12fed22d4f4414bd1f3400411c0e8da88ce4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee18d293036971832c8d9c3f7b9eca84ab779ff6768a2db3c6f7bdcf3433cdef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:25Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:25 crc kubenswrapper[4824]: I0121 11:11:25.754417 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:25 crc kubenswrapper[4824]: I0121 11:11:25.754457 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:25 crc kubenswrapper[4824]: I0121 11:11:25.754466 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:25 crc kubenswrapper[4824]: I0121 11:11:25.754479 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:25 crc kubenswrapper[4824]: I0121 11:11:25.754487 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:25Z","lastTransitionTime":"2026-01-21T11:11:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:25 crc kubenswrapper[4824]: I0121 11:11:25.856269 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:25 crc kubenswrapper[4824]: I0121 11:11:25.856304 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:25 crc kubenswrapper[4824]: I0121 11:11:25.856314 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:25 crc kubenswrapper[4824]: I0121 11:11:25.856326 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:25 crc kubenswrapper[4824]: I0121 11:11:25.856337 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:25Z","lastTransitionTime":"2026-01-21T11:11:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:25 crc kubenswrapper[4824]: I0121 11:11:25.958267 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:25 crc kubenswrapper[4824]: I0121 11:11:25.958306 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:25 crc kubenswrapper[4824]: I0121 11:11:25.958315 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:25 crc kubenswrapper[4824]: I0121 11:11:25.958339 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:25 crc kubenswrapper[4824]: I0121 11:11:25.958352 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:25Z","lastTransitionTime":"2026-01-21T11:11:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:25 crc kubenswrapper[4824]: I0121 11:11:25.996469 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-07 18:16:19.8902512 +0000 UTC Jan 21 11:11:26 crc kubenswrapper[4824]: I0121 11:11:26.048861 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 11:11:26 crc kubenswrapper[4824]: E0121 11:11:26.049006 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 21 11:11:26 crc kubenswrapper[4824]: I0121 11:11:26.059568 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:26 crc kubenswrapper[4824]: I0121 11:11:26.059598 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:26 crc kubenswrapper[4824]: I0121 11:11:26.059607 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:26 crc kubenswrapper[4824]: I0121 11:11:26.059620 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:26 crc kubenswrapper[4824]: I0121 11:11:26.059628 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:26Z","lastTransitionTime":"2026-01-21T11:11:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:26 crc kubenswrapper[4824]: I0121 11:11:26.161760 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:26 crc kubenswrapper[4824]: I0121 11:11:26.161791 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:26 crc kubenswrapper[4824]: I0121 11:11:26.161800 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:26 crc kubenswrapper[4824]: I0121 11:11:26.161815 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:26 crc kubenswrapper[4824]: I0121 11:11:26.161823 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:26Z","lastTransitionTime":"2026-01-21T11:11:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:26 crc kubenswrapper[4824]: I0121 11:11:26.263106 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:26 crc kubenswrapper[4824]: I0121 11:11:26.263304 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:26 crc kubenswrapper[4824]: I0121 11:11:26.263395 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:26 crc kubenswrapper[4824]: I0121 11:11:26.263454 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:26 crc kubenswrapper[4824]: I0121 11:11:26.263511 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:26Z","lastTransitionTime":"2026-01-21T11:11:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:26 crc kubenswrapper[4824]: I0121 11:11:26.365161 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:26 crc kubenswrapper[4824]: I0121 11:11:26.365187 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:26 crc kubenswrapper[4824]: I0121 11:11:26.365195 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:26 crc kubenswrapper[4824]: I0121 11:11:26.365209 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:26 crc kubenswrapper[4824]: I0121 11:11:26.365217 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:26Z","lastTransitionTime":"2026-01-21T11:11:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:26 crc kubenswrapper[4824]: I0121 11:11:26.467365 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:26 crc kubenswrapper[4824]: I0121 11:11:26.467387 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:26 crc kubenswrapper[4824]: I0121 11:11:26.467394 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:26 crc kubenswrapper[4824]: I0121 11:11:26.467405 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:26 crc kubenswrapper[4824]: I0121 11:11:26.467412 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:26Z","lastTransitionTime":"2026-01-21T11:11:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:26 crc kubenswrapper[4824]: I0121 11:11:26.569605 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:26 crc kubenswrapper[4824]: I0121 11:11:26.569641 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:26 crc kubenswrapper[4824]: I0121 11:11:26.569650 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:26 crc kubenswrapper[4824]: I0121 11:11:26.569664 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:26 crc kubenswrapper[4824]: I0121 11:11:26.569678 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:26Z","lastTransitionTime":"2026-01-21T11:11:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:26 crc kubenswrapper[4824]: I0121 11:11:26.671686 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:26 crc kubenswrapper[4824]: I0121 11:11:26.671720 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:26 crc kubenswrapper[4824]: I0121 11:11:26.671728 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:26 crc kubenswrapper[4824]: I0121 11:11:26.671743 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:26 crc kubenswrapper[4824]: I0121 11:11:26.671753 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:26Z","lastTransitionTime":"2026-01-21T11:11:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:26 crc kubenswrapper[4824]: I0121 11:11:26.773437 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:26 crc kubenswrapper[4824]: I0121 11:11:26.773465 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:26 crc kubenswrapper[4824]: I0121 11:11:26.773474 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:26 crc kubenswrapper[4824]: I0121 11:11:26.773484 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:26 crc kubenswrapper[4824]: I0121 11:11:26.773492 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:26Z","lastTransitionTime":"2026-01-21T11:11:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:26 crc kubenswrapper[4824]: I0121 11:11:26.875362 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:26 crc kubenswrapper[4824]: I0121 11:11:26.875391 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:26 crc kubenswrapper[4824]: I0121 11:11:26.875399 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:26 crc kubenswrapper[4824]: I0121 11:11:26.875410 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:26 crc kubenswrapper[4824]: I0121 11:11:26.875419 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:26Z","lastTransitionTime":"2026-01-21T11:11:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:26 crc kubenswrapper[4824]: I0121 11:11:26.977196 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:26 crc kubenswrapper[4824]: I0121 11:11:26.977357 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:26 crc kubenswrapper[4824]: I0121 11:11:26.977437 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:26 crc kubenswrapper[4824]: I0121 11:11:26.977497 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:26 crc kubenswrapper[4824]: I0121 11:11:26.977555 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:26Z","lastTransitionTime":"2026-01-21T11:11:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:26 crc kubenswrapper[4824]: I0121 11:11:26.996559 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-27 16:56:14.98029536 +0000 UTC Jan 21 11:11:27 crc kubenswrapper[4824]: I0121 11:11:27.049220 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 11:11:27 crc kubenswrapper[4824]: E0121 11:11:27.049379 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 21 11:11:27 crc kubenswrapper[4824]: I0121 11:11:27.049223 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 11:11:27 crc kubenswrapper[4824]: E0121 11:11:27.049548 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 21 11:11:27 crc kubenswrapper[4824]: I0121 11:11:27.078778 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:27 crc kubenswrapper[4824]: I0121 11:11:27.078826 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:27 crc kubenswrapper[4824]: I0121 11:11:27.078837 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:27 crc kubenswrapper[4824]: I0121 11:11:27.078849 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:27 crc kubenswrapper[4824]: I0121 11:11:27.078858 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:27Z","lastTransitionTime":"2026-01-21T11:11:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:27 crc kubenswrapper[4824]: I0121 11:11:27.180859 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:27 crc kubenswrapper[4824]: I0121 11:11:27.180895 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:27 crc kubenswrapper[4824]: I0121 11:11:27.180904 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:27 crc kubenswrapper[4824]: I0121 11:11:27.180917 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:27 crc kubenswrapper[4824]: I0121 11:11:27.180929 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:27Z","lastTransitionTime":"2026-01-21T11:11:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:27 crc kubenswrapper[4824]: I0121 11:11:27.197916 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-np2v8"] Jan 21 11:11:27 crc kubenswrapper[4824]: I0121 11:11:27.198259 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-np2v8" Jan 21 11:11:27 crc kubenswrapper[4824]: I0121 11:11:27.199481 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Jan 21 11:11:27 crc kubenswrapper[4824]: I0121 11:11:27.199678 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Jan 21 11:11:27 crc kubenswrapper[4824]: I0121 11:11:27.214914 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06f56d86-ec6f-4b57-b314-f508c0dd4d9f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d25c94f1ae0bbabb733f0e1ea1d1ba39562a92af73047a39aacc03013e239681\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca06234486b59cf990c69d9501e23f935328e2ba21013cd71ed64831ea749bcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b63e45d66bcd11ff3479cc3ec6b450017da8d21d36e967b8d496ae0f24b0ee9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52758bcd31f2ba64edf03adfd4d1be6428015abff5b64c7c029fa787085f9c29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2736e73342d2adb175b0721dd4f40cc42d26fca06d5eb3334927518531c3b107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b35ad7dd0570c191eb0080739304f3cd43f17b3d1dddf895549441ebe50deaa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b35ad7dd0570c191eb0080739304f3cd43f17b3d1dddf895549441ebe50deaa6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://80cd5ea5e6fc19a88c1aab5c61f827f103ab798a5c84da8dc4d45cea86d5aacb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80cd5ea5e6fc19a88c1aab5c61f827f103ab798a5c84da8dc4d45cea86d5aacb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://234d8b914e6a88480021647ae65d74e9b21c55b6788312a094d00c8a42677d48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://234d8b914e6a88480021647ae65d74e9b21c55b6788312a094d00c8a42677d48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:27Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:27 crc kubenswrapper[4824]: I0121 11:11:27.222770 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:27Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:27 crc kubenswrapper[4824]: I0121 11:11:27.230519 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86d3005cd61c0973659e1c1c0d3ef1ddd6fcf95bede5465f4f2099e64a74bb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:27Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:27 crc kubenswrapper[4824]: I0121 11:11:27.238155 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:27Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:27 crc kubenswrapper[4824]: I0121 11:11:27.245517 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x94nb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75d52224135b75133fc000611cd179240b9174f6492d3263341716c7e42ccb15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgr8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x94nb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:27Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:27 crc kubenswrapper[4824]: I0121 11:11:27.251801 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33f3d922-4ffe-409b-a49a-d88c85898260\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a9e57b39eed8300107e944b9fcf6a2ff1e75991f8f37387a3de615b14f9bc78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpwst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b1b37c6d2994021cfc30f85e92c40ae35677113b17aa066d52bc7a76ae5c153\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpwst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zqjg9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:27Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:27 crc kubenswrapper[4824]: I0121 11:11:27.260702 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gjtnh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a8d2cac-2d0f-4894-9bc0-2abd81699c47\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67c122a26811deaa9b70d19a1a65daee318ca3af2c626712482179d72fe67875\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2239186f1de46481cebfcafb1cab85cd320a205b5a0dfc6dd1b321d9baedeba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b2239186f1de46481cebfcafb1cab85cd320a205b5a0dfc6dd1b321d9baedeba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0db5ecd7759d974ae1b0d8065b3b8bd4d7957fae16eb5bab2d71c2fe69243816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0db5ecd7759d974ae1b0d8065b3b8bd4d7957fae16eb5bab2d71c2fe69243816\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba9b3278c208fc7aac630ca7a0daa3de0823f85d0799d6f1f278da80ae14c862\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba9b3278c208fc7aac630ca7a0daa3de0823f85d0799d6f1f278da80ae14c862\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79b8f6c016b43c3ac3ba7b22cca9a6b72c262cb17ddd6236c5d4970ecc09680c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79b8f6c016b43c3ac3ba7b22cca9a6b72c262cb17ddd6236c5d4970ecc09680c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38c4f218eb2ca7ea0c92bd0ae2ae9efeaa40746751608fbcf7f6c3f8d9e3c758\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38c4f218eb2ca7ea0c92bd0ae2ae9efeaa40746751608fbcf7f6c3f8d9e3c758\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d66ad9831d6029c8c350054525a00069e869d39c1b5638e06e9b772666bd8630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d66ad9831d6029c8c350054525a00069e869d39c1b5638e06e9b772666bd8630\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gjtnh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:27Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:27 crc kubenswrapper[4824]: I0121 11:11:27.267773 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d87422473838f5218c9fb646ab2515f2f42592754ffdf07b19f7a8e093fa48a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:27Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:27 crc kubenswrapper[4824]: I0121 11:11:27.278280 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://516c009d4720ab5b393c7482872f12fed22d4f4414bd1f3400411c0e8da88ce4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee18d293036971832c8d9c3f7b9eca84ab779ff6768a2db3c6f7bdcf3433cdef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:27Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:27 crc kubenswrapper[4824]: I0121 11:11:27.283057 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:27 crc kubenswrapper[4824]: I0121 11:11:27.283094 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:27 crc kubenswrapper[4824]: I0121 11:11:27.283104 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:27 crc kubenswrapper[4824]: I0121 11:11:27.283116 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:27 crc kubenswrapper[4824]: I0121 11:11:27.283124 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:27Z","lastTransitionTime":"2026-01-21T11:11:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:27 crc kubenswrapper[4824]: I0121 11:11:27.290385 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"26b1108a-4b98-4e39-a3cd-e0d055089fd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc319bbc06568cc811ded2b418b04b986a31f321fed492e8be50104af79bb30b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cee9cc9eec3fc815f1621a22a37805691288dbb60dd3ff64194a84be62abcddc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc4c96e33e069f7d08d33caf92cd0987668506f931fd134b246d5ce5e934f95c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ddbb0e5ae02e9d5d82a80ead7efc91f1b8e557a9fb6f39940e5530fae8de2d38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://64a9065254f889b1c43d63759283335bf78fc4fa1531a4baae472c06c075dae3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e94ef8a8f47dc75ecfbab2e3fb8e7097c1b73eed32d7ca947f11ea198cea880\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e688b12ddb605e11511f11a6b41f84cd31b6065797f6b7332b1255595bd27cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e688b12ddb605e11511f11a6b41f84cd31b6065797f6b7332b1255595bd27cb\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-21T11:11:23Z\\\",\\\"message\\\":\\\" crc\\\\nF0121 11:11:23.774847 6243 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:23Z is after 2025-08-24T17:21:41Z]\\\\nI0121 11:11:23.774842 6243 services_controller.go:451] Built service openshift-kube-storage-version-migrator-operator/metrics cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-kube-storage-version-migrator-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-storage-version-migrator-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeO\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:23Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-htkvk_openshift-ovn-kubernetes(26b1108a-4b98-4e39-a3cd-e0d055089fd8)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948e0497eaffa7b0d84b4085c9ed8d987eb2c6527c9a82c10e576ed9c9c1e21b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-htkvk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:27Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:27 crc kubenswrapper[4824]: I0121 11:11:27.297436 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-np2v8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4698abe4-0f33-4349-a2a4-614de80ae21b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n9qbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n9qbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-np2v8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:27Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:27 crc kubenswrapper[4824]: I0121 11:11:27.305836 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d31ab61-f94c-4f49-85d0-c9e91b83069b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a66b5c33884af00408dc35cf02f08ebd8570bd695f9f1fe558bf54fd06930431\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7514786602746dabfa690da48d48e233fe68d6dd6674aa5ecd489c0b3185f08\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61b14b6846f94cdeba8f1f61b3754e7663041f92c70356c4ee6cbb988a2e680f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a7023cfe4c098bcdd9c0477a8187f4ed7a0d2abdbe1e143436863948f441b4f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af11ad95bcf0c6bbd5fa4ab99bb592ce1c87f40d29c5ade7832414315820a569\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW0121 11:11:14.604290 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0121 11:11:14.604572 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0121 11:11:14.605875 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3284452394/tls.crt::/tmp/serving-cert-3284452394/tls.key\\\\\\\"\\\\nI0121 11:11:14.923427 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0121 11:11:14.928141 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0121 11:11:14.928162 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0121 11:11:14.928179 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0121 11:11:14.928184 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0121 11:11:14.933154 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0121 11:11:14.933230 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 11:11:14.933259 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 11:11:14.933282 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0121 11:11:14.933299 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0121 11:11:14.933315 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0121 11:11:14.933331 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0121 11:11:14.933177 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0121 11:11:14.934666 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48070c09fad511108ff22e44211cf26911509b8406d66948985cbfc3009d8233\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:27Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:27 crc kubenswrapper[4824]: I0121 11:11:27.307124 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n9qbj\" (UniqueName: \"kubernetes.io/projected/4698abe4-0f33-4349-a2a4-614de80ae21b-kube-api-access-n9qbj\") pod \"ovnkube-control-plane-749d76644c-np2v8\" (UID: \"4698abe4-0f33-4349-a2a4-614de80ae21b\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-np2v8" Jan 21 11:11:27 crc kubenswrapper[4824]: I0121 11:11:27.307241 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/4698abe4-0f33-4349-a2a4-614de80ae21b-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-np2v8\" (UID: \"4698abe4-0f33-4349-a2a4-614de80ae21b\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-np2v8" Jan 21 11:11:27 crc kubenswrapper[4824]: I0121 11:11:27.307335 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/4698abe4-0f33-4349-a2a4-614de80ae21b-env-overrides\") pod \"ovnkube-control-plane-749d76644c-np2v8\" (UID: \"4698abe4-0f33-4349-a2a4-614de80ae21b\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-np2v8" Jan 21 11:11:27 crc kubenswrapper[4824]: I0121 11:11:27.307425 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/4698abe4-0f33-4349-a2a4-614de80ae21b-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-np2v8\" (UID: \"4698abe4-0f33-4349-a2a4-614de80ae21b\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-np2v8" Jan 21 11:11:27 crc kubenswrapper[4824]: I0121 11:11:27.313468 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9bb99c0a-683b-44da-9c0e-e8d284e87d53\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05e076febd697ed802ce5c7ee92a8c2c636a1a2c6609ad7eb01f17dace247c4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43529d1b9e1e5e22b0d56b023946119da8aacbae54b68c26ff0a1a82dc890c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ea232ad87393685aa323ce6393ff3869f8f8d423c2406a501c69c355535adfa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc6c7eff9e17b8a8ecc88bbfe703b5a01dbf1e02d342cede4f6a263a6526af43\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:27Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:27 crc kubenswrapper[4824]: I0121 11:11:27.321075 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:27Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:27 crc kubenswrapper[4824]: I0121 11:11:27.327618 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lv4rn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1267539-9adf-466f-8027-8ede009d19a0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eed1b884aa53e1ee318169a0cc2c9ad10130aa363ae4ac55e4bf065237756f53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbmfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lv4rn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:27Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:27 crc kubenswrapper[4824]: I0121 11:11:27.334554 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jr64s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9768c60d-5bee-4286-bced-076e584b04e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91158f14d2e98ac9f875a5e7101df3107ef28aa5fc5d62218f460af98ff920fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flrnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:17Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jr64s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:27Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:27 crc kubenswrapper[4824]: I0121 11:11:27.384735 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:27 crc kubenswrapper[4824]: I0121 11:11:27.384762 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:27 crc kubenswrapper[4824]: I0121 11:11:27.384772 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:27 crc kubenswrapper[4824]: I0121 11:11:27.384785 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:27 crc kubenswrapper[4824]: I0121 11:11:27.384812 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:27Z","lastTransitionTime":"2026-01-21T11:11:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:27 crc kubenswrapper[4824]: I0121 11:11:27.408386 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n9qbj\" (UniqueName: \"kubernetes.io/projected/4698abe4-0f33-4349-a2a4-614de80ae21b-kube-api-access-n9qbj\") pod \"ovnkube-control-plane-749d76644c-np2v8\" (UID: \"4698abe4-0f33-4349-a2a4-614de80ae21b\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-np2v8" Jan 21 11:11:27 crc kubenswrapper[4824]: I0121 11:11:27.408434 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/4698abe4-0f33-4349-a2a4-614de80ae21b-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-np2v8\" (UID: \"4698abe4-0f33-4349-a2a4-614de80ae21b\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-np2v8" Jan 21 11:11:27 crc kubenswrapper[4824]: I0121 11:11:27.408468 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/4698abe4-0f33-4349-a2a4-614de80ae21b-env-overrides\") pod \"ovnkube-control-plane-749d76644c-np2v8\" (UID: \"4698abe4-0f33-4349-a2a4-614de80ae21b\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-np2v8" Jan 21 11:11:27 crc kubenswrapper[4824]: I0121 11:11:27.408482 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/4698abe4-0f33-4349-a2a4-614de80ae21b-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-np2v8\" (UID: \"4698abe4-0f33-4349-a2a4-614de80ae21b\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-np2v8" Jan 21 11:11:27 crc kubenswrapper[4824]: I0121 11:11:27.409037 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/4698abe4-0f33-4349-a2a4-614de80ae21b-env-overrides\") pod \"ovnkube-control-plane-749d76644c-np2v8\" (UID: \"4698abe4-0f33-4349-a2a4-614de80ae21b\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-np2v8" Jan 21 11:11:27 crc kubenswrapper[4824]: I0121 11:11:27.409107 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/4698abe4-0f33-4349-a2a4-614de80ae21b-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-np2v8\" (UID: \"4698abe4-0f33-4349-a2a4-614de80ae21b\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-np2v8" Jan 21 11:11:27 crc kubenswrapper[4824]: I0121 11:11:27.412470 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/4698abe4-0f33-4349-a2a4-614de80ae21b-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-np2v8\" (UID: \"4698abe4-0f33-4349-a2a4-614de80ae21b\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-np2v8" Jan 21 11:11:27 crc kubenswrapper[4824]: I0121 11:11:27.419924 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n9qbj\" (UniqueName: \"kubernetes.io/projected/4698abe4-0f33-4349-a2a4-614de80ae21b-kube-api-access-n9qbj\") pod \"ovnkube-control-plane-749d76644c-np2v8\" (UID: \"4698abe4-0f33-4349-a2a4-614de80ae21b\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-np2v8" Jan 21 11:11:27 crc kubenswrapper[4824]: I0121 11:11:27.486569 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:27 crc kubenswrapper[4824]: I0121 11:11:27.486607 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:27 crc kubenswrapper[4824]: I0121 11:11:27.486616 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:27 crc kubenswrapper[4824]: I0121 11:11:27.486631 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:27 crc kubenswrapper[4824]: I0121 11:11:27.486642 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:27Z","lastTransitionTime":"2026-01-21T11:11:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:27 crc kubenswrapper[4824]: I0121 11:11:27.506947 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-np2v8" Jan 21 11:11:27 crc kubenswrapper[4824]: W0121 11:11:27.515491 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4698abe4_0f33_4349_a2a4_614de80ae21b.slice/crio-83061edbe8885031d9d7ad1d9dc528be6ab148f6f36122e2040a9e71cd925887 WatchSource:0}: Error finding container 83061edbe8885031d9d7ad1d9dc528be6ab148f6f36122e2040a9e71cd925887: Status 404 returned error can't find the container with id 83061edbe8885031d9d7ad1d9dc528be6ab148f6f36122e2040a9e71cd925887 Jan 21 11:11:27 crc kubenswrapper[4824]: I0121 11:11:27.588561 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:27 crc kubenswrapper[4824]: I0121 11:11:27.588587 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:27 crc kubenswrapper[4824]: I0121 11:11:27.588595 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:27 crc kubenswrapper[4824]: I0121 11:11:27.588608 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:27 crc kubenswrapper[4824]: I0121 11:11:27.588616 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:27Z","lastTransitionTime":"2026-01-21T11:11:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:27 crc kubenswrapper[4824]: I0121 11:11:27.690901 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:27 crc kubenswrapper[4824]: I0121 11:11:27.690932 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:27 crc kubenswrapper[4824]: I0121 11:11:27.690941 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:27 crc kubenswrapper[4824]: I0121 11:11:27.690966 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:27 crc kubenswrapper[4824]: I0121 11:11:27.690975 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:27Z","lastTransitionTime":"2026-01-21T11:11:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:27 crc kubenswrapper[4824]: I0121 11:11:27.793074 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:27 crc kubenswrapper[4824]: I0121 11:11:27.793119 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:27 crc kubenswrapper[4824]: I0121 11:11:27.793130 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:27 crc kubenswrapper[4824]: I0121 11:11:27.793144 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:27 crc kubenswrapper[4824]: I0121 11:11:27.793153 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:27Z","lastTransitionTime":"2026-01-21T11:11:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:27 crc kubenswrapper[4824]: I0121 11:11:27.895294 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:27 crc kubenswrapper[4824]: I0121 11:11:27.895325 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:27 crc kubenswrapper[4824]: I0121 11:11:27.895334 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:27 crc kubenswrapper[4824]: I0121 11:11:27.895346 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:27 crc kubenswrapper[4824]: I0121 11:11:27.895355 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:27Z","lastTransitionTime":"2026-01-21T11:11:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:27 crc kubenswrapper[4824]: I0121 11:11:27.996906 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-20 14:43:16.754203739 +0000 UTC Jan 21 11:11:27 crc kubenswrapper[4824]: I0121 11:11:27.997056 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:27 crc kubenswrapper[4824]: I0121 11:11:27.997078 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:27 crc kubenswrapper[4824]: I0121 11:11:27.997086 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:27 crc kubenswrapper[4824]: I0121 11:11:27.997106 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:27 crc kubenswrapper[4824]: I0121 11:11:27.997116 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:27Z","lastTransitionTime":"2026-01-21T11:11:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.048897 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 11:11:28 crc kubenswrapper[4824]: E0121 11:11:28.049019 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.062302 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06f56d86-ec6f-4b57-b314-f508c0dd4d9f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d25c94f1ae0bbabb733f0e1ea1d1ba39562a92af73047a39aacc03013e239681\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca06234486b59cf990c69d9501e23f935328e2ba21013cd71ed64831ea749bcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b63e45d66bcd11ff3479cc3ec6b450017da8d21d36e967b8d496ae0f24b0ee9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52758bcd31f2ba64edf03adfd4d1be6428015abff5b64c7c029fa787085f9c29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2736e73342d2adb175b0721dd4f40cc42d26fca06d5eb3334927518531c3b107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b35ad7dd0570c191eb0080739304f3cd43f17b3d1dddf895549441ebe50deaa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b35ad7dd0570c191eb0080739304f3cd43f17b3d1dddf895549441ebe50deaa6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://80cd5ea5e6fc19a88c1aab5c61f827f103ab798a5c84da8dc4d45cea86d5aacb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80cd5ea5e6fc19a88c1aab5c61f827f103ab798a5c84da8dc4d45cea86d5aacb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://234d8b914e6a88480021647ae65d74e9b21c55b6788312a094d00c8a42677d48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://234d8b914e6a88480021647ae65d74e9b21c55b6788312a094d00c8a42677d48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:28Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.071311 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:28Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.076783 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.076810 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.076823 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.076838 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.076848 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:28Z","lastTransitionTime":"2026-01-21T11:11:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.081774 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gjtnh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a8d2cac-2d0f-4894-9bc0-2abd81699c47\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67c122a26811deaa9b70d19a1a65daee318ca3af2c626712482179d72fe67875\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2239186f1de46481cebfcafb1cab85cd320a205b5a0dfc6dd1b321d9baedeba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b2239186f1de46481cebfcafb1cab85cd320a205b5a0dfc6dd1b321d9baedeba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0db5ecd7759d974ae1b0d8065b3b8bd4d7957fae16eb5bab2d71c2fe69243816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0db5ecd7759d974ae1b0d8065b3b8bd4d7957fae16eb5bab2d71c2fe69243816\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba9b3278c208fc7aac630ca7a0daa3de0823f85d0799d6f1f278da80ae14c862\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba9b3278c208fc7aac630ca7a0daa3de0823f85d0799d6f1f278da80ae14c862\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79b8f6c016b43c3ac3ba7b22cca9a6b72c262cb17ddd6236c5d4970ecc09680c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79b8f6c016b43c3ac3ba7b22cca9a6b72c262cb17ddd6236c5d4970ecc09680c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38c4f218eb2ca7ea0c92bd0ae2ae9efeaa40746751608fbcf7f6c3f8d9e3c758\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38c4f218eb2ca7ea0c92bd0ae2ae9efeaa40746751608fbcf7f6c3f8d9e3c758\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d66ad9831d6029c8c350054525a00069e869d39c1b5638e06e9b772666bd8630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d66ad9831d6029c8c350054525a00069e869d39c1b5638e06e9b772666bd8630\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gjtnh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:28Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:28 crc kubenswrapper[4824]: E0121 11:11:28.085032 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"695c676c-720c-4921-9e71-56b96f2fb46b\\\",\\\"systemUUID\\\":\\\"d00624f5-786f-41dc-bdc1-a1875f167d65\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:28Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.087579 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.087607 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.087616 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.087627 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.087634 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:28Z","lastTransitionTime":"2026-01-21T11:11:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.096275 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86d3005cd61c0973659e1c1c0d3ef1ddd6fcf95bede5465f4f2099e64a74bb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:28Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:28 crc kubenswrapper[4824]: E0121 11:11:28.097270 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"695c676c-720c-4921-9e71-56b96f2fb46b\\\",\\\"systemUUID\\\":\\\"d00624f5-786f-41dc-bdc1-a1875f167d65\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:28Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.099822 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.099854 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.099863 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.099875 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.099883 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:28Z","lastTransitionTime":"2026-01-21T11:11:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.104952 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:28Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:28 crc kubenswrapper[4824]: E0121 11:11:28.107667 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"695c676c-720c-4921-9e71-56b96f2fb46b\\\",\\\"systemUUID\\\":\\\"d00624f5-786f-41dc-bdc1-a1875f167d65\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:28Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.109937 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.109990 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.110002 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.110014 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.110024 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:28Z","lastTransitionTime":"2026-01-21T11:11:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.115575 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x94nb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75d52224135b75133fc000611cd179240b9174f6492d3263341716c7e42ccb15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgr8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x94nb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:28Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:28 crc kubenswrapper[4824]: E0121 11:11:28.118327 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"695c676c-720c-4921-9e71-56b96f2fb46b\\\",\\\"systemUUID\\\":\\\"d00624f5-786f-41dc-bdc1-a1875f167d65\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:28Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.120607 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.120630 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.120639 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.120651 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.120660 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:28Z","lastTransitionTime":"2026-01-21T11:11:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.122517 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33f3d922-4ffe-409b-a49a-d88c85898260\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a9e57b39eed8300107e944b9fcf6a2ff1e75991f8f37387a3de615b14f9bc78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpwst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b1b37c6d2994021cfc30f85e92c40ae35677113b17aa066d52bc7a76ae5c153\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpwst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zqjg9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:28Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:28 crc kubenswrapper[4824]: E0121 11:11:28.129647 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"695c676c-720c-4921-9e71-56b96f2fb46b\\\",\\\"systemUUID\\\":\\\"d00624f5-786f-41dc-bdc1-a1875f167d65\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:28Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:28 crc kubenswrapper[4824]: E0121 11:11:28.129884 4824 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.130845 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.130869 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.130878 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.130888 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.130895 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:28Z","lastTransitionTime":"2026-01-21T11:11:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.131278 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d87422473838f5218c9fb646ab2515f2f42592754ffdf07b19f7a8e093fa48a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:28Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.139470 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://516c009d4720ab5b393c7482872f12fed22d4f4414bd1f3400411c0e8da88ce4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee18d293036971832c8d9c3f7b9eca84ab779ff6768a2db3c6f7bdcf3433cdef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:28Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.151343 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"26b1108a-4b98-4e39-a3cd-e0d055089fd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc319bbc06568cc811ded2b418b04b986a31f321fed492e8be50104af79bb30b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cee9cc9eec3fc815f1621a22a37805691288dbb60dd3ff64194a84be62abcddc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc4c96e33e069f7d08d33caf92cd0987668506f931fd134b246d5ce5e934f95c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ddbb0e5ae02e9d5d82a80ead7efc91f1b8e557a9fb6f39940e5530fae8de2d38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://64a9065254f889b1c43d63759283335bf78fc4fa1531a4baae472c06c075dae3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e94ef8a8f47dc75ecfbab2e3fb8e7097c1b73eed32d7ca947f11ea198cea880\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e688b12ddb605e11511f11a6b41f84cd31b6065797f6b7332b1255595bd27cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e688b12ddb605e11511f11a6b41f84cd31b6065797f6b7332b1255595bd27cb\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-21T11:11:23Z\\\",\\\"message\\\":\\\" crc\\\\nF0121 11:11:23.774847 6243 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:23Z is after 2025-08-24T17:21:41Z]\\\\nI0121 11:11:23.774842 6243 services_controller.go:451] Built service openshift-kube-storage-version-migrator-operator/metrics cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-kube-storage-version-migrator-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-storage-version-migrator-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeO\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:23Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-htkvk_openshift-ovn-kubernetes(26b1108a-4b98-4e39-a3cd-e0d055089fd8)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948e0497eaffa7b0d84b4085c9ed8d987eb2c6527c9a82c10e576ed9c9c1e21b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-htkvk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:28Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.158317 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-np2v8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4698abe4-0f33-4349-a2a4-614de80ae21b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n9qbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n9qbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-np2v8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:28Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.165924 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:28Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.172034 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lv4rn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1267539-9adf-466f-8027-8ede009d19a0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eed1b884aa53e1ee318169a0cc2c9ad10130aa363ae4ac55e4bf065237756f53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbmfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lv4rn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:28Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.179237 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jr64s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9768c60d-5bee-4286-bced-076e584b04e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91158f14d2e98ac9f875a5e7101df3107ef28aa5fc5d62218f460af98ff920fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flrnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:17Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jr64s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:28Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.187442 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d31ab61-f94c-4f49-85d0-c9e91b83069b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a66b5c33884af00408dc35cf02f08ebd8570bd695f9f1fe558bf54fd06930431\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7514786602746dabfa690da48d48e233fe68d6dd6674aa5ecd489c0b3185f08\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61b14b6846f94cdeba8f1f61b3754e7663041f92c70356c4ee6cbb988a2e680f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a7023cfe4c098bcdd9c0477a8187f4ed7a0d2abdbe1e143436863948f441b4f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af11ad95bcf0c6bbd5fa4ab99bb592ce1c87f40d29c5ade7832414315820a569\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW0121 11:11:14.604290 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0121 11:11:14.604572 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0121 11:11:14.605875 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3284452394/tls.crt::/tmp/serving-cert-3284452394/tls.key\\\\\\\"\\\\nI0121 11:11:14.923427 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0121 11:11:14.928141 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0121 11:11:14.928162 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0121 11:11:14.928179 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0121 11:11:14.928184 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0121 11:11:14.933154 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0121 11:11:14.933230 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 11:11:14.933259 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 11:11:14.933282 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0121 11:11:14.933299 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0121 11:11:14.933315 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0121 11:11:14.933331 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0121 11:11:14.933177 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0121 11:11:14.934666 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48070c09fad511108ff22e44211cf26911509b8406d66948985cbfc3009d8233\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:28Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.195637 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9bb99c0a-683b-44da-9c0e-e8d284e87d53\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05e076febd697ed802ce5c7ee92a8c2c636a1a2c6609ad7eb01f17dace247c4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43529d1b9e1e5e22b0d56b023946119da8aacbae54b68c26ff0a1a82dc890c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ea232ad87393685aa323ce6393ff3869f8f8d423c2406a501c69c355535adfa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc6c7eff9e17b8a8ecc88bbfe703b5a01dbf1e02d342cede4f6a263a6526af43\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:28Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.206709 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-np2v8" event={"ID":"4698abe4-0f33-4349-a2a4-614de80ae21b","Type":"ContainerStarted","Data":"050bd5bdded2162d5e9dac81713f4ed63cb575d1776e828b2f98e50d89f11b04"} Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.206746 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-np2v8" event={"ID":"4698abe4-0f33-4349-a2a4-614de80ae21b","Type":"ContainerStarted","Data":"191328471bb6f2b34cde2eb927da697619826bf0b6174ecb583c8c25d27a6654"} Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.206756 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-np2v8" event={"ID":"4698abe4-0f33-4349-a2a4-614de80ae21b","Type":"ContainerStarted","Data":"83061edbe8885031d9d7ad1d9dc528be6ab148f6f36122e2040a9e71cd925887"} Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.215755 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86d3005cd61c0973659e1c1c0d3ef1ddd6fcf95bede5465f4f2099e64a74bb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:28Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.223401 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:28Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.231617 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x94nb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75d52224135b75133fc000611cd179240b9174f6492d3263341716c7e42ccb15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgr8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x94nb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:28Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.232636 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.232667 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.232678 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.232692 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.232700 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:28Z","lastTransitionTime":"2026-01-21T11:11:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.238543 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33f3d922-4ffe-409b-a49a-d88c85898260\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a9e57b39eed8300107e944b9fcf6a2ff1e75991f8f37387a3de615b14f9bc78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpwst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b1b37c6d2994021cfc30f85e92c40ae35677113b17aa066d52bc7a76ae5c153\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpwst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zqjg9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:28Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.251256 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gjtnh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a8d2cac-2d0f-4894-9bc0-2abd81699c47\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67c122a26811deaa9b70d19a1a65daee318ca3af2c626712482179d72fe67875\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2239186f1de46481cebfcafb1cab85cd320a205b5a0dfc6dd1b321d9baedeba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b2239186f1de46481cebfcafb1cab85cd320a205b5a0dfc6dd1b321d9baedeba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0db5ecd7759d974ae1b0d8065b3b8bd4d7957fae16eb5bab2d71c2fe69243816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0db5ecd7759d974ae1b0d8065b3b8bd4d7957fae16eb5bab2d71c2fe69243816\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba9b3278c208fc7aac630ca7a0daa3de0823f85d0799d6f1f278da80ae14c862\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba9b3278c208fc7aac630ca7a0daa3de0823f85d0799d6f1f278da80ae14c862\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79b8f6c016b43c3ac3ba7b22cca9a6b72c262cb17ddd6236c5d4970ecc09680c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79b8f6c016b43c3ac3ba7b22cca9a6b72c262cb17ddd6236c5d4970ecc09680c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38c4f218eb2ca7ea0c92bd0ae2ae9efeaa40746751608fbcf7f6c3f8d9e3c758\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38c4f218eb2ca7ea0c92bd0ae2ae9efeaa40746751608fbcf7f6c3f8d9e3c758\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d66ad9831d6029c8c350054525a00069e869d39c1b5638e06e9b772666bd8630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d66ad9831d6029c8c350054525a00069e869d39c1b5638e06e9b772666bd8630\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gjtnh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:28Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.255492 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-9rmnw"] Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.255892 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9rmnw" Jan 21 11:11:28 crc kubenswrapper[4824]: E0121 11:11:28.255948 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9rmnw" podUID="8f9bd12b-22ee-44ed-a8b9-c100d043f691" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.258931 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d87422473838f5218c9fb646ab2515f2f42592754ffdf07b19f7a8e093fa48a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:28Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.266990 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://516c009d4720ab5b393c7482872f12fed22d4f4414bd1f3400411c0e8da88ce4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee18d293036971832c8d9c3f7b9eca84ab779ff6768a2db3c6f7bdcf3433cdef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:28Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.278620 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"26b1108a-4b98-4e39-a3cd-e0d055089fd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc319bbc06568cc811ded2b418b04b986a31f321fed492e8be50104af79bb30b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cee9cc9eec3fc815f1621a22a37805691288dbb60dd3ff64194a84be62abcddc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc4c96e33e069f7d08d33caf92cd0987668506f931fd134b246d5ce5e934f95c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ddbb0e5ae02e9d5d82a80ead7efc91f1b8e557a9fb6f39940e5530fae8de2d38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://64a9065254f889b1c43d63759283335bf78fc4fa1531a4baae472c06c075dae3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e94ef8a8f47dc75ecfbab2e3fb8e7097c1b73eed32d7ca947f11ea198cea880\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e688b12ddb605e11511f11a6b41f84cd31b6065797f6b7332b1255595bd27cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e688b12ddb605e11511f11a6b41f84cd31b6065797f6b7332b1255595bd27cb\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-21T11:11:23Z\\\",\\\"message\\\":\\\" crc\\\\nF0121 11:11:23.774847 6243 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:23Z is after 2025-08-24T17:21:41Z]\\\\nI0121 11:11:23.774842 6243 services_controller.go:451] Built service openshift-kube-storage-version-migrator-operator/metrics cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-kube-storage-version-migrator-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-storage-version-migrator-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeO\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:23Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-htkvk_openshift-ovn-kubernetes(26b1108a-4b98-4e39-a3cd-e0d055089fd8)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948e0497eaffa7b0d84b4085c9ed8d987eb2c6527c9a82c10e576ed9c9c1e21b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-htkvk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:28Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.285570 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-np2v8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4698abe4-0f33-4349-a2a4-614de80ae21b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://191328471bb6f2b34cde2eb927da697619826bf0b6174ecb583c8c25d27a6654\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n9qbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://050bd5bdded2162d5e9dac81713f4ed63cb575d1776e828b2f98e50d89f11b04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n9qbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-np2v8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:28Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.292088 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lv4rn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1267539-9adf-466f-8027-8ede009d19a0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eed1b884aa53e1ee318169a0cc2c9ad10130aa363ae4ac55e4bf065237756f53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbmfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lv4rn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:28Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.298249 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jr64s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9768c60d-5bee-4286-bced-076e584b04e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91158f14d2e98ac9f875a5e7101df3107ef28aa5fc5d62218f460af98ff920fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flrnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:17Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jr64s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:28Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.306736 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d31ab61-f94c-4f49-85d0-c9e91b83069b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a66b5c33884af00408dc35cf02f08ebd8570bd695f9f1fe558bf54fd06930431\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7514786602746dabfa690da48d48e233fe68d6dd6674aa5ecd489c0b3185f08\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61b14b6846f94cdeba8f1f61b3754e7663041f92c70356c4ee6cbb988a2e680f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a7023cfe4c098bcdd9c0477a8187f4ed7a0d2abdbe1e143436863948f441b4f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af11ad95bcf0c6bbd5fa4ab99bb592ce1c87f40d29c5ade7832414315820a569\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW0121 11:11:14.604290 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0121 11:11:14.604572 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0121 11:11:14.605875 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3284452394/tls.crt::/tmp/serving-cert-3284452394/tls.key\\\\\\\"\\\\nI0121 11:11:14.923427 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0121 11:11:14.928141 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0121 11:11:14.928162 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0121 11:11:14.928179 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0121 11:11:14.928184 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0121 11:11:14.933154 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0121 11:11:14.933230 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 11:11:14.933259 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 11:11:14.933282 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0121 11:11:14.933299 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0121 11:11:14.933315 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0121 11:11:14.933331 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0121 11:11:14.933177 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0121 11:11:14.934666 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48070c09fad511108ff22e44211cf26911509b8406d66948985cbfc3009d8233\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:28Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.314411 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9bb99c0a-683b-44da-9c0e-e8d284e87d53\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05e076febd697ed802ce5c7ee92a8c2c636a1a2c6609ad7eb01f17dace247c4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43529d1b9e1e5e22b0d56b023946119da8aacbae54b68c26ff0a1a82dc890c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ea232ad87393685aa323ce6393ff3869f8f8d423c2406a501c69c355535adfa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc6c7eff9e17b8a8ecc88bbfe703b5a01dbf1e02d342cede4f6a263a6526af43\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:28Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.316190 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8f9bd12b-22ee-44ed-a8b9-c100d043f691-metrics-certs\") pod \"network-metrics-daemon-9rmnw\" (UID: \"8f9bd12b-22ee-44ed-a8b9-c100d043f691\") " pod="openshift-multus/network-metrics-daemon-9rmnw" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.316226 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q2fgn\" (UniqueName: \"kubernetes.io/projected/8f9bd12b-22ee-44ed-a8b9-c100d043f691-kube-api-access-q2fgn\") pod \"network-metrics-daemon-9rmnw\" (UID: \"8f9bd12b-22ee-44ed-a8b9-c100d043f691\") " pod="openshift-multus/network-metrics-daemon-9rmnw" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.321742 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:28Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.334806 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.334830 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.334838 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.334851 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.334860 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:28Z","lastTransitionTime":"2026-01-21T11:11:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.335859 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06f56d86-ec6f-4b57-b314-f508c0dd4d9f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d25c94f1ae0bbabb733f0e1ea1d1ba39562a92af73047a39aacc03013e239681\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca06234486b59cf990c69d9501e23f935328e2ba21013cd71ed64831ea749bcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b63e45d66bcd11ff3479cc3ec6b450017da8d21d36e967b8d496ae0f24b0ee9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52758bcd31f2ba64edf03adfd4d1be6428015abff5b64c7c029fa787085f9c29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2736e73342d2adb175b0721dd4f40cc42d26fca06d5eb3334927518531c3b107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b35ad7dd0570c191eb0080739304f3cd43f17b3d1dddf895549441ebe50deaa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b35ad7dd0570c191eb0080739304f3cd43f17b3d1dddf895549441ebe50deaa6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://80cd5ea5e6fc19a88c1aab5c61f827f103ab798a5c84da8dc4d45cea86d5aacb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80cd5ea5e6fc19a88c1aab5c61f827f103ab798a5c84da8dc4d45cea86d5aacb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://234d8b914e6a88480021647ae65d74e9b21c55b6788312a094d00c8a42677d48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://234d8b914e6a88480021647ae65d74e9b21c55b6788312a094d00c8a42677d48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:28Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.344054 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:28Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.356885 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-np2v8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4698abe4-0f33-4349-a2a4-614de80ae21b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://191328471bb6f2b34cde2eb927da697619826bf0b6174ecb583c8c25d27a6654\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n9qbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://050bd5bdded2162d5e9dac81713f4ed63cb575d1776e828b2f98e50d89f11b04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n9qbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-np2v8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:28Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.371998 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d87422473838f5218c9fb646ab2515f2f42592754ffdf07b19f7a8e093fa48a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:28Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.384572 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://516c009d4720ab5b393c7482872f12fed22d4f4414bd1f3400411c0e8da88ce4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee18d293036971832c8d9c3f7b9eca84ab779ff6768a2db3c6f7bdcf3433cdef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:28Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.397910 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"26b1108a-4b98-4e39-a3cd-e0d055089fd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc319bbc06568cc811ded2b418b04b986a31f321fed492e8be50104af79bb30b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cee9cc9eec3fc815f1621a22a37805691288dbb60dd3ff64194a84be62abcddc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc4c96e33e069f7d08d33caf92cd0987668506f931fd134b246d5ce5e934f95c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ddbb0e5ae02e9d5d82a80ead7efc91f1b8e557a9fb6f39940e5530fae8de2d38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://64a9065254f889b1c43d63759283335bf78fc4fa1531a4baae472c06c075dae3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e94ef8a8f47dc75ecfbab2e3fb8e7097c1b73eed32d7ca947f11ea198cea880\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e688b12ddb605e11511f11a6b41f84cd31b6065797f6b7332b1255595bd27cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e688b12ddb605e11511f11a6b41f84cd31b6065797f6b7332b1255595bd27cb\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-21T11:11:23Z\\\",\\\"message\\\":\\\" crc\\\\nF0121 11:11:23.774847 6243 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:23Z is after 2025-08-24T17:21:41Z]\\\\nI0121 11:11:23.774842 6243 services_controller.go:451] Built service openshift-kube-storage-version-migrator-operator/metrics cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-kube-storage-version-migrator-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-storage-version-migrator-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeO\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:23Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-htkvk_openshift-ovn-kubernetes(26b1108a-4b98-4e39-a3cd-e0d055089fd8)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948e0497eaffa7b0d84b4085c9ed8d987eb2c6527c9a82c10e576ed9c9c1e21b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-htkvk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:28Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.406616 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:28Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.414251 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lv4rn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1267539-9adf-466f-8027-8ede009d19a0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eed1b884aa53e1ee318169a0cc2c9ad10130aa363ae4ac55e4bf065237756f53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbmfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lv4rn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:28Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.417550 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8f9bd12b-22ee-44ed-a8b9-c100d043f691-metrics-certs\") pod \"network-metrics-daemon-9rmnw\" (UID: \"8f9bd12b-22ee-44ed-a8b9-c100d043f691\") " pod="openshift-multus/network-metrics-daemon-9rmnw" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.417588 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q2fgn\" (UniqueName: \"kubernetes.io/projected/8f9bd12b-22ee-44ed-a8b9-c100d043f691-kube-api-access-q2fgn\") pod \"network-metrics-daemon-9rmnw\" (UID: \"8f9bd12b-22ee-44ed-a8b9-c100d043f691\") " pod="openshift-multus/network-metrics-daemon-9rmnw" Jan 21 11:11:28 crc kubenswrapper[4824]: E0121 11:11:28.417733 4824 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Jan 21 11:11:28 crc kubenswrapper[4824]: E0121 11:11:28.417812 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8f9bd12b-22ee-44ed-a8b9-c100d043f691-metrics-certs podName:8f9bd12b-22ee-44ed-a8b9-c100d043f691 nodeName:}" failed. No retries permitted until 2026-01-21 11:11:28.917794601 +0000 UTC m=+31.210823893 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/8f9bd12b-22ee-44ed-a8b9-c100d043f691-metrics-certs") pod "network-metrics-daemon-9rmnw" (UID: "8f9bd12b-22ee-44ed-a8b9-c100d043f691") : object "openshift-multus"/"metrics-daemon-secret" not registered Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.421430 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jr64s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9768c60d-5bee-4286-bced-076e584b04e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91158f14d2e98ac9f875a5e7101df3107ef28aa5fc5d62218f460af98ff920fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flrnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:17Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jr64s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:28Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.429706 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q2fgn\" (UniqueName: \"kubernetes.io/projected/8f9bd12b-22ee-44ed-a8b9-c100d043f691-kube-api-access-q2fgn\") pod \"network-metrics-daemon-9rmnw\" (UID: \"8f9bd12b-22ee-44ed-a8b9-c100d043f691\") " pod="openshift-multus/network-metrics-daemon-9rmnw" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.431463 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d31ab61-f94c-4f49-85d0-c9e91b83069b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a66b5c33884af00408dc35cf02f08ebd8570bd695f9f1fe558bf54fd06930431\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7514786602746dabfa690da48d48e233fe68d6dd6674aa5ecd489c0b3185f08\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61b14b6846f94cdeba8f1f61b3754e7663041f92c70356c4ee6cbb988a2e680f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a7023cfe4c098bcdd9c0477a8187f4ed7a0d2abdbe1e143436863948f441b4f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af11ad95bcf0c6bbd5fa4ab99bb592ce1c87f40d29c5ade7832414315820a569\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW0121 11:11:14.604290 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0121 11:11:14.604572 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0121 11:11:14.605875 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3284452394/tls.crt::/tmp/serving-cert-3284452394/tls.key\\\\\\\"\\\\nI0121 11:11:14.923427 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0121 11:11:14.928141 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0121 11:11:14.928162 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0121 11:11:14.928179 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0121 11:11:14.928184 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0121 11:11:14.933154 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0121 11:11:14.933230 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 11:11:14.933259 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 11:11:14.933282 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0121 11:11:14.933299 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0121 11:11:14.933315 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0121 11:11:14.933331 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0121 11:11:14.933177 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0121 11:11:14.934666 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48070c09fad511108ff22e44211cf26911509b8406d66948985cbfc3009d8233\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:28Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.436533 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.436676 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.436749 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.436815 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.436871 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:28Z","lastTransitionTime":"2026-01-21T11:11:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.441596 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9bb99c0a-683b-44da-9c0e-e8d284e87d53\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05e076febd697ed802ce5c7ee92a8c2c636a1a2c6609ad7eb01f17dace247c4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43529d1b9e1e5e22b0d56b023946119da8aacbae54b68c26ff0a1a82dc890c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ea232ad87393685aa323ce6393ff3869f8f8d423c2406a501c69c355535adfa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc6c7eff9e17b8a8ecc88bbfe703b5a01dbf1e02d342cede4f6a263a6526af43\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:28Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.454863 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06f56d86-ec6f-4b57-b314-f508c0dd4d9f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d25c94f1ae0bbabb733f0e1ea1d1ba39562a92af73047a39aacc03013e239681\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca06234486b59cf990c69d9501e23f935328e2ba21013cd71ed64831ea749bcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b63e45d66bcd11ff3479cc3ec6b450017da8d21d36e967b8d496ae0f24b0ee9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52758bcd31f2ba64edf03adfd4d1be6428015abff5b64c7c029fa787085f9c29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2736e73342d2adb175b0721dd4f40cc42d26fca06d5eb3334927518531c3b107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b35ad7dd0570c191eb0080739304f3cd43f17b3d1dddf895549441ebe50deaa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b35ad7dd0570c191eb0080739304f3cd43f17b3d1dddf895549441ebe50deaa6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://80cd5ea5e6fc19a88c1aab5c61f827f103ab798a5c84da8dc4d45cea86d5aacb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80cd5ea5e6fc19a88c1aab5c61f827f103ab798a5c84da8dc4d45cea86d5aacb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://234d8b914e6a88480021647ae65d74e9b21c55b6788312a094d00c8a42677d48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://234d8b914e6a88480021647ae65d74e9b21c55b6788312a094d00c8a42677d48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:28Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.464088 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:28Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.471693 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33f3d922-4ffe-409b-a49a-d88c85898260\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a9e57b39eed8300107e944b9fcf6a2ff1e75991f8f37387a3de615b14f9bc78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpwst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b1b37c6d2994021cfc30f85e92c40ae35677113b17aa066d52bc7a76ae5c153\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpwst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zqjg9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:28Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.482196 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gjtnh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a8d2cac-2d0f-4894-9bc0-2abd81699c47\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67c122a26811deaa9b70d19a1a65daee318ca3af2c626712482179d72fe67875\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2239186f1de46481cebfcafb1cab85cd320a205b5a0dfc6dd1b321d9baedeba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b2239186f1de46481cebfcafb1cab85cd320a205b5a0dfc6dd1b321d9baedeba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0db5ecd7759d974ae1b0d8065b3b8bd4d7957fae16eb5bab2d71c2fe69243816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0db5ecd7759d974ae1b0d8065b3b8bd4d7957fae16eb5bab2d71c2fe69243816\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba9b3278c208fc7aac630ca7a0daa3de0823f85d0799d6f1f278da80ae14c862\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba9b3278c208fc7aac630ca7a0daa3de0823f85d0799d6f1f278da80ae14c862\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79b8f6c016b43c3ac3ba7b22cca9a6b72c262cb17ddd6236c5d4970ecc09680c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79b8f6c016b43c3ac3ba7b22cca9a6b72c262cb17ddd6236c5d4970ecc09680c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38c4f218eb2ca7ea0c92bd0ae2ae9efeaa40746751608fbcf7f6c3f8d9e3c758\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38c4f218eb2ca7ea0c92bd0ae2ae9efeaa40746751608fbcf7f6c3f8d9e3c758\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d66ad9831d6029c8c350054525a00069e869d39c1b5638e06e9b772666bd8630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d66ad9831d6029c8c350054525a00069e869d39c1b5638e06e9b772666bd8630\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gjtnh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:28Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.490053 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-9rmnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f9bd12b-22ee-44ed-a8b9-c100d043f691\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2fgn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2fgn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:28Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-9rmnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:28Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.499118 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86d3005cd61c0973659e1c1c0d3ef1ddd6fcf95bede5465f4f2099e64a74bb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:28Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.508389 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:28Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.517352 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x94nb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75d52224135b75133fc000611cd179240b9174f6492d3263341716c7e42ccb15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgr8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x94nb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:28Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.538758 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.538777 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.538785 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.538796 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.538804 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:28Z","lastTransitionTime":"2026-01-21T11:11:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.640646 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.640732 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.640810 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.640874 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.640941 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:28Z","lastTransitionTime":"2026-01-21T11:11:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.731023 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.731665 4824 scope.go:117] "RemoveContainer" containerID="8e688b12ddb605e11511f11a6b41f84cd31b6065797f6b7332b1255595bd27cb" Jan 21 11:11:28 crc kubenswrapper[4824]: E0121 11:11:28.731852 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-htkvk_openshift-ovn-kubernetes(26b1108a-4b98-4e39-a3cd-e0d055089fd8)\"" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" podUID="26b1108a-4b98-4e39-a3cd-e0d055089fd8" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.743212 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.743244 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.743254 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.743266 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.743275 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:28Z","lastTransitionTime":"2026-01-21T11:11:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.844775 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.844808 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.844817 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.844828 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.844836 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:28Z","lastTransitionTime":"2026-01-21T11:11:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.921781 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8f9bd12b-22ee-44ed-a8b9-c100d043f691-metrics-certs\") pod \"network-metrics-daemon-9rmnw\" (UID: \"8f9bd12b-22ee-44ed-a8b9-c100d043f691\") " pod="openshift-multus/network-metrics-daemon-9rmnw" Jan 21 11:11:28 crc kubenswrapper[4824]: E0121 11:11:28.921892 4824 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Jan 21 11:11:28 crc kubenswrapper[4824]: E0121 11:11:28.921937 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8f9bd12b-22ee-44ed-a8b9-c100d043f691-metrics-certs podName:8f9bd12b-22ee-44ed-a8b9-c100d043f691 nodeName:}" failed. No retries permitted until 2026-01-21 11:11:29.921924535 +0000 UTC m=+32.214953827 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/8f9bd12b-22ee-44ed-a8b9-c100d043f691-metrics-certs") pod "network-metrics-daemon-9rmnw" (UID: "8f9bd12b-22ee-44ed-a8b9-c100d043f691") : object "openshift-multus"/"metrics-daemon-secret" not registered Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.946802 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.946836 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.946845 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.946857 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.946865 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:28Z","lastTransitionTime":"2026-01-21T11:11:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:28 crc kubenswrapper[4824]: I0121 11:11:28.997264 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-14 02:35:40.001102101 +0000 UTC Jan 21 11:11:29 crc kubenswrapper[4824]: I0121 11:11:29.048250 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 11:11:29 crc kubenswrapper[4824]: I0121 11:11:29.048311 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 11:11:29 crc kubenswrapper[4824]: E0121 11:11:29.048330 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 21 11:11:29 crc kubenswrapper[4824]: E0121 11:11:29.048406 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 21 11:11:29 crc kubenswrapper[4824]: I0121 11:11:29.048850 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:29 crc kubenswrapper[4824]: I0121 11:11:29.048876 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:29 crc kubenswrapper[4824]: I0121 11:11:29.048886 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:29 crc kubenswrapper[4824]: I0121 11:11:29.048905 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:29 crc kubenswrapper[4824]: I0121 11:11:29.048915 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:29Z","lastTransitionTime":"2026-01-21T11:11:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:29 crc kubenswrapper[4824]: I0121 11:11:29.151044 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:29 crc kubenswrapper[4824]: I0121 11:11:29.151082 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:29 crc kubenswrapper[4824]: I0121 11:11:29.151090 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:29 crc kubenswrapper[4824]: I0121 11:11:29.151113 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:29 crc kubenswrapper[4824]: I0121 11:11:29.151122 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:29Z","lastTransitionTime":"2026-01-21T11:11:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:29 crc kubenswrapper[4824]: I0121 11:11:29.253541 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:29 crc kubenswrapper[4824]: I0121 11:11:29.253572 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:29 crc kubenswrapper[4824]: I0121 11:11:29.253581 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:29 crc kubenswrapper[4824]: I0121 11:11:29.253591 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:29 crc kubenswrapper[4824]: I0121 11:11:29.253603 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:29Z","lastTransitionTime":"2026-01-21T11:11:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:29 crc kubenswrapper[4824]: I0121 11:11:29.355284 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:29 crc kubenswrapper[4824]: I0121 11:11:29.355306 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:29 crc kubenswrapper[4824]: I0121 11:11:29.355314 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:29 crc kubenswrapper[4824]: I0121 11:11:29.355323 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:29 crc kubenswrapper[4824]: I0121 11:11:29.355331 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:29Z","lastTransitionTime":"2026-01-21T11:11:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:29 crc kubenswrapper[4824]: I0121 11:11:29.456844 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:29 crc kubenswrapper[4824]: I0121 11:11:29.456875 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:29 crc kubenswrapper[4824]: I0121 11:11:29.456886 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:29 crc kubenswrapper[4824]: I0121 11:11:29.456898 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:29 crc kubenswrapper[4824]: I0121 11:11:29.456907 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:29Z","lastTransitionTime":"2026-01-21T11:11:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:29 crc kubenswrapper[4824]: I0121 11:11:29.558998 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:29 crc kubenswrapper[4824]: I0121 11:11:29.559038 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:29 crc kubenswrapper[4824]: I0121 11:11:29.559047 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:29 crc kubenswrapper[4824]: I0121 11:11:29.559060 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:29 crc kubenswrapper[4824]: I0121 11:11:29.559069 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:29Z","lastTransitionTime":"2026-01-21T11:11:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:29 crc kubenswrapper[4824]: I0121 11:11:29.661163 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:29 crc kubenswrapper[4824]: I0121 11:11:29.661198 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:29 crc kubenswrapper[4824]: I0121 11:11:29.661207 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:29 crc kubenswrapper[4824]: I0121 11:11:29.661218 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:29 crc kubenswrapper[4824]: I0121 11:11:29.661226 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:29Z","lastTransitionTime":"2026-01-21T11:11:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:29 crc kubenswrapper[4824]: I0121 11:11:29.763485 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:29 crc kubenswrapper[4824]: I0121 11:11:29.763529 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:29 crc kubenswrapper[4824]: I0121 11:11:29.763539 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:29 crc kubenswrapper[4824]: I0121 11:11:29.763552 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:29 crc kubenswrapper[4824]: I0121 11:11:29.763561 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:29Z","lastTransitionTime":"2026-01-21T11:11:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:29 crc kubenswrapper[4824]: I0121 11:11:29.865197 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:29 crc kubenswrapper[4824]: I0121 11:11:29.865231 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:29 crc kubenswrapper[4824]: I0121 11:11:29.865239 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:29 crc kubenswrapper[4824]: I0121 11:11:29.865251 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:29 crc kubenswrapper[4824]: I0121 11:11:29.865259 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:29Z","lastTransitionTime":"2026-01-21T11:11:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:29 crc kubenswrapper[4824]: I0121 11:11:29.930839 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8f9bd12b-22ee-44ed-a8b9-c100d043f691-metrics-certs\") pod \"network-metrics-daemon-9rmnw\" (UID: \"8f9bd12b-22ee-44ed-a8b9-c100d043f691\") " pod="openshift-multus/network-metrics-daemon-9rmnw" Jan 21 11:11:29 crc kubenswrapper[4824]: E0121 11:11:29.930929 4824 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Jan 21 11:11:29 crc kubenswrapper[4824]: E0121 11:11:29.930984 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8f9bd12b-22ee-44ed-a8b9-c100d043f691-metrics-certs podName:8f9bd12b-22ee-44ed-a8b9-c100d043f691 nodeName:}" failed. No retries permitted until 2026-01-21 11:11:31.930973149 +0000 UTC m=+34.224002442 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/8f9bd12b-22ee-44ed-a8b9-c100d043f691-metrics-certs") pod "network-metrics-daemon-9rmnw" (UID: "8f9bd12b-22ee-44ed-a8b9-c100d043f691") : object "openshift-multus"/"metrics-daemon-secret" not registered Jan 21 11:11:29 crc kubenswrapper[4824]: I0121 11:11:29.967184 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:29 crc kubenswrapper[4824]: I0121 11:11:29.967228 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:29 crc kubenswrapper[4824]: I0121 11:11:29.967238 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:29 crc kubenswrapper[4824]: I0121 11:11:29.967250 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:29 crc kubenswrapper[4824]: I0121 11:11:29.967258 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:29Z","lastTransitionTime":"2026-01-21T11:11:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:29 crc kubenswrapper[4824]: I0121 11:11:29.997429 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-02 14:38:13.779757788 +0000 UTC Jan 21 11:11:30 crc kubenswrapper[4824]: I0121 11:11:30.023019 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 21 11:11:30 crc kubenswrapper[4824]: I0121 11:11:30.032542 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:30Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:30 crc kubenswrapper[4824]: I0121 11:11:30.039985 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lv4rn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1267539-9adf-466f-8027-8ede009d19a0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eed1b884aa53e1ee318169a0cc2c9ad10130aa363ae4ac55e4bf065237756f53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbmfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lv4rn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:30Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:30 crc kubenswrapper[4824]: I0121 11:11:30.047061 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jr64s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9768c60d-5bee-4286-bced-076e584b04e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91158f14d2e98ac9f875a5e7101df3107ef28aa5fc5d62218f460af98ff920fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flrnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:17Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jr64s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:30Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:30 crc kubenswrapper[4824]: I0121 11:11:30.048437 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9rmnw" Jan 21 11:11:30 crc kubenswrapper[4824]: I0121 11:11:30.048480 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 11:11:30 crc kubenswrapper[4824]: E0121 11:11:30.048506 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9rmnw" podUID="8f9bd12b-22ee-44ed-a8b9-c100d043f691" Jan 21 11:11:30 crc kubenswrapper[4824]: E0121 11:11:30.048562 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 21 11:11:30 crc kubenswrapper[4824]: I0121 11:11:30.056079 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d31ab61-f94c-4f49-85d0-c9e91b83069b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a66b5c33884af00408dc35cf02f08ebd8570bd695f9f1fe558bf54fd06930431\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7514786602746dabfa690da48d48e233fe68d6dd6674aa5ecd489c0b3185f08\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61b14b6846f94cdeba8f1f61b3754e7663041f92c70356c4ee6cbb988a2e680f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a7023cfe4c098bcdd9c0477a8187f4ed7a0d2abdbe1e143436863948f441b4f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af11ad95bcf0c6bbd5fa4ab99bb592ce1c87f40d29c5ade7832414315820a569\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW0121 11:11:14.604290 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0121 11:11:14.604572 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0121 11:11:14.605875 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3284452394/tls.crt::/tmp/serving-cert-3284452394/tls.key\\\\\\\"\\\\nI0121 11:11:14.923427 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0121 11:11:14.928141 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0121 11:11:14.928162 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0121 11:11:14.928179 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0121 11:11:14.928184 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0121 11:11:14.933154 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0121 11:11:14.933230 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 11:11:14.933259 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 11:11:14.933282 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0121 11:11:14.933299 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0121 11:11:14.933315 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0121 11:11:14.933331 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0121 11:11:14.933177 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0121 11:11:14.934666 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48070c09fad511108ff22e44211cf26911509b8406d66948985cbfc3009d8233\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:30Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:30 crc kubenswrapper[4824]: I0121 11:11:30.064378 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9bb99c0a-683b-44da-9c0e-e8d284e87d53\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05e076febd697ed802ce5c7ee92a8c2c636a1a2c6609ad7eb01f17dace247c4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43529d1b9e1e5e22b0d56b023946119da8aacbae54b68c26ff0a1a82dc890c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ea232ad87393685aa323ce6393ff3869f8f8d423c2406a501c69c355535adfa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc6c7eff9e17b8a8ecc88bbfe703b5a01dbf1e02d342cede4f6a263a6526af43\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:30Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:30 crc kubenswrapper[4824]: I0121 11:11:30.068989 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:30 crc kubenswrapper[4824]: I0121 11:11:30.069022 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:30 crc kubenswrapper[4824]: I0121 11:11:30.069033 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:30 crc kubenswrapper[4824]: I0121 11:11:30.069046 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:30 crc kubenswrapper[4824]: I0121 11:11:30.069055 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:30Z","lastTransitionTime":"2026-01-21T11:11:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:30 crc kubenswrapper[4824]: I0121 11:11:30.076540 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06f56d86-ec6f-4b57-b314-f508c0dd4d9f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d25c94f1ae0bbabb733f0e1ea1d1ba39562a92af73047a39aacc03013e239681\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca06234486b59cf990c69d9501e23f935328e2ba21013cd71ed64831ea749bcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b63e45d66bcd11ff3479cc3ec6b450017da8d21d36e967b8d496ae0f24b0ee9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52758bcd31f2ba64edf03adfd4d1be6428015abff5b64c7c029fa787085f9c29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2736e73342d2adb175b0721dd4f40cc42d26fca06d5eb3334927518531c3b107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b35ad7dd0570c191eb0080739304f3cd43f17b3d1dddf895549441ebe50deaa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b35ad7dd0570c191eb0080739304f3cd43f17b3d1dddf895549441ebe50deaa6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://80cd5ea5e6fc19a88c1aab5c61f827f103ab798a5c84da8dc4d45cea86d5aacb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80cd5ea5e6fc19a88c1aab5c61f827f103ab798a5c84da8dc4d45cea86d5aacb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://234d8b914e6a88480021647ae65d74e9b21c55b6788312a094d00c8a42677d48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://234d8b914e6a88480021647ae65d74e9b21c55b6788312a094d00c8a42677d48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:30Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:30 crc kubenswrapper[4824]: I0121 11:11:30.084305 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:30Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:30 crc kubenswrapper[4824]: I0121 11:11:30.091951 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33f3d922-4ffe-409b-a49a-d88c85898260\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a9e57b39eed8300107e944b9fcf6a2ff1e75991f8f37387a3de615b14f9bc78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpwst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b1b37c6d2994021cfc30f85e92c40ae35677113b17aa066d52bc7a76ae5c153\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpwst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zqjg9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:30Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:30 crc kubenswrapper[4824]: I0121 11:11:30.101087 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gjtnh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a8d2cac-2d0f-4894-9bc0-2abd81699c47\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67c122a26811deaa9b70d19a1a65daee318ca3af2c626712482179d72fe67875\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2239186f1de46481cebfcafb1cab85cd320a205b5a0dfc6dd1b321d9baedeba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b2239186f1de46481cebfcafb1cab85cd320a205b5a0dfc6dd1b321d9baedeba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0db5ecd7759d974ae1b0d8065b3b8bd4d7957fae16eb5bab2d71c2fe69243816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0db5ecd7759d974ae1b0d8065b3b8bd4d7957fae16eb5bab2d71c2fe69243816\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba9b3278c208fc7aac630ca7a0daa3de0823f85d0799d6f1f278da80ae14c862\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba9b3278c208fc7aac630ca7a0daa3de0823f85d0799d6f1f278da80ae14c862\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79b8f6c016b43c3ac3ba7b22cca9a6b72c262cb17ddd6236c5d4970ecc09680c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79b8f6c016b43c3ac3ba7b22cca9a6b72c262cb17ddd6236c5d4970ecc09680c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38c4f218eb2ca7ea0c92bd0ae2ae9efeaa40746751608fbcf7f6c3f8d9e3c758\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38c4f218eb2ca7ea0c92bd0ae2ae9efeaa40746751608fbcf7f6c3f8d9e3c758\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d66ad9831d6029c8c350054525a00069e869d39c1b5638e06e9b772666bd8630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d66ad9831d6029c8c350054525a00069e869d39c1b5638e06e9b772666bd8630\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gjtnh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:30Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:30 crc kubenswrapper[4824]: I0121 11:11:30.108020 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-9rmnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f9bd12b-22ee-44ed-a8b9-c100d043f691\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2fgn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2fgn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:28Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-9rmnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:30Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:30 crc kubenswrapper[4824]: I0121 11:11:30.115943 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86d3005cd61c0973659e1c1c0d3ef1ddd6fcf95bede5465f4f2099e64a74bb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:30Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:30 crc kubenswrapper[4824]: I0121 11:11:30.127181 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:30Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:30 crc kubenswrapper[4824]: I0121 11:11:30.137243 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x94nb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75d52224135b75133fc000611cd179240b9174f6492d3263341716c7e42ccb15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgr8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x94nb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:30Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:30 crc kubenswrapper[4824]: I0121 11:11:30.144659 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-np2v8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4698abe4-0f33-4349-a2a4-614de80ae21b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://191328471bb6f2b34cde2eb927da697619826bf0b6174ecb583c8c25d27a6654\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n9qbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://050bd5bdded2162d5e9dac81713f4ed63cb575d1776e828b2f98e50d89f11b04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n9qbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-np2v8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:30Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:30 crc kubenswrapper[4824]: I0121 11:11:30.152191 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d87422473838f5218c9fb646ab2515f2f42592754ffdf07b19f7a8e093fa48a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:30Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:30 crc kubenswrapper[4824]: I0121 11:11:30.159990 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://516c009d4720ab5b393c7482872f12fed22d4f4414bd1f3400411c0e8da88ce4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee18d293036971832c8d9c3f7b9eca84ab779ff6768a2db3c6f7bdcf3433cdef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:30Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:30 crc kubenswrapper[4824]: I0121 11:11:30.171415 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:30 crc kubenswrapper[4824]: I0121 11:11:30.171445 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:30 crc kubenswrapper[4824]: I0121 11:11:30.171455 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:30 crc kubenswrapper[4824]: I0121 11:11:30.171468 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:30 crc kubenswrapper[4824]: I0121 11:11:30.171476 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:30Z","lastTransitionTime":"2026-01-21T11:11:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:30 crc kubenswrapper[4824]: I0121 11:11:30.173086 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"26b1108a-4b98-4e39-a3cd-e0d055089fd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc319bbc06568cc811ded2b418b04b986a31f321fed492e8be50104af79bb30b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cee9cc9eec3fc815f1621a22a37805691288dbb60dd3ff64194a84be62abcddc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc4c96e33e069f7d08d33caf92cd0987668506f931fd134b246d5ce5e934f95c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ddbb0e5ae02e9d5d82a80ead7efc91f1b8e557a9fb6f39940e5530fae8de2d38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://64a9065254f889b1c43d63759283335bf78fc4fa1531a4baae472c06c075dae3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e94ef8a8f47dc75ecfbab2e3fb8e7097c1b73eed32d7ca947f11ea198cea880\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e688b12ddb605e11511f11a6b41f84cd31b6065797f6b7332b1255595bd27cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e688b12ddb605e11511f11a6b41f84cd31b6065797f6b7332b1255595bd27cb\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-21T11:11:23Z\\\",\\\"message\\\":\\\" crc\\\\nF0121 11:11:23.774847 6243 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:23Z is after 2025-08-24T17:21:41Z]\\\\nI0121 11:11:23.774842 6243 services_controller.go:451] Built service openshift-kube-storage-version-migrator-operator/metrics cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-kube-storage-version-migrator-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-storage-version-migrator-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeO\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:23Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-htkvk_openshift-ovn-kubernetes(26b1108a-4b98-4e39-a3cd-e0d055089fd8)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948e0497eaffa7b0d84b4085c9ed8d987eb2c6527c9a82c10e576ed9c9c1e21b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-htkvk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:30Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:30 crc kubenswrapper[4824]: I0121 11:11:30.273397 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:30 crc kubenswrapper[4824]: I0121 11:11:30.273435 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:30 crc kubenswrapper[4824]: I0121 11:11:30.273443 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:30 crc kubenswrapper[4824]: I0121 11:11:30.273456 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:30 crc kubenswrapper[4824]: I0121 11:11:30.273464 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:30Z","lastTransitionTime":"2026-01-21T11:11:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:30 crc kubenswrapper[4824]: I0121 11:11:30.375277 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:30 crc kubenswrapper[4824]: I0121 11:11:30.375534 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:30 crc kubenswrapper[4824]: I0121 11:11:30.375604 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:30 crc kubenswrapper[4824]: I0121 11:11:30.375670 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:30 crc kubenswrapper[4824]: I0121 11:11:30.375747 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:30Z","lastTransitionTime":"2026-01-21T11:11:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:30 crc kubenswrapper[4824]: I0121 11:11:30.478380 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:30 crc kubenswrapper[4824]: I0121 11:11:30.478425 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:30 crc kubenswrapper[4824]: I0121 11:11:30.478435 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:30 crc kubenswrapper[4824]: I0121 11:11:30.478449 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:30 crc kubenswrapper[4824]: I0121 11:11:30.478458 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:30Z","lastTransitionTime":"2026-01-21T11:11:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:30 crc kubenswrapper[4824]: I0121 11:11:30.580443 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:30 crc kubenswrapper[4824]: I0121 11:11:30.580743 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:30 crc kubenswrapper[4824]: I0121 11:11:30.580753 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:30 crc kubenswrapper[4824]: I0121 11:11:30.580766 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:30 crc kubenswrapper[4824]: I0121 11:11:30.580776 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:30Z","lastTransitionTime":"2026-01-21T11:11:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:30 crc kubenswrapper[4824]: I0121 11:11:30.683137 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:30 crc kubenswrapper[4824]: I0121 11:11:30.683178 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:30 crc kubenswrapper[4824]: I0121 11:11:30.683190 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:30 crc kubenswrapper[4824]: I0121 11:11:30.683206 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:30 crc kubenswrapper[4824]: I0121 11:11:30.683216 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:30Z","lastTransitionTime":"2026-01-21T11:11:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:30 crc kubenswrapper[4824]: I0121 11:11:30.739742 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 21 11:11:30 crc kubenswrapper[4824]: I0121 11:11:30.739835 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 11:11:30 crc kubenswrapper[4824]: I0121 11:11:30.739894 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 11:11:30 crc kubenswrapper[4824]: E0121 11:11:30.740001 4824 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 21 11:11:30 crc kubenswrapper[4824]: E0121 11:11:30.740012 4824 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Jan 21 11:11:30 crc kubenswrapper[4824]: E0121 11:11:30.740000 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-21 11:11:46.739952343 +0000 UTC m=+49.032981645 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 11:11:30 crc kubenswrapper[4824]: E0121 11:11:30.740070 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-21 11:11:46.740054586 +0000 UTC m=+49.033083879 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 21 11:11:30 crc kubenswrapper[4824]: E0121 11:11:30.740082 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-21 11:11:46.740076047 +0000 UTC m=+49.033105339 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Jan 21 11:11:30 crc kubenswrapper[4824]: I0121 11:11:30.785412 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:30 crc kubenswrapper[4824]: I0121 11:11:30.785454 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:30 crc kubenswrapper[4824]: I0121 11:11:30.785463 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:30 crc kubenswrapper[4824]: I0121 11:11:30.785482 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:30 crc kubenswrapper[4824]: I0121 11:11:30.785490 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:30Z","lastTransitionTime":"2026-01-21T11:11:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:30 crc kubenswrapper[4824]: I0121 11:11:30.841275 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 11:11:30 crc kubenswrapper[4824]: I0121 11:11:30.841332 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 11:11:30 crc kubenswrapper[4824]: E0121 11:11:30.841455 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 21 11:11:30 crc kubenswrapper[4824]: E0121 11:11:30.841472 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 21 11:11:30 crc kubenswrapper[4824]: E0121 11:11:30.841482 4824 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 21 11:11:30 crc kubenswrapper[4824]: E0121 11:11:30.841487 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 21 11:11:30 crc kubenswrapper[4824]: E0121 11:11:30.841517 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 21 11:11:30 crc kubenswrapper[4824]: E0121 11:11:30.841526 4824 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 21 11:11:30 crc kubenswrapper[4824]: E0121 11:11:30.841531 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-01-21 11:11:46.841518704 +0000 UTC m=+49.134547996 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 21 11:11:30 crc kubenswrapper[4824]: E0121 11:11:30.841581 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-01-21 11:11:46.841564361 +0000 UTC m=+49.134593652 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 21 11:11:30 crc kubenswrapper[4824]: I0121 11:11:30.888217 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:30 crc kubenswrapper[4824]: I0121 11:11:30.888271 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:30 crc kubenswrapper[4824]: I0121 11:11:30.888282 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:30 crc kubenswrapper[4824]: I0121 11:11:30.888297 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:30 crc kubenswrapper[4824]: I0121 11:11:30.888313 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:30Z","lastTransitionTime":"2026-01-21T11:11:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:30 crc kubenswrapper[4824]: I0121 11:11:30.990443 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:30 crc kubenswrapper[4824]: I0121 11:11:30.990475 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:30 crc kubenswrapper[4824]: I0121 11:11:30.990484 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:30 crc kubenswrapper[4824]: I0121 11:11:30.990502 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:30 crc kubenswrapper[4824]: I0121 11:11:30.990511 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:30Z","lastTransitionTime":"2026-01-21T11:11:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:30 crc kubenswrapper[4824]: I0121 11:11:30.997498 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-30 06:33:00.023934931 +0000 UTC Jan 21 11:11:31 crc kubenswrapper[4824]: I0121 11:11:31.048434 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 11:11:31 crc kubenswrapper[4824]: E0121 11:11:31.048536 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 21 11:11:31 crc kubenswrapper[4824]: I0121 11:11:31.048435 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 11:11:31 crc kubenswrapper[4824]: E0121 11:11:31.048607 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 21 11:11:31 crc kubenswrapper[4824]: I0121 11:11:31.092917 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:31 crc kubenswrapper[4824]: I0121 11:11:31.092978 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:31 crc kubenswrapper[4824]: I0121 11:11:31.092989 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:31 crc kubenswrapper[4824]: I0121 11:11:31.093000 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:31 crc kubenswrapper[4824]: I0121 11:11:31.093008 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:31Z","lastTransitionTime":"2026-01-21T11:11:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:31 crc kubenswrapper[4824]: I0121 11:11:31.194605 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:31 crc kubenswrapper[4824]: I0121 11:11:31.194635 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:31 crc kubenswrapper[4824]: I0121 11:11:31.194644 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:31 crc kubenswrapper[4824]: I0121 11:11:31.194656 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:31 crc kubenswrapper[4824]: I0121 11:11:31.194664 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:31Z","lastTransitionTime":"2026-01-21T11:11:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:31 crc kubenswrapper[4824]: I0121 11:11:31.296810 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:31 crc kubenswrapper[4824]: I0121 11:11:31.296839 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:31 crc kubenswrapper[4824]: I0121 11:11:31.296848 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:31 crc kubenswrapper[4824]: I0121 11:11:31.296876 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:31 crc kubenswrapper[4824]: I0121 11:11:31.296885 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:31Z","lastTransitionTime":"2026-01-21T11:11:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:31 crc kubenswrapper[4824]: I0121 11:11:31.398854 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:31 crc kubenswrapper[4824]: I0121 11:11:31.398898 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:31 crc kubenswrapper[4824]: I0121 11:11:31.398907 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:31 crc kubenswrapper[4824]: I0121 11:11:31.398919 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:31 crc kubenswrapper[4824]: I0121 11:11:31.398927 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:31Z","lastTransitionTime":"2026-01-21T11:11:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:31 crc kubenswrapper[4824]: I0121 11:11:31.500405 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:31 crc kubenswrapper[4824]: I0121 11:11:31.500449 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:31 crc kubenswrapper[4824]: I0121 11:11:31.500456 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:31 crc kubenswrapper[4824]: I0121 11:11:31.500471 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:31 crc kubenswrapper[4824]: I0121 11:11:31.500481 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:31Z","lastTransitionTime":"2026-01-21T11:11:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:31 crc kubenswrapper[4824]: I0121 11:11:31.602569 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:31 crc kubenswrapper[4824]: I0121 11:11:31.602603 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:31 crc kubenswrapper[4824]: I0121 11:11:31.602611 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:31 crc kubenswrapper[4824]: I0121 11:11:31.602634 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:31 crc kubenswrapper[4824]: I0121 11:11:31.602642 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:31Z","lastTransitionTime":"2026-01-21T11:11:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:31 crc kubenswrapper[4824]: I0121 11:11:31.703868 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:31 crc kubenswrapper[4824]: I0121 11:11:31.703907 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:31 crc kubenswrapper[4824]: I0121 11:11:31.703916 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:31 crc kubenswrapper[4824]: I0121 11:11:31.703929 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:31 crc kubenswrapper[4824]: I0121 11:11:31.703937 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:31Z","lastTransitionTime":"2026-01-21T11:11:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:31 crc kubenswrapper[4824]: I0121 11:11:31.805769 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:31 crc kubenswrapper[4824]: I0121 11:11:31.805808 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:31 crc kubenswrapper[4824]: I0121 11:11:31.805816 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:31 crc kubenswrapper[4824]: I0121 11:11:31.805829 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:31 crc kubenswrapper[4824]: I0121 11:11:31.805838 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:31Z","lastTransitionTime":"2026-01-21T11:11:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:31 crc kubenswrapper[4824]: I0121 11:11:31.907651 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:31 crc kubenswrapper[4824]: I0121 11:11:31.907706 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:31 crc kubenswrapper[4824]: I0121 11:11:31.907717 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:31 crc kubenswrapper[4824]: I0121 11:11:31.907729 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:31 crc kubenswrapper[4824]: I0121 11:11:31.907738 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:31Z","lastTransitionTime":"2026-01-21T11:11:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:31 crc kubenswrapper[4824]: I0121 11:11:31.952615 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8f9bd12b-22ee-44ed-a8b9-c100d043f691-metrics-certs\") pod \"network-metrics-daemon-9rmnw\" (UID: \"8f9bd12b-22ee-44ed-a8b9-c100d043f691\") " pod="openshift-multus/network-metrics-daemon-9rmnw" Jan 21 11:11:31 crc kubenswrapper[4824]: E0121 11:11:31.952752 4824 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Jan 21 11:11:31 crc kubenswrapper[4824]: E0121 11:11:31.952797 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8f9bd12b-22ee-44ed-a8b9-c100d043f691-metrics-certs podName:8f9bd12b-22ee-44ed-a8b9-c100d043f691 nodeName:}" failed. No retries permitted until 2026-01-21 11:11:35.952786356 +0000 UTC m=+38.245815648 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/8f9bd12b-22ee-44ed-a8b9-c100d043f691-metrics-certs") pod "network-metrics-daemon-9rmnw" (UID: "8f9bd12b-22ee-44ed-a8b9-c100d043f691") : object "openshift-multus"/"metrics-daemon-secret" not registered Jan 21 11:11:31 crc kubenswrapper[4824]: I0121 11:11:31.998227 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-05 05:01:56.311426 +0000 UTC Jan 21 11:11:32 crc kubenswrapper[4824]: I0121 11:11:32.009630 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:32 crc kubenswrapper[4824]: I0121 11:11:32.009649 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:32 crc kubenswrapper[4824]: I0121 11:11:32.009660 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:32 crc kubenswrapper[4824]: I0121 11:11:32.009671 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:32 crc kubenswrapper[4824]: I0121 11:11:32.009680 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:32Z","lastTransitionTime":"2026-01-21T11:11:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:32 crc kubenswrapper[4824]: I0121 11:11:32.049205 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 11:11:32 crc kubenswrapper[4824]: I0121 11:11:32.049260 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9rmnw" Jan 21 11:11:32 crc kubenswrapper[4824]: E0121 11:11:32.049334 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 21 11:11:32 crc kubenswrapper[4824]: E0121 11:11:32.049408 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9rmnw" podUID="8f9bd12b-22ee-44ed-a8b9-c100d043f691" Jan 21 11:11:32 crc kubenswrapper[4824]: I0121 11:11:32.111681 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:32 crc kubenswrapper[4824]: I0121 11:11:32.111708 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:32 crc kubenswrapper[4824]: I0121 11:11:32.111716 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:32 crc kubenswrapper[4824]: I0121 11:11:32.111727 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:32 crc kubenswrapper[4824]: I0121 11:11:32.111735 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:32Z","lastTransitionTime":"2026-01-21T11:11:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:32 crc kubenswrapper[4824]: I0121 11:11:32.213309 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:32 crc kubenswrapper[4824]: I0121 11:11:32.213332 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:32 crc kubenswrapper[4824]: I0121 11:11:32.213340 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:32 crc kubenswrapper[4824]: I0121 11:11:32.213349 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:32 crc kubenswrapper[4824]: I0121 11:11:32.213356 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:32Z","lastTransitionTime":"2026-01-21T11:11:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:32 crc kubenswrapper[4824]: I0121 11:11:32.314947 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:32 crc kubenswrapper[4824]: I0121 11:11:32.314996 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:32 crc kubenswrapper[4824]: I0121 11:11:32.315005 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:32 crc kubenswrapper[4824]: I0121 11:11:32.315014 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:32 crc kubenswrapper[4824]: I0121 11:11:32.315021 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:32Z","lastTransitionTime":"2026-01-21T11:11:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:32 crc kubenswrapper[4824]: I0121 11:11:32.417109 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:32 crc kubenswrapper[4824]: I0121 11:11:32.417151 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:32 crc kubenswrapper[4824]: I0121 11:11:32.417161 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:32 crc kubenswrapper[4824]: I0121 11:11:32.417174 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:32 crc kubenswrapper[4824]: I0121 11:11:32.417183 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:32Z","lastTransitionTime":"2026-01-21T11:11:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:32 crc kubenswrapper[4824]: I0121 11:11:32.519578 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:32 crc kubenswrapper[4824]: I0121 11:11:32.519618 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:32 crc kubenswrapper[4824]: I0121 11:11:32.519629 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:32 crc kubenswrapper[4824]: I0121 11:11:32.519644 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:32 crc kubenswrapper[4824]: I0121 11:11:32.519652 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:32Z","lastTransitionTime":"2026-01-21T11:11:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:32 crc kubenswrapper[4824]: I0121 11:11:32.621429 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:32 crc kubenswrapper[4824]: I0121 11:11:32.621461 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:32 crc kubenswrapper[4824]: I0121 11:11:32.621468 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:32 crc kubenswrapper[4824]: I0121 11:11:32.621481 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:32 crc kubenswrapper[4824]: I0121 11:11:32.621490 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:32Z","lastTransitionTime":"2026-01-21T11:11:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:32 crc kubenswrapper[4824]: I0121 11:11:32.723048 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:32 crc kubenswrapper[4824]: I0121 11:11:32.723121 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:32 crc kubenswrapper[4824]: I0121 11:11:32.723148 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:32 crc kubenswrapper[4824]: I0121 11:11:32.723162 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:32 crc kubenswrapper[4824]: I0121 11:11:32.723173 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:32Z","lastTransitionTime":"2026-01-21T11:11:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:32 crc kubenswrapper[4824]: I0121 11:11:32.825102 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:32 crc kubenswrapper[4824]: I0121 11:11:32.825145 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:32 crc kubenswrapper[4824]: I0121 11:11:32.825153 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:32 crc kubenswrapper[4824]: I0121 11:11:32.825166 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:32 crc kubenswrapper[4824]: I0121 11:11:32.825175 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:32Z","lastTransitionTime":"2026-01-21T11:11:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:32 crc kubenswrapper[4824]: I0121 11:11:32.926772 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:32 crc kubenswrapper[4824]: I0121 11:11:32.926804 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:32 crc kubenswrapper[4824]: I0121 11:11:32.926813 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:32 crc kubenswrapper[4824]: I0121 11:11:32.926827 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:32 crc kubenswrapper[4824]: I0121 11:11:32.926837 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:32Z","lastTransitionTime":"2026-01-21T11:11:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:32 crc kubenswrapper[4824]: I0121 11:11:32.998521 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-02 11:08:18.52546233 +0000 UTC Jan 21 11:11:33 crc kubenswrapper[4824]: I0121 11:11:33.028985 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:33 crc kubenswrapper[4824]: I0121 11:11:33.029013 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:33 crc kubenswrapper[4824]: I0121 11:11:33.029020 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:33 crc kubenswrapper[4824]: I0121 11:11:33.029030 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:33 crc kubenswrapper[4824]: I0121 11:11:33.029039 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:33Z","lastTransitionTime":"2026-01-21T11:11:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:33 crc kubenswrapper[4824]: I0121 11:11:33.048397 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 11:11:33 crc kubenswrapper[4824]: I0121 11:11:33.048417 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 11:11:33 crc kubenswrapper[4824]: E0121 11:11:33.048477 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 21 11:11:33 crc kubenswrapper[4824]: E0121 11:11:33.048553 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 21 11:11:33 crc kubenswrapper[4824]: I0121 11:11:33.130341 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:33 crc kubenswrapper[4824]: I0121 11:11:33.130392 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:33 crc kubenswrapper[4824]: I0121 11:11:33.130402 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:33 crc kubenswrapper[4824]: I0121 11:11:33.130416 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:33 crc kubenswrapper[4824]: I0121 11:11:33.130425 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:33Z","lastTransitionTime":"2026-01-21T11:11:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:33 crc kubenswrapper[4824]: I0121 11:11:33.232228 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:33 crc kubenswrapper[4824]: I0121 11:11:33.232269 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:33 crc kubenswrapper[4824]: I0121 11:11:33.232277 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:33 crc kubenswrapper[4824]: I0121 11:11:33.232290 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:33 crc kubenswrapper[4824]: I0121 11:11:33.232299 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:33Z","lastTransitionTime":"2026-01-21T11:11:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:33 crc kubenswrapper[4824]: I0121 11:11:33.333608 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:33 crc kubenswrapper[4824]: I0121 11:11:33.333637 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:33 crc kubenswrapper[4824]: I0121 11:11:33.333645 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:33 crc kubenswrapper[4824]: I0121 11:11:33.333657 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:33 crc kubenswrapper[4824]: I0121 11:11:33.333665 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:33Z","lastTransitionTime":"2026-01-21T11:11:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:33 crc kubenswrapper[4824]: I0121 11:11:33.435453 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:33 crc kubenswrapper[4824]: I0121 11:11:33.435473 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:33 crc kubenswrapper[4824]: I0121 11:11:33.435482 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:33 crc kubenswrapper[4824]: I0121 11:11:33.435491 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:33 crc kubenswrapper[4824]: I0121 11:11:33.435499 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:33Z","lastTransitionTime":"2026-01-21T11:11:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:33 crc kubenswrapper[4824]: I0121 11:11:33.537505 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:33 crc kubenswrapper[4824]: I0121 11:11:33.537544 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:33 crc kubenswrapper[4824]: I0121 11:11:33.537553 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:33 crc kubenswrapper[4824]: I0121 11:11:33.537569 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:33 crc kubenswrapper[4824]: I0121 11:11:33.537577 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:33Z","lastTransitionTime":"2026-01-21T11:11:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:33 crc kubenswrapper[4824]: I0121 11:11:33.639336 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:33 crc kubenswrapper[4824]: I0121 11:11:33.639372 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:33 crc kubenswrapper[4824]: I0121 11:11:33.639381 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:33 crc kubenswrapper[4824]: I0121 11:11:33.639394 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:33 crc kubenswrapper[4824]: I0121 11:11:33.639402 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:33Z","lastTransitionTime":"2026-01-21T11:11:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:33 crc kubenswrapper[4824]: I0121 11:11:33.741300 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:33 crc kubenswrapper[4824]: I0121 11:11:33.741331 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:33 crc kubenswrapper[4824]: I0121 11:11:33.741340 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:33 crc kubenswrapper[4824]: I0121 11:11:33.741354 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:33 crc kubenswrapper[4824]: I0121 11:11:33.741362 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:33Z","lastTransitionTime":"2026-01-21T11:11:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:33 crc kubenswrapper[4824]: I0121 11:11:33.843135 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:33 crc kubenswrapper[4824]: I0121 11:11:33.843179 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:33 crc kubenswrapper[4824]: I0121 11:11:33.843188 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:33 crc kubenswrapper[4824]: I0121 11:11:33.843198 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:33 crc kubenswrapper[4824]: I0121 11:11:33.843205 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:33Z","lastTransitionTime":"2026-01-21T11:11:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:33 crc kubenswrapper[4824]: I0121 11:11:33.945225 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:33 crc kubenswrapper[4824]: I0121 11:11:33.945259 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:33 crc kubenswrapper[4824]: I0121 11:11:33.945268 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:33 crc kubenswrapper[4824]: I0121 11:11:33.945281 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:33 crc kubenswrapper[4824]: I0121 11:11:33.945288 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:33Z","lastTransitionTime":"2026-01-21T11:11:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:33 crc kubenswrapper[4824]: I0121 11:11:33.998819 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-18 06:43:51.897008128 +0000 UTC Jan 21 11:11:34 crc kubenswrapper[4824]: I0121 11:11:34.047407 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:34 crc kubenswrapper[4824]: I0121 11:11:34.047446 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:34 crc kubenswrapper[4824]: I0121 11:11:34.047457 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:34 crc kubenswrapper[4824]: I0121 11:11:34.047473 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:34 crc kubenswrapper[4824]: I0121 11:11:34.047483 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:34Z","lastTransitionTime":"2026-01-21T11:11:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:34 crc kubenswrapper[4824]: I0121 11:11:34.048546 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 11:11:34 crc kubenswrapper[4824]: I0121 11:11:34.048562 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9rmnw" Jan 21 11:11:34 crc kubenswrapper[4824]: E0121 11:11:34.048628 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 21 11:11:34 crc kubenswrapper[4824]: E0121 11:11:34.048685 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9rmnw" podUID="8f9bd12b-22ee-44ed-a8b9-c100d043f691" Jan 21 11:11:34 crc kubenswrapper[4824]: I0121 11:11:34.149328 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:34 crc kubenswrapper[4824]: I0121 11:11:34.149357 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:34 crc kubenswrapper[4824]: I0121 11:11:34.149365 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:34 crc kubenswrapper[4824]: I0121 11:11:34.149375 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:34 crc kubenswrapper[4824]: I0121 11:11:34.149383 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:34Z","lastTransitionTime":"2026-01-21T11:11:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:34 crc kubenswrapper[4824]: I0121 11:11:34.251478 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:34 crc kubenswrapper[4824]: I0121 11:11:34.251509 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:34 crc kubenswrapper[4824]: I0121 11:11:34.251518 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:34 crc kubenswrapper[4824]: I0121 11:11:34.251532 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:34 crc kubenswrapper[4824]: I0121 11:11:34.251544 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:34Z","lastTransitionTime":"2026-01-21T11:11:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:34 crc kubenswrapper[4824]: I0121 11:11:34.353000 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:34 crc kubenswrapper[4824]: I0121 11:11:34.353030 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:34 crc kubenswrapper[4824]: I0121 11:11:34.353040 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:34 crc kubenswrapper[4824]: I0121 11:11:34.353052 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:34 crc kubenswrapper[4824]: I0121 11:11:34.353061 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:34Z","lastTransitionTime":"2026-01-21T11:11:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:34 crc kubenswrapper[4824]: I0121 11:11:34.455325 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:34 crc kubenswrapper[4824]: I0121 11:11:34.455354 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:34 crc kubenswrapper[4824]: I0121 11:11:34.455363 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:34 crc kubenswrapper[4824]: I0121 11:11:34.455376 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:34 crc kubenswrapper[4824]: I0121 11:11:34.455385 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:34Z","lastTransitionTime":"2026-01-21T11:11:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:34 crc kubenswrapper[4824]: I0121 11:11:34.556763 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:34 crc kubenswrapper[4824]: I0121 11:11:34.556788 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:34 crc kubenswrapper[4824]: I0121 11:11:34.556797 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:34 crc kubenswrapper[4824]: I0121 11:11:34.556807 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:34 crc kubenswrapper[4824]: I0121 11:11:34.556814 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:34Z","lastTransitionTime":"2026-01-21T11:11:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:34 crc kubenswrapper[4824]: I0121 11:11:34.659252 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:34 crc kubenswrapper[4824]: I0121 11:11:34.659302 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:34 crc kubenswrapper[4824]: I0121 11:11:34.659312 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:34 crc kubenswrapper[4824]: I0121 11:11:34.659324 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:34 crc kubenswrapper[4824]: I0121 11:11:34.659333 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:34Z","lastTransitionTime":"2026-01-21T11:11:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:34 crc kubenswrapper[4824]: I0121 11:11:34.760886 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:34 crc kubenswrapper[4824]: I0121 11:11:34.760911 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:34 crc kubenswrapper[4824]: I0121 11:11:34.760919 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:34 crc kubenswrapper[4824]: I0121 11:11:34.760931 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:34 crc kubenswrapper[4824]: I0121 11:11:34.760940 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:34Z","lastTransitionTime":"2026-01-21T11:11:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:34 crc kubenswrapper[4824]: I0121 11:11:34.862457 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:34 crc kubenswrapper[4824]: I0121 11:11:34.862480 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:34 crc kubenswrapper[4824]: I0121 11:11:34.862488 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:34 crc kubenswrapper[4824]: I0121 11:11:34.862498 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:34 crc kubenswrapper[4824]: I0121 11:11:34.862505 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:34Z","lastTransitionTime":"2026-01-21T11:11:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:34 crc kubenswrapper[4824]: I0121 11:11:34.964136 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:34 crc kubenswrapper[4824]: I0121 11:11:34.964178 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:34 crc kubenswrapper[4824]: I0121 11:11:34.964187 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:34 crc kubenswrapper[4824]: I0121 11:11:34.964198 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:34 crc kubenswrapper[4824]: I0121 11:11:34.964206 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:34Z","lastTransitionTime":"2026-01-21T11:11:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:34 crc kubenswrapper[4824]: I0121 11:11:34.998905 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-14 21:12:04.790777236 +0000 UTC Jan 21 11:11:35 crc kubenswrapper[4824]: I0121 11:11:35.048488 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 11:11:35 crc kubenswrapper[4824]: E0121 11:11:35.048622 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 21 11:11:35 crc kubenswrapper[4824]: I0121 11:11:35.048492 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 11:11:35 crc kubenswrapper[4824]: E0121 11:11:35.048740 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 21 11:11:35 crc kubenswrapper[4824]: I0121 11:11:35.066324 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:35 crc kubenswrapper[4824]: I0121 11:11:35.066355 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:35 crc kubenswrapper[4824]: I0121 11:11:35.066365 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:35 crc kubenswrapper[4824]: I0121 11:11:35.066379 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:35 crc kubenswrapper[4824]: I0121 11:11:35.066388 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:35Z","lastTransitionTime":"2026-01-21T11:11:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:35 crc kubenswrapper[4824]: I0121 11:11:35.168314 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:35 crc kubenswrapper[4824]: I0121 11:11:35.168339 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:35 crc kubenswrapper[4824]: I0121 11:11:35.168347 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:35 crc kubenswrapper[4824]: I0121 11:11:35.168357 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:35 crc kubenswrapper[4824]: I0121 11:11:35.168364 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:35Z","lastTransitionTime":"2026-01-21T11:11:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:35 crc kubenswrapper[4824]: I0121 11:11:35.269751 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:35 crc kubenswrapper[4824]: I0121 11:11:35.269784 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:35 crc kubenswrapper[4824]: I0121 11:11:35.269792 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:35 crc kubenswrapper[4824]: I0121 11:11:35.269804 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:35 crc kubenswrapper[4824]: I0121 11:11:35.269814 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:35Z","lastTransitionTime":"2026-01-21T11:11:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:35 crc kubenswrapper[4824]: I0121 11:11:35.372060 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:35 crc kubenswrapper[4824]: I0121 11:11:35.372096 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:35 crc kubenswrapper[4824]: I0121 11:11:35.372107 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:35 crc kubenswrapper[4824]: I0121 11:11:35.372121 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:35 crc kubenswrapper[4824]: I0121 11:11:35.372131 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:35Z","lastTransitionTime":"2026-01-21T11:11:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:35 crc kubenswrapper[4824]: I0121 11:11:35.473577 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:35 crc kubenswrapper[4824]: I0121 11:11:35.473614 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:35 crc kubenswrapper[4824]: I0121 11:11:35.473622 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:35 crc kubenswrapper[4824]: I0121 11:11:35.473637 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:35 crc kubenswrapper[4824]: I0121 11:11:35.473646 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:35Z","lastTransitionTime":"2026-01-21T11:11:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:35 crc kubenswrapper[4824]: I0121 11:11:35.575531 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:35 crc kubenswrapper[4824]: I0121 11:11:35.575558 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:35 crc kubenswrapper[4824]: I0121 11:11:35.575566 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:35 crc kubenswrapper[4824]: I0121 11:11:35.575576 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:35 crc kubenswrapper[4824]: I0121 11:11:35.575584 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:35Z","lastTransitionTime":"2026-01-21T11:11:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:35 crc kubenswrapper[4824]: I0121 11:11:35.677914 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:35 crc kubenswrapper[4824]: I0121 11:11:35.677944 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:35 crc kubenswrapper[4824]: I0121 11:11:35.677952 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:35 crc kubenswrapper[4824]: I0121 11:11:35.677975 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:35 crc kubenswrapper[4824]: I0121 11:11:35.677984 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:35Z","lastTransitionTime":"2026-01-21T11:11:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:35 crc kubenswrapper[4824]: I0121 11:11:35.779771 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:35 crc kubenswrapper[4824]: I0121 11:11:35.779803 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:35 crc kubenswrapper[4824]: I0121 11:11:35.779811 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:35 crc kubenswrapper[4824]: I0121 11:11:35.779824 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:35 crc kubenswrapper[4824]: I0121 11:11:35.779833 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:35Z","lastTransitionTime":"2026-01-21T11:11:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:35 crc kubenswrapper[4824]: I0121 11:11:35.881081 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:35 crc kubenswrapper[4824]: I0121 11:11:35.881110 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:35 crc kubenswrapper[4824]: I0121 11:11:35.881119 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:35 crc kubenswrapper[4824]: I0121 11:11:35.881129 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:35 crc kubenswrapper[4824]: I0121 11:11:35.881137 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:35Z","lastTransitionTime":"2026-01-21T11:11:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:35 crc kubenswrapper[4824]: I0121 11:11:35.982736 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:35 crc kubenswrapper[4824]: I0121 11:11:35.982775 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:35 crc kubenswrapper[4824]: I0121 11:11:35.982785 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:35 crc kubenswrapper[4824]: I0121 11:11:35.982798 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:35 crc kubenswrapper[4824]: I0121 11:11:35.982815 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:35Z","lastTransitionTime":"2026-01-21T11:11:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:35 crc kubenswrapper[4824]: I0121 11:11:35.986122 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8f9bd12b-22ee-44ed-a8b9-c100d043f691-metrics-certs\") pod \"network-metrics-daemon-9rmnw\" (UID: \"8f9bd12b-22ee-44ed-a8b9-c100d043f691\") " pod="openshift-multus/network-metrics-daemon-9rmnw" Jan 21 11:11:35 crc kubenswrapper[4824]: E0121 11:11:35.986225 4824 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Jan 21 11:11:35 crc kubenswrapper[4824]: E0121 11:11:35.986270 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8f9bd12b-22ee-44ed-a8b9-c100d043f691-metrics-certs podName:8f9bd12b-22ee-44ed-a8b9-c100d043f691 nodeName:}" failed. No retries permitted until 2026-01-21 11:11:43.986258054 +0000 UTC m=+46.279287346 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/8f9bd12b-22ee-44ed-a8b9-c100d043f691-metrics-certs") pod "network-metrics-daemon-9rmnw" (UID: "8f9bd12b-22ee-44ed-a8b9-c100d043f691") : object "openshift-multus"/"metrics-daemon-secret" not registered Jan 21 11:11:35 crc kubenswrapper[4824]: I0121 11:11:35.999493 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-09 19:38:12.287513634 +0000 UTC Jan 21 11:11:36 crc kubenswrapper[4824]: I0121 11:11:36.048943 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9rmnw" Jan 21 11:11:36 crc kubenswrapper[4824]: I0121 11:11:36.048949 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 11:11:36 crc kubenswrapper[4824]: E0121 11:11:36.049158 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 21 11:11:36 crc kubenswrapper[4824]: E0121 11:11:36.049065 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9rmnw" podUID="8f9bd12b-22ee-44ed-a8b9-c100d043f691" Jan 21 11:11:36 crc kubenswrapper[4824]: I0121 11:11:36.084704 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:36 crc kubenswrapper[4824]: I0121 11:11:36.084734 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:36 crc kubenswrapper[4824]: I0121 11:11:36.084744 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:36 crc kubenswrapper[4824]: I0121 11:11:36.084754 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:36 crc kubenswrapper[4824]: I0121 11:11:36.084763 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:36Z","lastTransitionTime":"2026-01-21T11:11:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:36 crc kubenswrapper[4824]: I0121 11:11:36.186142 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:36 crc kubenswrapper[4824]: I0121 11:11:36.186187 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:36 crc kubenswrapper[4824]: I0121 11:11:36.186196 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:36 crc kubenswrapper[4824]: I0121 11:11:36.186208 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:36 crc kubenswrapper[4824]: I0121 11:11:36.186217 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:36Z","lastTransitionTime":"2026-01-21T11:11:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:36 crc kubenswrapper[4824]: I0121 11:11:36.287539 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:36 crc kubenswrapper[4824]: I0121 11:11:36.287567 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:36 crc kubenswrapper[4824]: I0121 11:11:36.287576 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:36 crc kubenswrapper[4824]: I0121 11:11:36.287586 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:36 crc kubenswrapper[4824]: I0121 11:11:36.287594 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:36Z","lastTransitionTime":"2026-01-21T11:11:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:36 crc kubenswrapper[4824]: I0121 11:11:36.389074 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:36 crc kubenswrapper[4824]: I0121 11:11:36.389110 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:36 crc kubenswrapper[4824]: I0121 11:11:36.389118 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:36 crc kubenswrapper[4824]: I0121 11:11:36.389132 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:36 crc kubenswrapper[4824]: I0121 11:11:36.389140 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:36Z","lastTransitionTime":"2026-01-21T11:11:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:36 crc kubenswrapper[4824]: I0121 11:11:36.490990 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:36 crc kubenswrapper[4824]: I0121 11:11:36.491013 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:36 crc kubenswrapper[4824]: I0121 11:11:36.491022 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:36 crc kubenswrapper[4824]: I0121 11:11:36.491033 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:36 crc kubenswrapper[4824]: I0121 11:11:36.491041 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:36Z","lastTransitionTime":"2026-01-21T11:11:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:36 crc kubenswrapper[4824]: I0121 11:11:36.591988 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:36 crc kubenswrapper[4824]: I0121 11:11:36.592011 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:36 crc kubenswrapper[4824]: I0121 11:11:36.592020 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:36 crc kubenswrapper[4824]: I0121 11:11:36.592032 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:36 crc kubenswrapper[4824]: I0121 11:11:36.592040 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:36Z","lastTransitionTime":"2026-01-21T11:11:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:36 crc kubenswrapper[4824]: I0121 11:11:36.693345 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:36 crc kubenswrapper[4824]: I0121 11:11:36.693375 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:36 crc kubenswrapper[4824]: I0121 11:11:36.693384 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:36 crc kubenswrapper[4824]: I0121 11:11:36.693395 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:36 crc kubenswrapper[4824]: I0121 11:11:36.693403 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:36Z","lastTransitionTime":"2026-01-21T11:11:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:36 crc kubenswrapper[4824]: I0121 11:11:36.795211 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:36 crc kubenswrapper[4824]: I0121 11:11:36.795240 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:36 crc kubenswrapper[4824]: I0121 11:11:36.795249 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:36 crc kubenswrapper[4824]: I0121 11:11:36.795260 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:36 crc kubenswrapper[4824]: I0121 11:11:36.795268 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:36Z","lastTransitionTime":"2026-01-21T11:11:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:36 crc kubenswrapper[4824]: I0121 11:11:36.896646 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:36 crc kubenswrapper[4824]: I0121 11:11:36.896673 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:36 crc kubenswrapper[4824]: I0121 11:11:36.896680 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:36 crc kubenswrapper[4824]: I0121 11:11:36.896689 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:36 crc kubenswrapper[4824]: I0121 11:11:36.896698 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:36Z","lastTransitionTime":"2026-01-21T11:11:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:36 crc kubenswrapper[4824]: I0121 11:11:36.998706 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:36 crc kubenswrapper[4824]: I0121 11:11:36.998737 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:36 crc kubenswrapper[4824]: I0121 11:11:36.998744 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:36 crc kubenswrapper[4824]: I0121 11:11:36.998755 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:36 crc kubenswrapper[4824]: I0121 11:11:36.998763 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:36Z","lastTransitionTime":"2026-01-21T11:11:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:36 crc kubenswrapper[4824]: I0121 11:11:36.999809 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-18 11:30:17.363330206 +0000 UTC Jan 21 11:11:37 crc kubenswrapper[4824]: I0121 11:11:37.049292 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 11:11:37 crc kubenswrapper[4824]: I0121 11:11:37.049306 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 11:11:37 crc kubenswrapper[4824]: E0121 11:11:37.049385 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 21 11:11:37 crc kubenswrapper[4824]: E0121 11:11:37.049466 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 21 11:11:37 crc kubenswrapper[4824]: I0121 11:11:37.100162 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:37 crc kubenswrapper[4824]: I0121 11:11:37.100202 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:37 crc kubenswrapper[4824]: I0121 11:11:37.100211 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:37 crc kubenswrapper[4824]: I0121 11:11:37.100221 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:37 crc kubenswrapper[4824]: I0121 11:11:37.100230 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:37Z","lastTransitionTime":"2026-01-21T11:11:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:37 crc kubenswrapper[4824]: I0121 11:11:37.201641 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:37 crc kubenswrapper[4824]: I0121 11:11:37.201663 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:37 crc kubenswrapper[4824]: I0121 11:11:37.201670 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:37 crc kubenswrapper[4824]: I0121 11:11:37.201680 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:37 crc kubenswrapper[4824]: I0121 11:11:37.201688 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:37Z","lastTransitionTime":"2026-01-21T11:11:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:37 crc kubenswrapper[4824]: I0121 11:11:37.303669 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:37 crc kubenswrapper[4824]: I0121 11:11:37.303689 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:37 crc kubenswrapper[4824]: I0121 11:11:37.303696 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:37 crc kubenswrapper[4824]: I0121 11:11:37.303705 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:37 crc kubenswrapper[4824]: I0121 11:11:37.303713 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:37Z","lastTransitionTime":"2026-01-21T11:11:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:37 crc kubenswrapper[4824]: I0121 11:11:37.405553 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:37 crc kubenswrapper[4824]: I0121 11:11:37.405580 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:37 crc kubenswrapper[4824]: I0121 11:11:37.405593 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:37 crc kubenswrapper[4824]: I0121 11:11:37.405603 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:37 crc kubenswrapper[4824]: I0121 11:11:37.405609 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:37Z","lastTransitionTime":"2026-01-21T11:11:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:37 crc kubenswrapper[4824]: I0121 11:11:37.507813 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:37 crc kubenswrapper[4824]: I0121 11:11:37.507840 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:37 crc kubenswrapper[4824]: I0121 11:11:37.507848 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:37 crc kubenswrapper[4824]: I0121 11:11:37.507857 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:37 crc kubenswrapper[4824]: I0121 11:11:37.507864 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:37Z","lastTransitionTime":"2026-01-21T11:11:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:37 crc kubenswrapper[4824]: I0121 11:11:37.609436 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:37 crc kubenswrapper[4824]: I0121 11:11:37.609460 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:37 crc kubenswrapper[4824]: I0121 11:11:37.609468 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:37 crc kubenswrapper[4824]: I0121 11:11:37.609477 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:37 crc kubenswrapper[4824]: I0121 11:11:37.609484 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:37Z","lastTransitionTime":"2026-01-21T11:11:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:37 crc kubenswrapper[4824]: I0121 11:11:37.711479 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:37 crc kubenswrapper[4824]: I0121 11:11:37.711520 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:37 crc kubenswrapper[4824]: I0121 11:11:37.711538 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:37 crc kubenswrapper[4824]: I0121 11:11:37.711552 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:37 crc kubenswrapper[4824]: I0121 11:11:37.711565 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:37Z","lastTransitionTime":"2026-01-21T11:11:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:37 crc kubenswrapper[4824]: I0121 11:11:37.813276 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:37 crc kubenswrapper[4824]: I0121 11:11:37.813306 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:37 crc kubenswrapper[4824]: I0121 11:11:37.813315 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:37 crc kubenswrapper[4824]: I0121 11:11:37.813326 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:37 crc kubenswrapper[4824]: I0121 11:11:37.813338 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:37Z","lastTransitionTime":"2026-01-21T11:11:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:37 crc kubenswrapper[4824]: I0121 11:11:37.914696 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:37 crc kubenswrapper[4824]: I0121 11:11:37.914740 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:37 crc kubenswrapper[4824]: I0121 11:11:37.914749 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:37 crc kubenswrapper[4824]: I0121 11:11:37.914768 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:37 crc kubenswrapper[4824]: I0121 11:11:37.914775 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:37Z","lastTransitionTime":"2026-01-21T11:11:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.000406 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-28 01:34:46.449565449 +0000 UTC Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.016691 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.016714 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.016722 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.016733 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.016740 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:38Z","lastTransitionTime":"2026-01-21T11:11:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.049103 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9rmnw" Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.049197 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 11:11:38 crc kubenswrapper[4824]: E0121 11:11:38.049284 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9rmnw" podUID="8f9bd12b-22ee-44ed-a8b9-c100d043f691" Jan 21 11:11:38 crc kubenswrapper[4824]: E0121 11:11:38.049369 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.058676 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86d3005cd61c0973659e1c1c0d3ef1ddd6fcf95bede5465f4f2099e64a74bb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:38Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.065778 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:38Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.073215 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x94nb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75d52224135b75133fc000611cd179240b9174f6492d3263341716c7e42ccb15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgr8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x94nb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:38Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.080224 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33f3d922-4ffe-409b-a49a-d88c85898260\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a9e57b39eed8300107e944b9fcf6a2ff1e75991f8f37387a3de615b14f9bc78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpwst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b1b37c6d2994021cfc30f85e92c40ae35677113b17aa066d52bc7a76ae5c153\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpwst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zqjg9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:38Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.089950 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gjtnh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a8d2cac-2d0f-4894-9bc0-2abd81699c47\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67c122a26811deaa9b70d19a1a65daee318ca3af2c626712482179d72fe67875\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2239186f1de46481cebfcafb1cab85cd320a205b5a0dfc6dd1b321d9baedeba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b2239186f1de46481cebfcafb1cab85cd320a205b5a0dfc6dd1b321d9baedeba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0db5ecd7759d974ae1b0d8065b3b8bd4d7957fae16eb5bab2d71c2fe69243816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0db5ecd7759d974ae1b0d8065b3b8bd4d7957fae16eb5bab2d71c2fe69243816\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba9b3278c208fc7aac630ca7a0daa3de0823f85d0799d6f1f278da80ae14c862\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba9b3278c208fc7aac630ca7a0daa3de0823f85d0799d6f1f278da80ae14c862\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79b8f6c016b43c3ac3ba7b22cca9a6b72c262cb17ddd6236c5d4970ecc09680c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79b8f6c016b43c3ac3ba7b22cca9a6b72c262cb17ddd6236c5d4970ecc09680c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38c4f218eb2ca7ea0c92bd0ae2ae9efeaa40746751608fbcf7f6c3f8d9e3c758\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38c4f218eb2ca7ea0c92bd0ae2ae9efeaa40746751608fbcf7f6c3f8d9e3c758\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d66ad9831d6029c8c350054525a00069e869d39c1b5638e06e9b772666bd8630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d66ad9831d6029c8c350054525a00069e869d39c1b5638e06e9b772666bd8630\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gjtnh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:38Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.097749 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-9rmnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f9bd12b-22ee-44ed-a8b9-c100d043f691\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2fgn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2fgn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:28Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-9rmnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:38Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.106718 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d87422473838f5218c9fb646ab2515f2f42592754ffdf07b19f7a8e093fa48a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:38Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.114393 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://516c009d4720ab5b393c7482872f12fed22d4f4414bd1f3400411c0e8da88ce4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee18d293036971832c8d9c3f7b9eca84ab779ff6768a2db3c6f7bdcf3433cdef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:38Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.118999 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.119042 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.119053 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.119064 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.119073 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:38Z","lastTransitionTime":"2026-01-21T11:11:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.125883 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"26b1108a-4b98-4e39-a3cd-e0d055089fd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc319bbc06568cc811ded2b418b04b986a31f321fed492e8be50104af79bb30b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cee9cc9eec3fc815f1621a22a37805691288dbb60dd3ff64194a84be62abcddc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc4c96e33e069f7d08d33caf92cd0987668506f931fd134b246d5ce5e934f95c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ddbb0e5ae02e9d5d82a80ead7efc91f1b8e557a9fb6f39940e5530fae8de2d38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://64a9065254f889b1c43d63759283335bf78fc4fa1531a4baae472c06c075dae3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e94ef8a8f47dc75ecfbab2e3fb8e7097c1b73eed32d7ca947f11ea198cea880\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e688b12ddb605e11511f11a6b41f84cd31b6065797f6b7332b1255595bd27cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e688b12ddb605e11511f11a6b41f84cd31b6065797f6b7332b1255595bd27cb\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-21T11:11:23Z\\\",\\\"message\\\":\\\" crc\\\\nF0121 11:11:23.774847 6243 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:23Z is after 2025-08-24T17:21:41Z]\\\\nI0121 11:11:23.774842 6243 services_controller.go:451] Built service openshift-kube-storage-version-migrator-operator/metrics cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-kube-storage-version-migrator-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-storage-version-migrator-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeO\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:23Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-htkvk_openshift-ovn-kubernetes(26b1108a-4b98-4e39-a3cd-e0d055089fd8)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948e0497eaffa7b0d84b4085c9ed8d987eb2c6527c9a82c10e576ed9c9c1e21b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-htkvk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:38Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.133469 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-np2v8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4698abe4-0f33-4349-a2a4-614de80ae21b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://191328471bb6f2b34cde2eb927da697619826bf0b6174ecb583c8c25d27a6654\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n9qbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://050bd5bdded2162d5e9dac81713f4ed63cb575d1776e828b2f98e50d89f11b04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n9qbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-np2v8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:38Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.142199 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d31ab61-f94c-4f49-85d0-c9e91b83069b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a66b5c33884af00408dc35cf02f08ebd8570bd695f9f1fe558bf54fd06930431\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7514786602746dabfa690da48d48e233fe68d6dd6674aa5ecd489c0b3185f08\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61b14b6846f94cdeba8f1f61b3754e7663041f92c70356c4ee6cbb988a2e680f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a7023cfe4c098bcdd9c0477a8187f4ed7a0d2abdbe1e143436863948f441b4f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af11ad95bcf0c6bbd5fa4ab99bb592ce1c87f40d29c5ade7832414315820a569\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW0121 11:11:14.604290 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0121 11:11:14.604572 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0121 11:11:14.605875 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3284452394/tls.crt::/tmp/serving-cert-3284452394/tls.key\\\\\\\"\\\\nI0121 11:11:14.923427 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0121 11:11:14.928141 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0121 11:11:14.928162 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0121 11:11:14.928179 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0121 11:11:14.928184 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0121 11:11:14.933154 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0121 11:11:14.933230 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 11:11:14.933259 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 11:11:14.933282 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0121 11:11:14.933299 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0121 11:11:14.933315 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0121 11:11:14.933331 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0121 11:11:14.933177 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0121 11:11:14.934666 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48070c09fad511108ff22e44211cf26911509b8406d66948985cbfc3009d8233\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:38Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.150044 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9bb99c0a-683b-44da-9c0e-e8d284e87d53\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05e076febd697ed802ce5c7ee92a8c2c636a1a2c6609ad7eb01f17dace247c4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43529d1b9e1e5e22b0d56b023946119da8aacbae54b68c26ff0a1a82dc890c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ea232ad87393685aa323ce6393ff3869f8f8d423c2406a501c69c355535adfa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc6c7eff9e17b8a8ecc88bbfe703b5a01dbf1e02d342cede4f6a263a6526af43\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:38Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.157656 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:38Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.161831 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.161854 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.161863 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.161874 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.161882 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:38Z","lastTransitionTime":"2026-01-21T11:11:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.165666 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lv4rn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1267539-9adf-466f-8027-8ede009d19a0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eed1b884aa53e1ee318169a0cc2c9ad10130aa363ae4ac55e4bf065237756f53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbmfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lv4rn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:38Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:38 crc kubenswrapper[4824]: E0121 11:11:38.169261 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"695c676c-720c-4921-9e71-56b96f2fb46b\\\",\\\"systemUUID\\\":\\\"d00624f5-786f-41dc-bdc1-a1875f167d65\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:38Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.171486 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.171506 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.171515 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.171524 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.171532 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:38Z","lastTransitionTime":"2026-01-21T11:11:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.172029 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jr64s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9768c60d-5bee-4286-bced-076e584b04e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91158f14d2e98ac9f875a5e7101df3107ef28aa5fc5d62218f460af98ff920fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flrnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:17Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jr64s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:38Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:38 crc kubenswrapper[4824]: E0121 11:11:38.178596 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"695c676c-720c-4921-9e71-56b96f2fb46b\\\",\\\"systemUUID\\\":\\\"d00624f5-786f-41dc-bdc1-a1875f167d65\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:38Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.181121 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.181149 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.181158 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.181167 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.181174 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:38Z","lastTransitionTime":"2026-01-21T11:11:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.184104 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06f56d86-ec6f-4b57-b314-f508c0dd4d9f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d25c94f1ae0bbabb733f0e1ea1d1ba39562a92af73047a39aacc03013e239681\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca06234486b59cf990c69d9501e23f935328e2ba21013cd71ed64831ea749bcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b63e45d66bcd11ff3479cc3ec6b450017da8d21d36e967b8d496ae0f24b0ee9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52758bcd31f2ba64edf03adfd4d1be6428015abff5b64c7c029fa787085f9c29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2736e73342d2adb175b0721dd4f40cc42d26fca06d5eb3334927518531c3b107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b35ad7dd0570c191eb0080739304f3cd43f17b3d1dddf895549441ebe50deaa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b35ad7dd0570c191eb0080739304f3cd43f17b3d1dddf895549441ebe50deaa6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://80cd5ea5e6fc19a88c1aab5c61f827f103ab798a5c84da8dc4d45cea86d5aacb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80cd5ea5e6fc19a88c1aab5c61f827f103ab798a5c84da8dc4d45cea86d5aacb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://234d8b914e6a88480021647ae65d74e9b21c55b6788312a094d00c8a42677d48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://234d8b914e6a88480021647ae65d74e9b21c55b6788312a094d00c8a42677d48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:38Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:38 crc kubenswrapper[4824]: E0121 11:11:38.189086 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"695c676c-720c-4921-9e71-56b96f2fb46b\\\",\\\"systemUUID\\\":\\\"d00624f5-786f-41dc-bdc1-a1875f167d65\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:38Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.191222 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:38Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.191523 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.191544 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.191553 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.191562 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.191568 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:38Z","lastTransitionTime":"2026-01-21T11:11:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:38 crc kubenswrapper[4824]: E0121 11:11:38.199295 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"695c676c-720c-4921-9e71-56b96f2fb46b\\\",\\\"systemUUID\\\":\\\"d00624f5-786f-41dc-bdc1-a1875f167d65\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:38Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.201630 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.201655 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.201663 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.201672 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.201679 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:38Z","lastTransitionTime":"2026-01-21T11:11:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:38 crc kubenswrapper[4824]: E0121 11:11:38.208951 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"695c676c-720c-4921-9e71-56b96f2fb46b\\\",\\\"systemUUID\\\":\\\"d00624f5-786f-41dc-bdc1-a1875f167d65\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:38Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:38 crc kubenswrapper[4824]: E0121 11:11:38.209093 4824 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.221459 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.221495 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.221507 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.221523 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.221533 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:38Z","lastTransitionTime":"2026-01-21T11:11:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.323627 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.323678 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.323688 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.323700 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.323711 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:38Z","lastTransitionTime":"2026-01-21T11:11:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.426348 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.426409 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.426419 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.426431 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.426440 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:38Z","lastTransitionTime":"2026-01-21T11:11:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.528380 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.528415 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.528423 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.528443 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.528451 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:38Z","lastTransitionTime":"2026-01-21T11:11:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.630028 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.630053 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.630063 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.630073 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.630082 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:38Z","lastTransitionTime":"2026-01-21T11:11:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.732122 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.732152 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.732162 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.732175 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.732194 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:38Z","lastTransitionTime":"2026-01-21T11:11:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.834507 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.834542 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.834551 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.834582 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.834591 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:38Z","lastTransitionTime":"2026-01-21T11:11:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.936613 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.936639 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.936647 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.936660 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:38 crc kubenswrapper[4824]: I0121 11:11:38.936668 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:38Z","lastTransitionTime":"2026-01-21T11:11:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:39 crc kubenswrapper[4824]: I0121 11:11:39.001309 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-23 15:44:27.184701429 +0000 UTC Jan 21 11:11:39 crc kubenswrapper[4824]: I0121 11:11:39.039283 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:39 crc kubenswrapper[4824]: I0121 11:11:39.039333 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:39 crc kubenswrapper[4824]: I0121 11:11:39.039343 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:39 crc kubenswrapper[4824]: I0121 11:11:39.039355 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:39 crc kubenswrapper[4824]: I0121 11:11:39.039363 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:39Z","lastTransitionTime":"2026-01-21T11:11:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:39 crc kubenswrapper[4824]: I0121 11:11:39.048580 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 11:11:39 crc kubenswrapper[4824]: I0121 11:11:39.048620 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 11:11:39 crc kubenswrapper[4824]: E0121 11:11:39.048694 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 21 11:11:39 crc kubenswrapper[4824]: E0121 11:11:39.048778 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 21 11:11:39 crc kubenswrapper[4824]: I0121 11:11:39.141925 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:39 crc kubenswrapper[4824]: I0121 11:11:39.141979 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:39 crc kubenswrapper[4824]: I0121 11:11:39.141989 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:39 crc kubenswrapper[4824]: I0121 11:11:39.142002 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:39 crc kubenswrapper[4824]: I0121 11:11:39.142011 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:39Z","lastTransitionTime":"2026-01-21T11:11:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:39 crc kubenswrapper[4824]: I0121 11:11:39.244179 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:39 crc kubenswrapper[4824]: I0121 11:11:39.244227 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:39 crc kubenswrapper[4824]: I0121 11:11:39.244238 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:39 crc kubenswrapper[4824]: I0121 11:11:39.244250 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:39 crc kubenswrapper[4824]: I0121 11:11:39.244259 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:39Z","lastTransitionTime":"2026-01-21T11:11:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:39 crc kubenswrapper[4824]: I0121 11:11:39.346390 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:39 crc kubenswrapper[4824]: I0121 11:11:39.346429 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:39 crc kubenswrapper[4824]: I0121 11:11:39.346438 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:39 crc kubenswrapper[4824]: I0121 11:11:39.346453 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:39 crc kubenswrapper[4824]: I0121 11:11:39.346463 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:39Z","lastTransitionTime":"2026-01-21T11:11:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:39 crc kubenswrapper[4824]: I0121 11:11:39.448798 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:39 crc kubenswrapper[4824]: I0121 11:11:39.448836 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:39 crc kubenswrapper[4824]: I0121 11:11:39.448848 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:39 crc kubenswrapper[4824]: I0121 11:11:39.448862 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:39 crc kubenswrapper[4824]: I0121 11:11:39.448871 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:39Z","lastTransitionTime":"2026-01-21T11:11:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:39 crc kubenswrapper[4824]: I0121 11:11:39.551081 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:39 crc kubenswrapper[4824]: I0121 11:11:39.551115 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:39 crc kubenswrapper[4824]: I0121 11:11:39.551125 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:39 crc kubenswrapper[4824]: I0121 11:11:39.551138 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:39 crc kubenswrapper[4824]: I0121 11:11:39.551145 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:39Z","lastTransitionTime":"2026-01-21T11:11:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:39 crc kubenswrapper[4824]: I0121 11:11:39.653323 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:39 crc kubenswrapper[4824]: I0121 11:11:39.653378 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:39 crc kubenswrapper[4824]: I0121 11:11:39.653387 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:39 crc kubenswrapper[4824]: I0121 11:11:39.653401 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:39 crc kubenswrapper[4824]: I0121 11:11:39.653409 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:39Z","lastTransitionTime":"2026-01-21T11:11:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:39 crc kubenswrapper[4824]: I0121 11:11:39.755449 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:39 crc kubenswrapper[4824]: I0121 11:11:39.755486 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:39 crc kubenswrapper[4824]: I0121 11:11:39.755494 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:39 crc kubenswrapper[4824]: I0121 11:11:39.755507 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:39 crc kubenswrapper[4824]: I0121 11:11:39.755517 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:39Z","lastTransitionTime":"2026-01-21T11:11:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:39 crc kubenswrapper[4824]: I0121 11:11:39.857533 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:39 crc kubenswrapper[4824]: I0121 11:11:39.857578 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:39 crc kubenswrapper[4824]: I0121 11:11:39.857587 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:39 crc kubenswrapper[4824]: I0121 11:11:39.857606 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:39 crc kubenswrapper[4824]: I0121 11:11:39.857614 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:39Z","lastTransitionTime":"2026-01-21T11:11:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:39 crc kubenswrapper[4824]: I0121 11:11:39.959932 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:39 crc kubenswrapper[4824]: I0121 11:11:39.960000 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:39 crc kubenswrapper[4824]: I0121 11:11:39.960016 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:39 crc kubenswrapper[4824]: I0121 11:11:39.960029 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:39 crc kubenswrapper[4824]: I0121 11:11:39.960036 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:39Z","lastTransitionTime":"2026-01-21T11:11:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:40 crc kubenswrapper[4824]: I0121 11:11:40.001475 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-07 20:49:52.696614283 +0000 UTC Jan 21 11:11:40 crc kubenswrapper[4824]: I0121 11:11:40.049047 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9rmnw" Jan 21 11:11:40 crc kubenswrapper[4824]: I0121 11:11:40.049088 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 11:11:40 crc kubenswrapper[4824]: E0121 11:11:40.049146 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9rmnw" podUID="8f9bd12b-22ee-44ed-a8b9-c100d043f691" Jan 21 11:11:40 crc kubenswrapper[4824]: E0121 11:11:40.049280 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 21 11:11:40 crc kubenswrapper[4824]: I0121 11:11:40.050437 4824 scope.go:117] "RemoveContainer" containerID="8e688b12ddb605e11511f11a6b41f84cd31b6065797f6b7332b1255595bd27cb" Jan 21 11:11:40 crc kubenswrapper[4824]: I0121 11:11:40.061435 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:40 crc kubenswrapper[4824]: I0121 11:11:40.061478 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:40 crc kubenswrapper[4824]: I0121 11:11:40.061489 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:40 crc kubenswrapper[4824]: I0121 11:11:40.061503 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:40 crc kubenswrapper[4824]: I0121 11:11:40.061515 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:40Z","lastTransitionTime":"2026-01-21T11:11:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:40 crc kubenswrapper[4824]: I0121 11:11:40.164472 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:40 crc kubenswrapper[4824]: I0121 11:11:40.164499 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:40 crc kubenswrapper[4824]: I0121 11:11:40.164508 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:40 crc kubenswrapper[4824]: I0121 11:11:40.164523 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:40 crc kubenswrapper[4824]: I0121 11:11:40.164532 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:40Z","lastTransitionTime":"2026-01-21T11:11:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:40 crc kubenswrapper[4824]: I0121 11:11:40.235663 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-htkvk_26b1108a-4b98-4e39-a3cd-e0d055089fd8/ovnkube-controller/1.log" Jan 21 11:11:40 crc kubenswrapper[4824]: I0121 11:11:40.238439 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" event={"ID":"26b1108a-4b98-4e39-a3cd-e0d055089fd8","Type":"ContainerStarted","Data":"2c655fe698872c3a7804fed5e2e75f300110a816eae4412891f3433762ed0161"} Jan 21 11:11:40 crc kubenswrapper[4824]: I0121 11:11:40.238815 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" Jan 21 11:11:40 crc kubenswrapper[4824]: I0121 11:11:40.247993 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-9rmnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f9bd12b-22ee-44ed-a8b9-c100d043f691\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2fgn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2fgn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:28Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-9rmnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:40Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:40 crc kubenswrapper[4824]: I0121 11:11:40.258367 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86d3005cd61c0973659e1c1c0d3ef1ddd6fcf95bede5465f4f2099e64a74bb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:40Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:40 crc kubenswrapper[4824]: I0121 11:11:40.266344 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:40 crc kubenswrapper[4824]: I0121 11:11:40.266381 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:40 crc kubenswrapper[4824]: I0121 11:11:40.266390 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:40 crc kubenswrapper[4824]: I0121 11:11:40.266407 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:40 crc kubenswrapper[4824]: I0121 11:11:40.266419 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:40Z","lastTransitionTime":"2026-01-21T11:11:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:40 crc kubenswrapper[4824]: I0121 11:11:40.266641 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:40Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:40 crc kubenswrapper[4824]: I0121 11:11:40.277205 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x94nb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75d52224135b75133fc000611cd179240b9174f6492d3263341716c7e42ccb15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgr8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x94nb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:40Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:40 crc kubenswrapper[4824]: I0121 11:11:40.286468 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33f3d922-4ffe-409b-a49a-d88c85898260\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a9e57b39eed8300107e944b9fcf6a2ff1e75991f8f37387a3de615b14f9bc78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpwst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b1b37c6d2994021cfc30f85e92c40ae35677113b17aa066d52bc7a76ae5c153\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpwst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zqjg9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:40Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:40 crc kubenswrapper[4824]: I0121 11:11:40.301368 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gjtnh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a8d2cac-2d0f-4894-9bc0-2abd81699c47\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67c122a26811deaa9b70d19a1a65daee318ca3af2c626712482179d72fe67875\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2239186f1de46481cebfcafb1cab85cd320a205b5a0dfc6dd1b321d9baedeba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b2239186f1de46481cebfcafb1cab85cd320a205b5a0dfc6dd1b321d9baedeba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0db5ecd7759d974ae1b0d8065b3b8bd4d7957fae16eb5bab2d71c2fe69243816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0db5ecd7759d974ae1b0d8065b3b8bd4d7957fae16eb5bab2d71c2fe69243816\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba9b3278c208fc7aac630ca7a0daa3de0823f85d0799d6f1f278da80ae14c862\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba9b3278c208fc7aac630ca7a0daa3de0823f85d0799d6f1f278da80ae14c862\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79b8f6c016b43c3ac3ba7b22cca9a6b72c262cb17ddd6236c5d4970ecc09680c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79b8f6c016b43c3ac3ba7b22cca9a6b72c262cb17ddd6236c5d4970ecc09680c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38c4f218eb2ca7ea0c92bd0ae2ae9efeaa40746751608fbcf7f6c3f8d9e3c758\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38c4f218eb2ca7ea0c92bd0ae2ae9efeaa40746751608fbcf7f6c3f8d9e3c758\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d66ad9831d6029c8c350054525a00069e869d39c1b5638e06e9b772666bd8630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d66ad9831d6029c8c350054525a00069e869d39c1b5638e06e9b772666bd8630\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gjtnh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:40Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:40 crc kubenswrapper[4824]: I0121 11:11:40.311837 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d87422473838f5218c9fb646ab2515f2f42592754ffdf07b19f7a8e093fa48a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:40Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:40 crc kubenswrapper[4824]: I0121 11:11:40.320828 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://516c009d4720ab5b393c7482872f12fed22d4f4414bd1f3400411c0e8da88ce4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee18d293036971832c8d9c3f7b9eca84ab779ff6768a2db3c6f7bdcf3433cdef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:40Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:40 crc kubenswrapper[4824]: I0121 11:11:40.335608 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"26b1108a-4b98-4e39-a3cd-e0d055089fd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc319bbc06568cc811ded2b418b04b986a31f321fed492e8be50104af79bb30b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cee9cc9eec3fc815f1621a22a37805691288dbb60dd3ff64194a84be62abcddc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc4c96e33e069f7d08d33caf92cd0987668506f931fd134b246d5ce5e934f95c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ddbb0e5ae02e9d5d82a80ead7efc91f1b8e557a9fb6f39940e5530fae8de2d38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://64a9065254f889b1c43d63759283335bf78fc4fa1531a4baae472c06c075dae3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e94ef8a8f47dc75ecfbab2e3fb8e7097c1b73eed32d7ca947f11ea198cea880\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c655fe698872c3a7804fed5e2e75f300110a816eae4412891f3433762ed0161\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e688b12ddb605e11511f11a6b41f84cd31b6065797f6b7332b1255595bd27cb\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-21T11:11:23Z\\\",\\\"message\\\":\\\" crc\\\\nF0121 11:11:23.774847 6243 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:23Z is after 2025-08-24T17:21:41Z]\\\\nI0121 11:11:23.774842 6243 services_controller.go:451] Built service openshift-kube-storage-version-migrator-operator/metrics cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-kube-storage-version-migrator-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-storage-version-migrator-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeO\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:23Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948e0497eaffa7b0d84b4085c9ed8d987eb2c6527c9a82c10e576ed9c9c1e21b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-htkvk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:40Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:40 crc kubenswrapper[4824]: I0121 11:11:40.348845 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-np2v8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4698abe4-0f33-4349-a2a4-614de80ae21b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://191328471bb6f2b34cde2eb927da697619826bf0b6174ecb583c8c25d27a6654\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n9qbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://050bd5bdded2162d5e9dac81713f4ed63cb575d1776e828b2f98e50d89f11b04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n9qbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-np2v8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:40Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:40 crc kubenswrapper[4824]: I0121 11:11:40.358177 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lv4rn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1267539-9adf-466f-8027-8ede009d19a0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eed1b884aa53e1ee318169a0cc2c9ad10130aa363ae4ac55e4bf065237756f53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbmfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lv4rn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:40Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:40 crc kubenswrapper[4824]: I0121 11:11:40.369392 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:40 crc kubenswrapper[4824]: I0121 11:11:40.369426 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:40 crc kubenswrapper[4824]: I0121 11:11:40.369436 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:40 crc kubenswrapper[4824]: I0121 11:11:40.369452 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:40 crc kubenswrapper[4824]: I0121 11:11:40.369462 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:40Z","lastTransitionTime":"2026-01-21T11:11:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:40 crc kubenswrapper[4824]: I0121 11:11:40.371399 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jr64s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9768c60d-5bee-4286-bced-076e584b04e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91158f14d2e98ac9f875a5e7101df3107ef28aa5fc5d62218f460af98ff920fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flrnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:17Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jr64s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:40Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:40 crc kubenswrapper[4824]: I0121 11:11:40.383711 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d31ab61-f94c-4f49-85d0-c9e91b83069b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a66b5c33884af00408dc35cf02f08ebd8570bd695f9f1fe558bf54fd06930431\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7514786602746dabfa690da48d48e233fe68d6dd6674aa5ecd489c0b3185f08\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61b14b6846f94cdeba8f1f61b3754e7663041f92c70356c4ee6cbb988a2e680f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a7023cfe4c098bcdd9c0477a8187f4ed7a0d2abdbe1e143436863948f441b4f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af11ad95bcf0c6bbd5fa4ab99bb592ce1c87f40d29c5ade7832414315820a569\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW0121 11:11:14.604290 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0121 11:11:14.604572 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0121 11:11:14.605875 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3284452394/tls.crt::/tmp/serving-cert-3284452394/tls.key\\\\\\\"\\\\nI0121 11:11:14.923427 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0121 11:11:14.928141 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0121 11:11:14.928162 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0121 11:11:14.928179 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0121 11:11:14.928184 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0121 11:11:14.933154 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0121 11:11:14.933230 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 11:11:14.933259 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 11:11:14.933282 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0121 11:11:14.933299 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0121 11:11:14.933315 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0121 11:11:14.933331 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0121 11:11:14.933177 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0121 11:11:14.934666 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48070c09fad511108ff22e44211cf26911509b8406d66948985cbfc3009d8233\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:40Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:40 crc kubenswrapper[4824]: I0121 11:11:40.393342 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9bb99c0a-683b-44da-9c0e-e8d284e87d53\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05e076febd697ed802ce5c7ee92a8c2c636a1a2c6609ad7eb01f17dace247c4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43529d1b9e1e5e22b0d56b023946119da8aacbae54b68c26ff0a1a82dc890c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ea232ad87393685aa323ce6393ff3869f8f8d423c2406a501c69c355535adfa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc6c7eff9e17b8a8ecc88bbfe703b5a01dbf1e02d342cede4f6a263a6526af43\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:40Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:40 crc kubenswrapper[4824]: I0121 11:11:40.402048 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:40Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:40 crc kubenswrapper[4824]: I0121 11:11:40.415367 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06f56d86-ec6f-4b57-b314-f508c0dd4d9f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d25c94f1ae0bbabb733f0e1ea1d1ba39562a92af73047a39aacc03013e239681\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca06234486b59cf990c69d9501e23f935328e2ba21013cd71ed64831ea749bcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b63e45d66bcd11ff3479cc3ec6b450017da8d21d36e967b8d496ae0f24b0ee9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52758bcd31f2ba64edf03adfd4d1be6428015abff5b64c7c029fa787085f9c29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2736e73342d2adb175b0721dd4f40cc42d26fca06d5eb3334927518531c3b107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b35ad7dd0570c191eb0080739304f3cd43f17b3d1dddf895549441ebe50deaa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b35ad7dd0570c191eb0080739304f3cd43f17b3d1dddf895549441ebe50deaa6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://80cd5ea5e6fc19a88c1aab5c61f827f103ab798a5c84da8dc4d45cea86d5aacb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80cd5ea5e6fc19a88c1aab5c61f827f103ab798a5c84da8dc4d45cea86d5aacb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://234d8b914e6a88480021647ae65d74e9b21c55b6788312a094d00c8a42677d48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://234d8b914e6a88480021647ae65d74e9b21c55b6788312a094d00c8a42677d48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:40Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:40 crc kubenswrapper[4824]: I0121 11:11:40.423979 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:40Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:40 crc kubenswrapper[4824]: I0121 11:11:40.471598 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:40 crc kubenswrapper[4824]: I0121 11:11:40.471642 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:40 crc kubenswrapper[4824]: I0121 11:11:40.471651 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:40 crc kubenswrapper[4824]: I0121 11:11:40.471669 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:40 crc kubenswrapper[4824]: I0121 11:11:40.471679 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:40Z","lastTransitionTime":"2026-01-21T11:11:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:40 crc kubenswrapper[4824]: I0121 11:11:40.573704 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:40 crc kubenswrapper[4824]: I0121 11:11:40.573761 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:40 crc kubenswrapper[4824]: I0121 11:11:40.573771 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:40 crc kubenswrapper[4824]: I0121 11:11:40.573785 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:40 crc kubenswrapper[4824]: I0121 11:11:40.573794 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:40Z","lastTransitionTime":"2026-01-21T11:11:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:40 crc kubenswrapper[4824]: I0121 11:11:40.675864 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:40 crc kubenswrapper[4824]: I0121 11:11:40.675904 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:40 crc kubenswrapper[4824]: I0121 11:11:40.675914 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:40 crc kubenswrapper[4824]: I0121 11:11:40.675927 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:40 crc kubenswrapper[4824]: I0121 11:11:40.675937 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:40Z","lastTransitionTime":"2026-01-21T11:11:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:40 crc kubenswrapper[4824]: I0121 11:11:40.778169 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:40 crc kubenswrapper[4824]: I0121 11:11:40.778219 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:40 crc kubenswrapper[4824]: I0121 11:11:40.778228 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:40 crc kubenswrapper[4824]: I0121 11:11:40.778245 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:40 crc kubenswrapper[4824]: I0121 11:11:40.778253 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:40Z","lastTransitionTime":"2026-01-21T11:11:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:40 crc kubenswrapper[4824]: I0121 11:11:40.880278 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:40 crc kubenswrapper[4824]: I0121 11:11:40.880318 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:40 crc kubenswrapper[4824]: I0121 11:11:40.880326 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:40 crc kubenswrapper[4824]: I0121 11:11:40.880341 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:40 crc kubenswrapper[4824]: I0121 11:11:40.880348 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:40Z","lastTransitionTime":"2026-01-21T11:11:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:40 crc kubenswrapper[4824]: I0121 11:11:40.982138 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:40 crc kubenswrapper[4824]: I0121 11:11:40.982180 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:40 crc kubenswrapper[4824]: I0121 11:11:40.982189 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:40 crc kubenswrapper[4824]: I0121 11:11:40.982416 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:40 crc kubenswrapper[4824]: I0121 11:11:40.982425 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:40Z","lastTransitionTime":"2026-01-21T11:11:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:41 crc kubenswrapper[4824]: I0121 11:11:41.002547 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-06 15:17:40.095088858 +0000 UTC Jan 21 11:11:41 crc kubenswrapper[4824]: I0121 11:11:41.049240 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 11:11:41 crc kubenswrapper[4824]: I0121 11:11:41.049262 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 11:11:41 crc kubenswrapper[4824]: E0121 11:11:41.049403 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 21 11:11:41 crc kubenswrapper[4824]: E0121 11:11:41.049503 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 21 11:11:41 crc kubenswrapper[4824]: I0121 11:11:41.084565 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:41 crc kubenswrapper[4824]: I0121 11:11:41.084596 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:41 crc kubenswrapper[4824]: I0121 11:11:41.084606 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:41 crc kubenswrapper[4824]: I0121 11:11:41.084618 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:41 crc kubenswrapper[4824]: I0121 11:11:41.084627 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:41Z","lastTransitionTime":"2026-01-21T11:11:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:41 crc kubenswrapper[4824]: I0121 11:11:41.186612 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:41 crc kubenswrapper[4824]: I0121 11:11:41.186642 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:41 crc kubenswrapper[4824]: I0121 11:11:41.186651 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:41 crc kubenswrapper[4824]: I0121 11:11:41.186665 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:41 crc kubenswrapper[4824]: I0121 11:11:41.186675 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:41Z","lastTransitionTime":"2026-01-21T11:11:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:41 crc kubenswrapper[4824]: I0121 11:11:41.242410 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-htkvk_26b1108a-4b98-4e39-a3cd-e0d055089fd8/ovnkube-controller/2.log" Jan 21 11:11:41 crc kubenswrapper[4824]: I0121 11:11:41.242841 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-htkvk_26b1108a-4b98-4e39-a3cd-e0d055089fd8/ovnkube-controller/1.log" Jan 21 11:11:41 crc kubenswrapper[4824]: I0121 11:11:41.244940 4824 generic.go:334] "Generic (PLEG): container finished" podID="26b1108a-4b98-4e39-a3cd-e0d055089fd8" containerID="2c655fe698872c3a7804fed5e2e75f300110a816eae4412891f3433762ed0161" exitCode=1 Jan 21 11:11:41 crc kubenswrapper[4824]: I0121 11:11:41.244983 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" event={"ID":"26b1108a-4b98-4e39-a3cd-e0d055089fd8","Type":"ContainerDied","Data":"2c655fe698872c3a7804fed5e2e75f300110a816eae4412891f3433762ed0161"} Jan 21 11:11:41 crc kubenswrapper[4824]: I0121 11:11:41.245029 4824 scope.go:117] "RemoveContainer" containerID="8e688b12ddb605e11511f11a6b41f84cd31b6065797f6b7332b1255595bd27cb" Jan 21 11:11:41 crc kubenswrapper[4824]: I0121 11:11:41.245822 4824 scope.go:117] "RemoveContainer" containerID="2c655fe698872c3a7804fed5e2e75f300110a816eae4412891f3433762ed0161" Jan 21 11:11:41 crc kubenswrapper[4824]: E0121 11:11:41.246042 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-htkvk_openshift-ovn-kubernetes(26b1108a-4b98-4e39-a3cd-e0d055089fd8)\"" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" podUID="26b1108a-4b98-4e39-a3cd-e0d055089fd8" Jan 21 11:11:41 crc kubenswrapper[4824]: I0121 11:11:41.259163 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33f3d922-4ffe-409b-a49a-d88c85898260\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a9e57b39eed8300107e944b9fcf6a2ff1e75991f8f37387a3de615b14f9bc78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpwst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b1b37c6d2994021cfc30f85e92c40ae35677113b17aa066d52bc7a76ae5c153\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpwst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zqjg9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:41Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:41 crc kubenswrapper[4824]: I0121 11:11:41.269479 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gjtnh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a8d2cac-2d0f-4894-9bc0-2abd81699c47\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67c122a26811deaa9b70d19a1a65daee318ca3af2c626712482179d72fe67875\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2239186f1de46481cebfcafb1cab85cd320a205b5a0dfc6dd1b321d9baedeba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b2239186f1de46481cebfcafb1cab85cd320a205b5a0dfc6dd1b321d9baedeba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0db5ecd7759d974ae1b0d8065b3b8bd4d7957fae16eb5bab2d71c2fe69243816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0db5ecd7759d974ae1b0d8065b3b8bd4d7957fae16eb5bab2d71c2fe69243816\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba9b3278c208fc7aac630ca7a0daa3de0823f85d0799d6f1f278da80ae14c862\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba9b3278c208fc7aac630ca7a0daa3de0823f85d0799d6f1f278da80ae14c862\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79b8f6c016b43c3ac3ba7b22cca9a6b72c262cb17ddd6236c5d4970ecc09680c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79b8f6c016b43c3ac3ba7b22cca9a6b72c262cb17ddd6236c5d4970ecc09680c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38c4f218eb2ca7ea0c92bd0ae2ae9efeaa40746751608fbcf7f6c3f8d9e3c758\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38c4f218eb2ca7ea0c92bd0ae2ae9efeaa40746751608fbcf7f6c3f8d9e3c758\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d66ad9831d6029c8c350054525a00069e869d39c1b5638e06e9b772666bd8630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d66ad9831d6029c8c350054525a00069e869d39c1b5638e06e9b772666bd8630\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gjtnh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:41Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:41 crc kubenswrapper[4824]: I0121 11:11:41.277303 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-9rmnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f9bd12b-22ee-44ed-a8b9-c100d043f691\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2fgn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2fgn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:28Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-9rmnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:41Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:41 crc kubenswrapper[4824]: I0121 11:11:41.286737 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86d3005cd61c0973659e1c1c0d3ef1ddd6fcf95bede5465f4f2099e64a74bb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:41Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:41 crc kubenswrapper[4824]: I0121 11:11:41.287930 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:41 crc kubenswrapper[4824]: I0121 11:11:41.287975 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:41 crc kubenswrapper[4824]: I0121 11:11:41.287984 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:41 crc kubenswrapper[4824]: I0121 11:11:41.287997 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:41 crc kubenswrapper[4824]: I0121 11:11:41.288005 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:41Z","lastTransitionTime":"2026-01-21T11:11:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:41 crc kubenswrapper[4824]: I0121 11:11:41.294484 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:41Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:41 crc kubenswrapper[4824]: I0121 11:11:41.302591 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x94nb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75d52224135b75133fc000611cd179240b9174f6492d3263341716c7e42ccb15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgr8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x94nb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:41Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:41 crc kubenswrapper[4824]: I0121 11:11:41.310238 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-np2v8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4698abe4-0f33-4349-a2a4-614de80ae21b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://191328471bb6f2b34cde2eb927da697619826bf0b6174ecb583c8c25d27a6654\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n9qbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://050bd5bdded2162d5e9dac81713f4ed63cb575d1776e828b2f98e50d89f11b04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n9qbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-np2v8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:41Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:41 crc kubenswrapper[4824]: I0121 11:11:41.318335 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d87422473838f5218c9fb646ab2515f2f42592754ffdf07b19f7a8e093fa48a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:41Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:41 crc kubenswrapper[4824]: I0121 11:11:41.326633 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://516c009d4720ab5b393c7482872f12fed22d4f4414bd1f3400411c0e8da88ce4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee18d293036971832c8d9c3f7b9eca84ab779ff6768a2db3c6f7bdcf3433cdef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:41Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:41 crc kubenswrapper[4824]: I0121 11:11:41.339374 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"26b1108a-4b98-4e39-a3cd-e0d055089fd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc319bbc06568cc811ded2b418b04b986a31f321fed492e8be50104af79bb30b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cee9cc9eec3fc815f1621a22a37805691288dbb60dd3ff64194a84be62abcddc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc4c96e33e069f7d08d33caf92cd0987668506f931fd134b246d5ce5e934f95c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ddbb0e5ae02e9d5d82a80ead7efc91f1b8e557a9fb6f39940e5530fae8de2d38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://64a9065254f889b1c43d63759283335bf78fc4fa1531a4baae472c06c075dae3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e94ef8a8f47dc75ecfbab2e3fb8e7097c1b73eed32d7ca947f11ea198cea880\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c655fe698872c3a7804fed5e2e75f300110a816eae4412891f3433762ed0161\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e688b12ddb605e11511f11a6b41f84cd31b6065797f6b7332b1255595bd27cb\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-21T11:11:23Z\\\",\\\"message\\\":\\\" crc\\\\nF0121 11:11:23.774847 6243 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:23Z is after 2025-08-24T17:21:41Z]\\\\nI0121 11:11:23.774842 6243 services_controller.go:451] Built service openshift-kube-storage-version-migrator-operator/metrics cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-kube-storage-version-migrator-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-storage-version-migrator-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeO\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:23Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2c655fe698872c3a7804fed5e2e75f300110a816eae4412891f3433762ed0161\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-21T11:11:40Z\\\",\\\"message\\\":\\\"ps://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:40Z is after 2025-08-24T17:21:41Z]\\\\nI0121 11:11:40.696180 6500 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf\\\\nI0121 11:11:40.696191 6500 obj_retry.go:303] Retry object setup: *v1.Pod openshift-kube-apiserver/kube-apiserver-crc\\\\nI0121 11:11:40.696193 6500 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf in node crc\\\\nI0121 11:11:40.696185 6500 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-apiserver-operator/metrics]} name:Service_openshift-apiserver-operator/metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.38:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {8b82f026-5975-4a1b-bb18-08d5d51147ec}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0121 11:11:40.696197 6500 obj_retry.go:365] Adding new object: *v1.Pod openshift-kube-apiserver/kube\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948e0497eaffa7b0d84b4085c9ed8d987eb2c6527c9a82c10e576ed9c9c1e21b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-htkvk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:41Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:41 crc kubenswrapper[4824]: I0121 11:11:41.350101 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:41Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:41 crc kubenswrapper[4824]: I0121 11:11:41.356700 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lv4rn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1267539-9adf-466f-8027-8ede009d19a0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eed1b884aa53e1ee318169a0cc2c9ad10130aa363ae4ac55e4bf065237756f53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbmfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lv4rn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:41Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:41 crc kubenswrapper[4824]: I0121 11:11:41.363360 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jr64s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9768c60d-5bee-4286-bced-076e584b04e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91158f14d2e98ac9f875a5e7101df3107ef28aa5fc5d62218f460af98ff920fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flrnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:17Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jr64s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:41Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:41 crc kubenswrapper[4824]: I0121 11:11:41.373379 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d31ab61-f94c-4f49-85d0-c9e91b83069b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a66b5c33884af00408dc35cf02f08ebd8570bd695f9f1fe558bf54fd06930431\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7514786602746dabfa690da48d48e233fe68d6dd6674aa5ecd489c0b3185f08\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61b14b6846f94cdeba8f1f61b3754e7663041f92c70356c4ee6cbb988a2e680f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a7023cfe4c098bcdd9c0477a8187f4ed7a0d2abdbe1e143436863948f441b4f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af11ad95bcf0c6bbd5fa4ab99bb592ce1c87f40d29c5ade7832414315820a569\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW0121 11:11:14.604290 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0121 11:11:14.604572 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0121 11:11:14.605875 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3284452394/tls.crt::/tmp/serving-cert-3284452394/tls.key\\\\\\\"\\\\nI0121 11:11:14.923427 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0121 11:11:14.928141 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0121 11:11:14.928162 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0121 11:11:14.928179 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0121 11:11:14.928184 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0121 11:11:14.933154 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0121 11:11:14.933230 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 11:11:14.933259 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 11:11:14.933282 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0121 11:11:14.933299 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0121 11:11:14.933315 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0121 11:11:14.933331 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0121 11:11:14.933177 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0121 11:11:14.934666 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48070c09fad511108ff22e44211cf26911509b8406d66948985cbfc3009d8233\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:41Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:41 crc kubenswrapper[4824]: I0121 11:11:41.381433 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9bb99c0a-683b-44da-9c0e-e8d284e87d53\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05e076febd697ed802ce5c7ee92a8c2c636a1a2c6609ad7eb01f17dace247c4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43529d1b9e1e5e22b0d56b023946119da8aacbae54b68c26ff0a1a82dc890c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ea232ad87393685aa323ce6393ff3869f8f8d423c2406a501c69c355535adfa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc6c7eff9e17b8a8ecc88bbfe703b5a01dbf1e02d342cede4f6a263a6526af43\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:41Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:41 crc kubenswrapper[4824]: I0121 11:11:41.391848 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:41 crc kubenswrapper[4824]: I0121 11:11:41.391883 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:41 crc kubenswrapper[4824]: I0121 11:11:41.391895 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:41 crc kubenswrapper[4824]: I0121 11:11:41.391913 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:41 crc kubenswrapper[4824]: I0121 11:11:41.391923 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:41Z","lastTransitionTime":"2026-01-21T11:11:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:41 crc kubenswrapper[4824]: I0121 11:11:41.394152 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06f56d86-ec6f-4b57-b314-f508c0dd4d9f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d25c94f1ae0bbabb733f0e1ea1d1ba39562a92af73047a39aacc03013e239681\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca06234486b59cf990c69d9501e23f935328e2ba21013cd71ed64831ea749bcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b63e45d66bcd11ff3479cc3ec6b450017da8d21d36e967b8d496ae0f24b0ee9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52758bcd31f2ba64edf03adfd4d1be6428015abff5b64c7c029fa787085f9c29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2736e73342d2adb175b0721dd4f40cc42d26fca06d5eb3334927518531c3b107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b35ad7dd0570c191eb0080739304f3cd43f17b3d1dddf895549441ebe50deaa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b35ad7dd0570c191eb0080739304f3cd43f17b3d1dddf895549441ebe50deaa6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://80cd5ea5e6fc19a88c1aab5c61f827f103ab798a5c84da8dc4d45cea86d5aacb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80cd5ea5e6fc19a88c1aab5c61f827f103ab798a5c84da8dc4d45cea86d5aacb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://234d8b914e6a88480021647ae65d74e9b21c55b6788312a094d00c8a42677d48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://234d8b914e6a88480021647ae65d74e9b21c55b6788312a094d00c8a42677d48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:41Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:41 crc kubenswrapper[4824]: I0121 11:11:41.402485 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:41Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:41 crc kubenswrapper[4824]: I0121 11:11:41.494662 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:41 crc kubenswrapper[4824]: I0121 11:11:41.494702 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:41 crc kubenswrapper[4824]: I0121 11:11:41.494715 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:41 crc kubenswrapper[4824]: I0121 11:11:41.494735 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:41 crc kubenswrapper[4824]: I0121 11:11:41.494746 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:41Z","lastTransitionTime":"2026-01-21T11:11:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:41 crc kubenswrapper[4824]: I0121 11:11:41.596594 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:41 crc kubenswrapper[4824]: I0121 11:11:41.596628 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:41 crc kubenswrapper[4824]: I0121 11:11:41.596638 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:41 crc kubenswrapper[4824]: I0121 11:11:41.596656 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:41 crc kubenswrapper[4824]: I0121 11:11:41.596668 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:41Z","lastTransitionTime":"2026-01-21T11:11:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:41 crc kubenswrapper[4824]: I0121 11:11:41.698495 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:41 crc kubenswrapper[4824]: I0121 11:11:41.698526 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:41 crc kubenswrapper[4824]: I0121 11:11:41.698534 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:41 crc kubenswrapper[4824]: I0121 11:11:41.698548 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:41 crc kubenswrapper[4824]: I0121 11:11:41.698558 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:41Z","lastTransitionTime":"2026-01-21T11:11:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:41 crc kubenswrapper[4824]: I0121 11:11:41.800953 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:41 crc kubenswrapper[4824]: I0121 11:11:41.801018 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:41 crc kubenswrapper[4824]: I0121 11:11:41.801027 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:41 crc kubenswrapper[4824]: I0121 11:11:41.801042 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:41 crc kubenswrapper[4824]: I0121 11:11:41.801051 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:41Z","lastTransitionTime":"2026-01-21T11:11:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:41 crc kubenswrapper[4824]: I0121 11:11:41.902935 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:41 crc kubenswrapper[4824]: I0121 11:11:41.902999 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:41 crc kubenswrapper[4824]: I0121 11:11:41.903010 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:41 crc kubenswrapper[4824]: I0121 11:11:41.903026 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:41 crc kubenswrapper[4824]: I0121 11:11:41.903037 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:41Z","lastTransitionTime":"2026-01-21T11:11:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:42 crc kubenswrapper[4824]: I0121 11:11:42.003083 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-16 01:17:21.564019552 +0000 UTC Jan 21 11:11:42 crc kubenswrapper[4824]: I0121 11:11:42.004532 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:42 crc kubenswrapper[4824]: I0121 11:11:42.004560 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:42 crc kubenswrapper[4824]: I0121 11:11:42.004571 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:42 crc kubenswrapper[4824]: I0121 11:11:42.004583 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:42 crc kubenswrapper[4824]: I0121 11:11:42.004592 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:42Z","lastTransitionTime":"2026-01-21T11:11:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:42 crc kubenswrapper[4824]: I0121 11:11:42.048502 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9rmnw" Jan 21 11:11:42 crc kubenswrapper[4824]: E0121 11:11:42.048615 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9rmnw" podUID="8f9bd12b-22ee-44ed-a8b9-c100d043f691" Jan 21 11:11:42 crc kubenswrapper[4824]: I0121 11:11:42.048686 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 11:11:42 crc kubenswrapper[4824]: E0121 11:11:42.048815 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 21 11:11:42 crc kubenswrapper[4824]: I0121 11:11:42.106430 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:42 crc kubenswrapper[4824]: I0121 11:11:42.106454 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:42 crc kubenswrapper[4824]: I0121 11:11:42.106462 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:42 crc kubenswrapper[4824]: I0121 11:11:42.106473 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:42 crc kubenswrapper[4824]: I0121 11:11:42.106481 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:42Z","lastTransitionTime":"2026-01-21T11:11:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:42 crc kubenswrapper[4824]: I0121 11:11:42.207822 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:42 crc kubenswrapper[4824]: I0121 11:11:42.207849 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:42 crc kubenswrapper[4824]: I0121 11:11:42.207858 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:42 crc kubenswrapper[4824]: I0121 11:11:42.207868 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:42 crc kubenswrapper[4824]: I0121 11:11:42.207875 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:42Z","lastTransitionTime":"2026-01-21T11:11:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:42 crc kubenswrapper[4824]: I0121 11:11:42.248329 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-htkvk_26b1108a-4b98-4e39-a3cd-e0d055089fd8/ovnkube-controller/2.log" Jan 21 11:11:42 crc kubenswrapper[4824]: I0121 11:11:42.250494 4824 scope.go:117] "RemoveContainer" containerID="2c655fe698872c3a7804fed5e2e75f300110a816eae4412891f3433762ed0161" Jan 21 11:11:42 crc kubenswrapper[4824]: E0121 11:11:42.250618 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-htkvk_openshift-ovn-kubernetes(26b1108a-4b98-4e39-a3cd-e0d055089fd8)\"" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" podUID="26b1108a-4b98-4e39-a3cd-e0d055089fd8" Jan 21 11:11:42 crc kubenswrapper[4824]: I0121 11:11:42.259822 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d87422473838f5218c9fb646ab2515f2f42592754ffdf07b19f7a8e093fa48a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:42Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:42 crc kubenswrapper[4824]: I0121 11:11:42.268021 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://516c009d4720ab5b393c7482872f12fed22d4f4414bd1f3400411c0e8da88ce4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee18d293036971832c8d9c3f7b9eca84ab779ff6768a2db3c6f7bdcf3433cdef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:42Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:42 crc kubenswrapper[4824]: I0121 11:11:42.279986 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"26b1108a-4b98-4e39-a3cd-e0d055089fd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc319bbc06568cc811ded2b418b04b986a31f321fed492e8be50104af79bb30b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cee9cc9eec3fc815f1621a22a37805691288dbb60dd3ff64194a84be62abcddc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc4c96e33e069f7d08d33caf92cd0987668506f931fd134b246d5ce5e934f95c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ddbb0e5ae02e9d5d82a80ead7efc91f1b8e557a9fb6f39940e5530fae8de2d38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://64a9065254f889b1c43d63759283335bf78fc4fa1531a4baae472c06c075dae3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e94ef8a8f47dc75ecfbab2e3fb8e7097c1b73eed32d7ca947f11ea198cea880\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c655fe698872c3a7804fed5e2e75f300110a816eae4412891f3433762ed0161\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2c655fe698872c3a7804fed5e2e75f300110a816eae4412891f3433762ed0161\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-21T11:11:40Z\\\",\\\"message\\\":\\\"ps://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:40Z is after 2025-08-24T17:21:41Z]\\\\nI0121 11:11:40.696180 6500 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf\\\\nI0121 11:11:40.696191 6500 obj_retry.go:303] Retry object setup: *v1.Pod openshift-kube-apiserver/kube-apiserver-crc\\\\nI0121 11:11:40.696193 6500 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf in node crc\\\\nI0121 11:11:40.696185 6500 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-apiserver-operator/metrics]} name:Service_openshift-apiserver-operator/metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.38:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {8b82f026-5975-4a1b-bb18-08d5d51147ec}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0121 11:11:40.696197 6500 obj_retry.go:365] Adding new object: *v1.Pod openshift-kube-apiserver/kube\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:40Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-htkvk_openshift-ovn-kubernetes(26b1108a-4b98-4e39-a3cd-e0d055089fd8)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948e0497eaffa7b0d84b4085c9ed8d987eb2c6527c9a82c10e576ed9c9c1e21b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-htkvk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:42Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:42 crc kubenswrapper[4824]: I0121 11:11:42.287570 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-np2v8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4698abe4-0f33-4349-a2a4-614de80ae21b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://191328471bb6f2b34cde2eb927da697619826bf0b6174ecb583c8c25d27a6654\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n9qbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://050bd5bdded2162d5e9dac81713f4ed63cb575d1776e828b2f98e50d89f11b04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n9qbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-np2v8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:42Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:42 crc kubenswrapper[4824]: I0121 11:11:42.294551 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jr64s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9768c60d-5bee-4286-bced-076e584b04e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91158f14d2e98ac9f875a5e7101df3107ef28aa5fc5d62218f460af98ff920fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flrnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:17Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jr64s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:42Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:42 crc kubenswrapper[4824]: I0121 11:11:42.302765 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d31ab61-f94c-4f49-85d0-c9e91b83069b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a66b5c33884af00408dc35cf02f08ebd8570bd695f9f1fe558bf54fd06930431\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7514786602746dabfa690da48d48e233fe68d6dd6674aa5ecd489c0b3185f08\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61b14b6846f94cdeba8f1f61b3754e7663041f92c70356c4ee6cbb988a2e680f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a7023cfe4c098bcdd9c0477a8187f4ed7a0d2abdbe1e143436863948f441b4f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af11ad95bcf0c6bbd5fa4ab99bb592ce1c87f40d29c5ade7832414315820a569\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW0121 11:11:14.604290 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0121 11:11:14.604572 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0121 11:11:14.605875 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3284452394/tls.crt::/tmp/serving-cert-3284452394/tls.key\\\\\\\"\\\\nI0121 11:11:14.923427 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0121 11:11:14.928141 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0121 11:11:14.928162 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0121 11:11:14.928179 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0121 11:11:14.928184 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0121 11:11:14.933154 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0121 11:11:14.933230 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 11:11:14.933259 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 11:11:14.933282 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0121 11:11:14.933299 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0121 11:11:14.933315 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0121 11:11:14.933331 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0121 11:11:14.933177 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0121 11:11:14.934666 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48070c09fad511108ff22e44211cf26911509b8406d66948985cbfc3009d8233\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:42Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:42 crc kubenswrapper[4824]: I0121 11:11:42.309606 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:42 crc kubenswrapper[4824]: I0121 11:11:42.309636 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:42 crc kubenswrapper[4824]: I0121 11:11:42.309648 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:42 crc kubenswrapper[4824]: I0121 11:11:42.309658 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:42 crc kubenswrapper[4824]: I0121 11:11:42.309666 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:42Z","lastTransitionTime":"2026-01-21T11:11:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:42 crc kubenswrapper[4824]: I0121 11:11:42.312416 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9bb99c0a-683b-44da-9c0e-e8d284e87d53\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05e076febd697ed802ce5c7ee92a8c2c636a1a2c6609ad7eb01f17dace247c4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43529d1b9e1e5e22b0d56b023946119da8aacbae54b68c26ff0a1a82dc890c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ea232ad87393685aa323ce6393ff3869f8f8d423c2406a501c69c355535adfa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc6c7eff9e17b8a8ecc88bbfe703b5a01dbf1e02d342cede4f6a263a6526af43\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:42Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:42 crc kubenswrapper[4824]: I0121 11:11:42.321186 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:42Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:42 crc kubenswrapper[4824]: I0121 11:11:42.330257 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lv4rn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1267539-9adf-466f-8027-8ede009d19a0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eed1b884aa53e1ee318169a0cc2c9ad10130aa363ae4ac55e4bf065237756f53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbmfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lv4rn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:42Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:42 crc kubenswrapper[4824]: I0121 11:11:42.342612 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06f56d86-ec6f-4b57-b314-f508c0dd4d9f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d25c94f1ae0bbabb733f0e1ea1d1ba39562a92af73047a39aacc03013e239681\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca06234486b59cf990c69d9501e23f935328e2ba21013cd71ed64831ea749bcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b63e45d66bcd11ff3479cc3ec6b450017da8d21d36e967b8d496ae0f24b0ee9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52758bcd31f2ba64edf03adfd4d1be6428015abff5b64c7c029fa787085f9c29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2736e73342d2adb175b0721dd4f40cc42d26fca06d5eb3334927518531c3b107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b35ad7dd0570c191eb0080739304f3cd43f17b3d1dddf895549441ebe50deaa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b35ad7dd0570c191eb0080739304f3cd43f17b3d1dddf895549441ebe50deaa6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://80cd5ea5e6fc19a88c1aab5c61f827f103ab798a5c84da8dc4d45cea86d5aacb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80cd5ea5e6fc19a88c1aab5c61f827f103ab798a5c84da8dc4d45cea86d5aacb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://234d8b914e6a88480021647ae65d74e9b21c55b6788312a094d00c8a42677d48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://234d8b914e6a88480021647ae65d74e9b21c55b6788312a094d00c8a42677d48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:42Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:42 crc kubenswrapper[4824]: I0121 11:11:42.354824 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:42Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:42 crc kubenswrapper[4824]: I0121 11:11:42.364169 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86d3005cd61c0973659e1c1c0d3ef1ddd6fcf95bede5465f4f2099e64a74bb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:42Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:42 crc kubenswrapper[4824]: I0121 11:11:42.373481 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:42Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:42 crc kubenswrapper[4824]: I0121 11:11:42.382418 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x94nb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75d52224135b75133fc000611cd179240b9174f6492d3263341716c7e42ccb15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgr8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x94nb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:42Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:42 crc kubenswrapper[4824]: I0121 11:11:42.389784 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33f3d922-4ffe-409b-a49a-d88c85898260\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a9e57b39eed8300107e944b9fcf6a2ff1e75991f8f37387a3de615b14f9bc78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpwst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b1b37c6d2994021cfc30f85e92c40ae35677113b17aa066d52bc7a76ae5c153\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpwst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zqjg9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:42Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:42 crc kubenswrapper[4824]: I0121 11:11:42.399020 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gjtnh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a8d2cac-2d0f-4894-9bc0-2abd81699c47\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67c122a26811deaa9b70d19a1a65daee318ca3af2c626712482179d72fe67875\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2239186f1de46481cebfcafb1cab85cd320a205b5a0dfc6dd1b321d9baedeba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b2239186f1de46481cebfcafb1cab85cd320a205b5a0dfc6dd1b321d9baedeba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0db5ecd7759d974ae1b0d8065b3b8bd4d7957fae16eb5bab2d71c2fe69243816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0db5ecd7759d974ae1b0d8065b3b8bd4d7957fae16eb5bab2d71c2fe69243816\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba9b3278c208fc7aac630ca7a0daa3de0823f85d0799d6f1f278da80ae14c862\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba9b3278c208fc7aac630ca7a0daa3de0823f85d0799d6f1f278da80ae14c862\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79b8f6c016b43c3ac3ba7b22cca9a6b72c262cb17ddd6236c5d4970ecc09680c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79b8f6c016b43c3ac3ba7b22cca9a6b72c262cb17ddd6236c5d4970ecc09680c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38c4f218eb2ca7ea0c92bd0ae2ae9efeaa40746751608fbcf7f6c3f8d9e3c758\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38c4f218eb2ca7ea0c92bd0ae2ae9efeaa40746751608fbcf7f6c3f8d9e3c758\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d66ad9831d6029c8c350054525a00069e869d39c1b5638e06e9b772666bd8630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d66ad9831d6029c8c350054525a00069e869d39c1b5638e06e9b772666bd8630\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gjtnh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:42Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:42 crc kubenswrapper[4824]: I0121 11:11:42.407941 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-9rmnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f9bd12b-22ee-44ed-a8b9-c100d043f691\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2fgn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2fgn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:28Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-9rmnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:42Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:42 crc kubenswrapper[4824]: I0121 11:11:42.411681 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:42 crc kubenswrapper[4824]: I0121 11:11:42.411710 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:42 crc kubenswrapper[4824]: I0121 11:11:42.411720 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:42 crc kubenswrapper[4824]: I0121 11:11:42.411734 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:42 crc kubenswrapper[4824]: I0121 11:11:42.411744 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:42Z","lastTransitionTime":"2026-01-21T11:11:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:42 crc kubenswrapper[4824]: I0121 11:11:42.514485 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:42 crc kubenswrapper[4824]: I0121 11:11:42.514542 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:42 crc kubenswrapper[4824]: I0121 11:11:42.514551 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:42 crc kubenswrapper[4824]: I0121 11:11:42.514564 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:42 crc kubenswrapper[4824]: I0121 11:11:42.514574 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:42Z","lastTransitionTime":"2026-01-21T11:11:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:42 crc kubenswrapper[4824]: I0121 11:11:42.616656 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:42 crc kubenswrapper[4824]: I0121 11:11:42.616686 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:42 crc kubenswrapper[4824]: I0121 11:11:42.616698 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:42 crc kubenswrapper[4824]: I0121 11:11:42.616712 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:42 crc kubenswrapper[4824]: I0121 11:11:42.616720 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:42Z","lastTransitionTime":"2026-01-21T11:11:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:42 crc kubenswrapper[4824]: I0121 11:11:42.718326 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:42 crc kubenswrapper[4824]: I0121 11:11:42.718357 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:42 crc kubenswrapper[4824]: I0121 11:11:42.718365 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:42 crc kubenswrapper[4824]: I0121 11:11:42.718377 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:42 crc kubenswrapper[4824]: I0121 11:11:42.718387 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:42Z","lastTransitionTime":"2026-01-21T11:11:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:42 crc kubenswrapper[4824]: I0121 11:11:42.819918 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:42 crc kubenswrapper[4824]: I0121 11:11:42.819941 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:42 crc kubenswrapper[4824]: I0121 11:11:42.819951 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:42 crc kubenswrapper[4824]: I0121 11:11:42.819979 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:42 crc kubenswrapper[4824]: I0121 11:11:42.819989 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:42Z","lastTransitionTime":"2026-01-21T11:11:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:42 crc kubenswrapper[4824]: I0121 11:11:42.921348 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:42 crc kubenswrapper[4824]: I0121 11:11:42.921369 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:42 crc kubenswrapper[4824]: I0121 11:11:42.921378 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:42 crc kubenswrapper[4824]: I0121 11:11:42.921390 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:42 crc kubenswrapper[4824]: I0121 11:11:42.921416 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:42Z","lastTransitionTime":"2026-01-21T11:11:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:43 crc kubenswrapper[4824]: I0121 11:11:43.004017 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-25 19:01:58.423277146 +0000 UTC Jan 21 11:11:43 crc kubenswrapper[4824]: I0121 11:11:43.023999 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:43 crc kubenswrapper[4824]: I0121 11:11:43.024064 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:43 crc kubenswrapper[4824]: I0121 11:11:43.024076 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:43 crc kubenswrapper[4824]: I0121 11:11:43.024101 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:43 crc kubenswrapper[4824]: I0121 11:11:43.024405 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:43Z","lastTransitionTime":"2026-01-21T11:11:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:43 crc kubenswrapper[4824]: I0121 11:11:43.048517 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 11:11:43 crc kubenswrapper[4824]: I0121 11:11:43.048548 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 11:11:43 crc kubenswrapper[4824]: E0121 11:11:43.048694 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 21 11:11:43 crc kubenswrapper[4824]: E0121 11:11:43.048739 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 21 11:11:43 crc kubenswrapper[4824]: I0121 11:11:43.126126 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:43 crc kubenswrapper[4824]: I0121 11:11:43.126171 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:43 crc kubenswrapper[4824]: I0121 11:11:43.126181 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:43 crc kubenswrapper[4824]: I0121 11:11:43.126197 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:43 crc kubenswrapper[4824]: I0121 11:11:43.126207 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:43Z","lastTransitionTime":"2026-01-21T11:11:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:43 crc kubenswrapper[4824]: I0121 11:11:43.229123 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:43 crc kubenswrapper[4824]: I0121 11:11:43.229149 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:43 crc kubenswrapper[4824]: I0121 11:11:43.229159 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:43 crc kubenswrapper[4824]: I0121 11:11:43.229170 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:43 crc kubenswrapper[4824]: I0121 11:11:43.229177 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:43Z","lastTransitionTime":"2026-01-21T11:11:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:43 crc kubenswrapper[4824]: I0121 11:11:43.331482 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:43 crc kubenswrapper[4824]: I0121 11:11:43.331523 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:43 crc kubenswrapper[4824]: I0121 11:11:43.331537 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:43 crc kubenswrapper[4824]: I0121 11:11:43.331553 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:43 crc kubenswrapper[4824]: I0121 11:11:43.331565 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:43Z","lastTransitionTime":"2026-01-21T11:11:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:43 crc kubenswrapper[4824]: I0121 11:11:43.434257 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:43 crc kubenswrapper[4824]: I0121 11:11:43.434293 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:43 crc kubenswrapper[4824]: I0121 11:11:43.434304 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:43 crc kubenswrapper[4824]: I0121 11:11:43.434317 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:43 crc kubenswrapper[4824]: I0121 11:11:43.434328 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:43Z","lastTransitionTime":"2026-01-21T11:11:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:43 crc kubenswrapper[4824]: I0121 11:11:43.536872 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:43 crc kubenswrapper[4824]: I0121 11:11:43.536921 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:43 crc kubenswrapper[4824]: I0121 11:11:43.536931 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:43 crc kubenswrapper[4824]: I0121 11:11:43.536948 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:43 crc kubenswrapper[4824]: I0121 11:11:43.536973 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:43Z","lastTransitionTime":"2026-01-21T11:11:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:43 crc kubenswrapper[4824]: I0121 11:11:43.639298 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:43 crc kubenswrapper[4824]: I0121 11:11:43.639326 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:43 crc kubenswrapper[4824]: I0121 11:11:43.639336 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:43 crc kubenswrapper[4824]: I0121 11:11:43.639347 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:43 crc kubenswrapper[4824]: I0121 11:11:43.639382 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:43Z","lastTransitionTime":"2026-01-21T11:11:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:43 crc kubenswrapper[4824]: I0121 11:11:43.741562 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:43 crc kubenswrapper[4824]: I0121 11:11:43.741618 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:43 crc kubenswrapper[4824]: I0121 11:11:43.741630 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:43 crc kubenswrapper[4824]: I0121 11:11:43.741643 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:43 crc kubenswrapper[4824]: I0121 11:11:43.741653 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:43Z","lastTransitionTime":"2026-01-21T11:11:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:43 crc kubenswrapper[4824]: I0121 11:11:43.843608 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:43 crc kubenswrapper[4824]: I0121 11:11:43.843663 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:43 crc kubenswrapper[4824]: I0121 11:11:43.843675 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:43 crc kubenswrapper[4824]: I0121 11:11:43.843694 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:43 crc kubenswrapper[4824]: I0121 11:11:43.843707 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:43Z","lastTransitionTime":"2026-01-21T11:11:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:43 crc kubenswrapper[4824]: I0121 11:11:43.945773 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:43 crc kubenswrapper[4824]: I0121 11:11:43.945821 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:43 crc kubenswrapper[4824]: I0121 11:11:43.945830 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:43 crc kubenswrapper[4824]: I0121 11:11:43.945847 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:43 crc kubenswrapper[4824]: I0121 11:11:43.945857 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:43Z","lastTransitionTime":"2026-01-21T11:11:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:44 crc kubenswrapper[4824]: I0121 11:11:44.004629 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-01 00:01:07.384168044 +0000 UTC Jan 21 11:11:44 crc kubenswrapper[4824]: I0121 11:11:44.048316 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 11:11:44 crc kubenswrapper[4824]: I0121 11:11:44.048346 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9rmnw" Jan 21 11:11:44 crc kubenswrapper[4824]: E0121 11:11:44.048455 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 21 11:11:44 crc kubenswrapper[4824]: E0121 11:11:44.048655 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9rmnw" podUID="8f9bd12b-22ee-44ed-a8b9-c100d043f691" Jan 21 11:11:44 crc kubenswrapper[4824]: I0121 11:11:44.048675 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:44 crc kubenswrapper[4824]: I0121 11:11:44.048726 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:44 crc kubenswrapper[4824]: I0121 11:11:44.048738 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:44 crc kubenswrapper[4824]: I0121 11:11:44.048754 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:44 crc kubenswrapper[4824]: I0121 11:11:44.048765 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:44Z","lastTransitionTime":"2026-01-21T11:11:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:44 crc kubenswrapper[4824]: I0121 11:11:44.056220 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8f9bd12b-22ee-44ed-a8b9-c100d043f691-metrics-certs\") pod \"network-metrics-daemon-9rmnw\" (UID: \"8f9bd12b-22ee-44ed-a8b9-c100d043f691\") " pod="openshift-multus/network-metrics-daemon-9rmnw" Jan 21 11:11:44 crc kubenswrapper[4824]: E0121 11:11:44.056382 4824 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Jan 21 11:11:44 crc kubenswrapper[4824]: E0121 11:11:44.056448 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8f9bd12b-22ee-44ed-a8b9-c100d043f691-metrics-certs podName:8f9bd12b-22ee-44ed-a8b9-c100d043f691 nodeName:}" failed. No retries permitted until 2026-01-21 11:12:00.056431943 +0000 UTC m=+62.349461235 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/8f9bd12b-22ee-44ed-a8b9-c100d043f691-metrics-certs") pod "network-metrics-daemon-9rmnw" (UID: "8f9bd12b-22ee-44ed-a8b9-c100d043f691") : object "openshift-multus"/"metrics-daemon-secret" not registered Jan 21 11:11:44 crc kubenswrapper[4824]: I0121 11:11:44.151035 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:44 crc kubenswrapper[4824]: I0121 11:11:44.151071 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:44 crc kubenswrapper[4824]: I0121 11:11:44.151081 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:44 crc kubenswrapper[4824]: I0121 11:11:44.151095 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:44 crc kubenswrapper[4824]: I0121 11:11:44.151105 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:44Z","lastTransitionTime":"2026-01-21T11:11:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:44 crc kubenswrapper[4824]: I0121 11:11:44.252885 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:44 crc kubenswrapper[4824]: I0121 11:11:44.252925 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:44 crc kubenswrapper[4824]: I0121 11:11:44.252936 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:44 crc kubenswrapper[4824]: I0121 11:11:44.252952 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:44 crc kubenswrapper[4824]: I0121 11:11:44.252986 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:44Z","lastTransitionTime":"2026-01-21T11:11:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:44 crc kubenswrapper[4824]: I0121 11:11:44.354996 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:44 crc kubenswrapper[4824]: I0121 11:11:44.355027 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:44 crc kubenswrapper[4824]: I0121 11:11:44.355036 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:44 crc kubenswrapper[4824]: I0121 11:11:44.355048 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:44 crc kubenswrapper[4824]: I0121 11:11:44.355056 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:44Z","lastTransitionTime":"2026-01-21T11:11:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:44 crc kubenswrapper[4824]: I0121 11:11:44.457314 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:44 crc kubenswrapper[4824]: I0121 11:11:44.457345 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:44 crc kubenswrapper[4824]: I0121 11:11:44.457354 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:44 crc kubenswrapper[4824]: I0121 11:11:44.457365 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:44 crc kubenswrapper[4824]: I0121 11:11:44.457376 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:44Z","lastTransitionTime":"2026-01-21T11:11:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:44 crc kubenswrapper[4824]: I0121 11:11:44.559251 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:44 crc kubenswrapper[4824]: I0121 11:11:44.559285 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:44 crc kubenswrapper[4824]: I0121 11:11:44.559293 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:44 crc kubenswrapper[4824]: I0121 11:11:44.559312 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:44 crc kubenswrapper[4824]: I0121 11:11:44.559321 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:44Z","lastTransitionTime":"2026-01-21T11:11:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:44 crc kubenswrapper[4824]: I0121 11:11:44.662395 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:44 crc kubenswrapper[4824]: I0121 11:11:44.662426 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:44 crc kubenswrapper[4824]: I0121 11:11:44.662435 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:44 crc kubenswrapper[4824]: I0121 11:11:44.662448 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:44 crc kubenswrapper[4824]: I0121 11:11:44.662457 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:44Z","lastTransitionTime":"2026-01-21T11:11:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:44 crc kubenswrapper[4824]: I0121 11:11:44.763940 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:44 crc kubenswrapper[4824]: I0121 11:11:44.764034 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:44 crc kubenswrapper[4824]: I0121 11:11:44.764046 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:44 crc kubenswrapper[4824]: I0121 11:11:44.764071 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:44 crc kubenswrapper[4824]: I0121 11:11:44.764086 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:44Z","lastTransitionTime":"2026-01-21T11:11:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:44 crc kubenswrapper[4824]: I0121 11:11:44.866422 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:44 crc kubenswrapper[4824]: I0121 11:11:44.866465 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:44 crc kubenswrapper[4824]: I0121 11:11:44.866475 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:44 crc kubenswrapper[4824]: I0121 11:11:44.866490 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:44 crc kubenswrapper[4824]: I0121 11:11:44.866502 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:44Z","lastTransitionTime":"2026-01-21T11:11:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:44 crc kubenswrapper[4824]: I0121 11:11:44.968686 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:44 crc kubenswrapper[4824]: I0121 11:11:44.968726 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:44 crc kubenswrapper[4824]: I0121 11:11:44.968736 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:44 crc kubenswrapper[4824]: I0121 11:11:44.968749 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:44 crc kubenswrapper[4824]: I0121 11:11:44.968759 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:44Z","lastTransitionTime":"2026-01-21T11:11:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:45 crc kubenswrapper[4824]: I0121 11:11:45.004881 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-14 16:56:48.493862819 +0000 UTC Jan 21 11:11:45 crc kubenswrapper[4824]: I0121 11:11:45.048647 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 11:11:45 crc kubenswrapper[4824]: I0121 11:11:45.048706 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 11:11:45 crc kubenswrapper[4824]: E0121 11:11:45.048774 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 21 11:11:45 crc kubenswrapper[4824]: E0121 11:11:45.049075 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 21 11:11:45 crc kubenswrapper[4824]: I0121 11:11:45.071507 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:45 crc kubenswrapper[4824]: I0121 11:11:45.071544 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:45 crc kubenswrapper[4824]: I0121 11:11:45.071556 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:45 crc kubenswrapper[4824]: I0121 11:11:45.071571 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:45 crc kubenswrapper[4824]: I0121 11:11:45.071581 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:45Z","lastTransitionTime":"2026-01-21T11:11:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:45 crc kubenswrapper[4824]: I0121 11:11:45.174195 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:45 crc kubenswrapper[4824]: I0121 11:11:45.174233 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:45 crc kubenswrapper[4824]: I0121 11:11:45.174259 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:45 crc kubenswrapper[4824]: I0121 11:11:45.174274 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:45 crc kubenswrapper[4824]: I0121 11:11:45.174284 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:45Z","lastTransitionTime":"2026-01-21T11:11:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:45 crc kubenswrapper[4824]: I0121 11:11:45.276757 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:45 crc kubenswrapper[4824]: I0121 11:11:45.276792 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:45 crc kubenswrapper[4824]: I0121 11:11:45.276800 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:45 crc kubenswrapper[4824]: I0121 11:11:45.276814 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:45 crc kubenswrapper[4824]: I0121 11:11:45.276822 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:45Z","lastTransitionTime":"2026-01-21T11:11:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:45 crc kubenswrapper[4824]: I0121 11:11:45.379403 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:45 crc kubenswrapper[4824]: I0121 11:11:45.379450 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:45 crc kubenswrapper[4824]: I0121 11:11:45.379463 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:45 crc kubenswrapper[4824]: I0121 11:11:45.379479 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:45 crc kubenswrapper[4824]: I0121 11:11:45.379488 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:45Z","lastTransitionTime":"2026-01-21T11:11:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:45 crc kubenswrapper[4824]: I0121 11:11:45.481108 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:45 crc kubenswrapper[4824]: I0121 11:11:45.481145 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:45 crc kubenswrapper[4824]: I0121 11:11:45.481153 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:45 crc kubenswrapper[4824]: I0121 11:11:45.481166 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:45 crc kubenswrapper[4824]: I0121 11:11:45.481175 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:45Z","lastTransitionTime":"2026-01-21T11:11:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:45 crc kubenswrapper[4824]: I0121 11:11:45.583558 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:45 crc kubenswrapper[4824]: I0121 11:11:45.583606 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:45 crc kubenswrapper[4824]: I0121 11:11:45.583618 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:45 crc kubenswrapper[4824]: I0121 11:11:45.583634 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:45 crc kubenswrapper[4824]: I0121 11:11:45.583643 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:45Z","lastTransitionTime":"2026-01-21T11:11:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:45 crc kubenswrapper[4824]: I0121 11:11:45.685929 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:45 crc kubenswrapper[4824]: I0121 11:11:45.685994 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:45 crc kubenswrapper[4824]: I0121 11:11:45.686007 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:45 crc kubenswrapper[4824]: I0121 11:11:45.686033 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:45 crc kubenswrapper[4824]: I0121 11:11:45.686043 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:45Z","lastTransitionTime":"2026-01-21T11:11:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:45 crc kubenswrapper[4824]: I0121 11:11:45.788070 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:45 crc kubenswrapper[4824]: I0121 11:11:45.788136 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:45 crc kubenswrapper[4824]: I0121 11:11:45.788149 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:45 crc kubenswrapper[4824]: I0121 11:11:45.788178 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:45 crc kubenswrapper[4824]: I0121 11:11:45.788193 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:45Z","lastTransitionTime":"2026-01-21T11:11:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:45 crc kubenswrapper[4824]: I0121 11:11:45.889723 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:45 crc kubenswrapper[4824]: I0121 11:11:45.889762 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:45 crc kubenswrapper[4824]: I0121 11:11:45.889771 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:45 crc kubenswrapper[4824]: I0121 11:11:45.889786 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:45 crc kubenswrapper[4824]: I0121 11:11:45.889795 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:45Z","lastTransitionTime":"2026-01-21T11:11:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:45 crc kubenswrapper[4824]: I0121 11:11:45.991815 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:45 crc kubenswrapper[4824]: I0121 11:11:45.991849 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:45 crc kubenswrapper[4824]: I0121 11:11:45.991858 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:45 crc kubenswrapper[4824]: I0121 11:11:45.991870 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:45 crc kubenswrapper[4824]: I0121 11:11:45.991879 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:45Z","lastTransitionTime":"2026-01-21T11:11:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:46 crc kubenswrapper[4824]: I0121 11:11:46.005995 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-04 02:15:42.067088464 +0000 UTC Jan 21 11:11:46 crc kubenswrapper[4824]: I0121 11:11:46.049270 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9rmnw" Jan 21 11:11:46 crc kubenswrapper[4824]: E0121 11:11:46.049383 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9rmnw" podUID="8f9bd12b-22ee-44ed-a8b9-c100d043f691" Jan 21 11:11:46 crc kubenswrapper[4824]: I0121 11:11:46.049276 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 11:11:46 crc kubenswrapper[4824]: E0121 11:11:46.049725 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 21 11:11:46 crc kubenswrapper[4824]: I0121 11:11:46.094146 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:46 crc kubenswrapper[4824]: I0121 11:11:46.094174 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:46 crc kubenswrapper[4824]: I0121 11:11:46.094184 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:46 crc kubenswrapper[4824]: I0121 11:11:46.094195 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:46 crc kubenswrapper[4824]: I0121 11:11:46.094205 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:46Z","lastTransitionTime":"2026-01-21T11:11:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:46 crc kubenswrapper[4824]: I0121 11:11:46.196173 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:46 crc kubenswrapper[4824]: I0121 11:11:46.196223 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:46 crc kubenswrapper[4824]: I0121 11:11:46.196233 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:46 crc kubenswrapper[4824]: I0121 11:11:46.196245 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:46 crc kubenswrapper[4824]: I0121 11:11:46.196264 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:46Z","lastTransitionTime":"2026-01-21T11:11:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:46 crc kubenswrapper[4824]: I0121 11:11:46.298524 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:46 crc kubenswrapper[4824]: I0121 11:11:46.298563 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:46 crc kubenswrapper[4824]: I0121 11:11:46.298572 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:46 crc kubenswrapper[4824]: I0121 11:11:46.298585 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:46 crc kubenswrapper[4824]: I0121 11:11:46.298593 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:46Z","lastTransitionTime":"2026-01-21T11:11:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:46 crc kubenswrapper[4824]: I0121 11:11:46.400356 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:46 crc kubenswrapper[4824]: I0121 11:11:46.400392 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:46 crc kubenswrapper[4824]: I0121 11:11:46.400402 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:46 crc kubenswrapper[4824]: I0121 11:11:46.400416 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:46 crc kubenswrapper[4824]: I0121 11:11:46.400426 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:46Z","lastTransitionTime":"2026-01-21T11:11:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:46 crc kubenswrapper[4824]: I0121 11:11:46.501813 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:46 crc kubenswrapper[4824]: I0121 11:11:46.501845 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:46 crc kubenswrapper[4824]: I0121 11:11:46.501852 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:46 crc kubenswrapper[4824]: I0121 11:11:46.501864 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:46 crc kubenswrapper[4824]: I0121 11:11:46.501872 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:46Z","lastTransitionTime":"2026-01-21T11:11:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:46 crc kubenswrapper[4824]: I0121 11:11:46.603553 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:46 crc kubenswrapper[4824]: I0121 11:11:46.603578 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:46 crc kubenswrapper[4824]: I0121 11:11:46.603586 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:46 crc kubenswrapper[4824]: I0121 11:11:46.603597 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:46 crc kubenswrapper[4824]: I0121 11:11:46.603606 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:46Z","lastTransitionTime":"2026-01-21T11:11:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:46 crc kubenswrapper[4824]: I0121 11:11:46.705137 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:46 crc kubenswrapper[4824]: I0121 11:11:46.705169 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:46 crc kubenswrapper[4824]: I0121 11:11:46.705177 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:46 crc kubenswrapper[4824]: I0121 11:11:46.705191 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:46 crc kubenswrapper[4824]: I0121 11:11:46.705199 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:46Z","lastTransitionTime":"2026-01-21T11:11:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:46 crc kubenswrapper[4824]: I0121 11:11:46.782168 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 21 11:11:46 crc kubenswrapper[4824]: I0121 11:11:46.782240 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 11:11:46 crc kubenswrapper[4824]: I0121 11:11:46.782278 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 11:11:46 crc kubenswrapper[4824]: E0121 11:11:46.782331 4824 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Jan 21 11:11:46 crc kubenswrapper[4824]: E0121 11:11:46.782333 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-21 11:12:18.782312245 +0000 UTC m=+81.075341547 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 11:11:46 crc kubenswrapper[4824]: E0121 11:11:46.782410 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-21 11:12:18.782389001 +0000 UTC m=+81.075418303 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Jan 21 11:11:46 crc kubenswrapper[4824]: E0121 11:11:46.782421 4824 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 21 11:11:46 crc kubenswrapper[4824]: E0121 11:11:46.782473 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-21 11:12:18.7824626 +0000 UTC m=+81.075491892 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 21 11:11:46 crc kubenswrapper[4824]: I0121 11:11:46.807454 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:46 crc kubenswrapper[4824]: I0121 11:11:46.807500 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:46 crc kubenswrapper[4824]: I0121 11:11:46.807510 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:46 crc kubenswrapper[4824]: I0121 11:11:46.807520 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:46 crc kubenswrapper[4824]: I0121 11:11:46.807529 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:46Z","lastTransitionTime":"2026-01-21T11:11:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:46 crc kubenswrapper[4824]: I0121 11:11:46.883183 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 11:11:46 crc kubenswrapper[4824]: I0121 11:11:46.883240 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 11:11:46 crc kubenswrapper[4824]: E0121 11:11:46.883307 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 21 11:11:46 crc kubenswrapper[4824]: E0121 11:11:46.883331 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 21 11:11:46 crc kubenswrapper[4824]: E0121 11:11:46.883343 4824 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 21 11:11:46 crc kubenswrapper[4824]: E0121 11:11:46.883383 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-01-21 11:12:18.883371316 +0000 UTC m=+81.176400618 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 21 11:11:46 crc kubenswrapper[4824]: E0121 11:11:46.883331 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 21 11:11:46 crc kubenswrapper[4824]: E0121 11:11:46.883465 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 21 11:11:46 crc kubenswrapper[4824]: E0121 11:11:46.883477 4824 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 21 11:11:46 crc kubenswrapper[4824]: E0121 11:11:46.883508 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-01-21 11:12:18.883499678 +0000 UTC m=+81.176528971 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 21 11:11:46 crc kubenswrapper[4824]: I0121 11:11:46.909266 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:46 crc kubenswrapper[4824]: I0121 11:11:46.909291 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:46 crc kubenswrapper[4824]: I0121 11:11:46.909299 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:46 crc kubenswrapper[4824]: I0121 11:11:46.909310 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:46 crc kubenswrapper[4824]: I0121 11:11:46.909318 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:46Z","lastTransitionTime":"2026-01-21T11:11:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:47 crc kubenswrapper[4824]: I0121 11:11:47.006790 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-31 13:54:49.079161465 +0000 UTC Jan 21 11:11:47 crc kubenswrapper[4824]: I0121 11:11:47.011054 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:47 crc kubenswrapper[4824]: I0121 11:11:47.011085 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:47 crc kubenswrapper[4824]: I0121 11:11:47.011094 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:47 crc kubenswrapper[4824]: I0121 11:11:47.011106 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:47 crc kubenswrapper[4824]: I0121 11:11:47.011115 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:47Z","lastTransitionTime":"2026-01-21T11:11:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:47 crc kubenswrapper[4824]: I0121 11:11:47.048647 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 11:11:47 crc kubenswrapper[4824]: I0121 11:11:47.048692 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 11:11:47 crc kubenswrapper[4824]: E0121 11:11:47.048731 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 21 11:11:47 crc kubenswrapper[4824]: E0121 11:11:47.048798 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 21 11:11:47 crc kubenswrapper[4824]: I0121 11:11:47.112573 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:47 crc kubenswrapper[4824]: I0121 11:11:47.112601 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:47 crc kubenswrapper[4824]: I0121 11:11:47.112610 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:47 crc kubenswrapper[4824]: I0121 11:11:47.112620 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:47 crc kubenswrapper[4824]: I0121 11:11:47.112630 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:47Z","lastTransitionTime":"2026-01-21T11:11:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:47 crc kubenswrapper[4824]: I0121 11:11:47.214285 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:47 crc kubenswrapper[4824]: I0121 11:11:47.214315 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:47 crc kubenswrapper[4824]: I0121 11:11:47.214323 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:47 crc kubenswrapper[4824]: I0121 11:11:47.214336 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:47 crc kubenswrapper[4824]: I0121 11:11:47.214344 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:47Z","lastTransitionTime":"2026-01-21T11:11:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:47 crc kubenswrapper[4824]: I0121 11:11:47.316408 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:47 crc kubenswrapper[4824]: I0121 11:11:47.316441 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:47 crc kubenswrapper[4824]: I0121 11:11:47.316449 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:47 crc kubenswrapper[4824]: I0121 11:11:47.316460 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:47 crc kubenswrapper[4824]: I0121 11:11:47.316468 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:47Z","lastTransitionTime":"2026-01-21T11:11:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:47 crc kubenswrapper[4824]: I0121 11:11:47.418463 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:47 crc kubenswrapper[4824]: I0121 11:11:47.418497 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:47 crc kubenswrapper[4824]: I0121 11:11:47.418505 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:47 crc kubenswrapper[4824]: I0121 11:11:47.418521 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:47 crc kubenswrapper[4824]: I0121 11:11:47.418529 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:47Z","lastTransitionTime":"2026-01-21T11:11:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:47 crc kubenswrapper[4824]: I0121 11:11:47.519699 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:47 crc kubenswrapper[4824]: I0121 11:11:47.519733 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:47 crc kubenswrapper[4824]: I0121 11:11:47.519741 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:47 crc kubenswrapper[4824]: I0121 11:11:47.519752 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:47 crc kubenswrapper[4824]: I0121 11:11:47.519761 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:47Z","lastTransitionTime":"2026-01-21T11:11:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:47 crc kubenswrapper[4824]: I0121 11:11:47.621236 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:47 crc kubenswrapper[4824]: I0121 11:11:47.621281 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:47 crc kubenswrapper[4824]: I0121 11:11:47.621290 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:47 crc kubenswrapper[4824]: I0121 11:11:47.621301 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:47 crc kubenswrapper[4824]: I0121 11:11:47.621309 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:47Z","lastTransitionTime":"2026-01-21T11:11:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:47 crc kubenswrapper[4824]: I0121 11:11:47.723108 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:47 crc kubenswrapper[4824]: I0121 11:11:47.723141 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:47 crc kubenswrapper[4824]: I0121 11:11:47.723150 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:47 crc kubenswrapper[4824]: I0121 11:11:47.723161 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:47 crc kubenswrapper[4824]: I0121 11:11:47.723169 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:47Z","lastTransitionTime":"2026-01-21T11:11:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:47 crc kubenswrapper[4824]: I0121 11:11:47.825044 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:47 crc kubenswrapper[4824]: I0121 11:11:47.825067 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:47 crc kubenswrapper[4824]: I0121 11:11:47.825075 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:47 crc kubenswrapper[4824]: I0121 11:11:47.825086 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:47 crc kubenswrapper[4824]: I0121 11:11:47.825093 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:47Z","lastTransitionTime":"2026-01-21T11:11:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:47 crc kubenswrapper[4824]: I0121 11:11:47.926509 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:47 crc kubenswrapper[4824]: I0121 11:11:47.926542 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:47 crc kubenswrapper[4824]: I0121 11:11:47.926551 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:47 crc kubenswrapper[4824]: I0121 11:11:47.926562 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:47 crc kubenswrapper[4824]: I0121 11:11:47.926570 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:47Z","lastTransitionTime":"2026-01-21T11:11:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.007697 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-29 18:26:34.155619471 +0000 UTC Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.027691 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.027717 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.027725 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.027735 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.027744 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:48Z","lastTransitionTime":"2026-01-21T11:11:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.049231 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 11:11:48 crc kubenswrapper[4824]: E0121 11:11:48.049330 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.049388 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9rmnw" Jan 21 11:11:48 crc kubenswrapper[4824]: E0121 11:11:48.049508 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9rmnw" podUID="8f9bd12b-22ee-44ed-a8b9-c100d043f691" Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.057557 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lv4rn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1267539-9adf-466f-8027-8ede009d19a0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eed1b884aa53e1ee318169a0cc2c9ad10130aa363ae4ac55e4bf065237756f53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbmfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lv4rn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:48Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.064918 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jr64s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9768c60d-5bee-4286-bced-076e584b04e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91158f14d2e98ac9f875a5e7101df3107ef28aa5fc5d62218f460af98ff920fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flrnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:17Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jr64s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:48Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.073668 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d31ab61-f94c-4f49-85d0-c9e91b83069b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a66b5c33884af00408dc35cf02f08ebd8570bd695f9f1fe558bf54fd06930431\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7514786602746dabfa690da48d48e233fe68d6dd6674aa5ecd489c0b3185f08\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61b14b6846f94cdeba8f1f61b3754e7663041f92c70356c4ee6cbb988a2e680f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a7023cfe4c098bcdd9c0477a8187f4ed7a0d2abdbe1e143436863948f441b4f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af11ad95bcf0c6bbd5fa4ab99bb592ce1c87f40d29c5ade7832414315820a569\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW0121 11:11:14.604290 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0121 11:11:14.604572 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0121 11:11:14.605875 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3284452394/tls.crt::/tmp/serving-cert-3284452394/tls.key\\\\\\\"\\\\nI0121 11:11:14.923427 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0121 11:11:14.928141 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0121 11:11:14.928162 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0121 11:11:14.928179 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0121 11:11:14.928184 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0121 11:11:14.933154 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0121 11:11:14.933230 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 11:11:14.933259 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 11:11:14.933282 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0121 11:11:14.933299 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0121 11:11:14.933315 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0121 11:11:14.933331 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0121 11:11:14.933177 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0121 11:11:14.934666 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48070c09fad511108ff22e44211cf26911509b8406d66948985cbfc3009d8233\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:48Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.081249 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9bb99c0a-683b-44da-9c0e-e8d284e87d53\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05e076febd697ed802ce5c7ee92a8c2c636a1a2c6609ad7eb01f17dace247c4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43529d1b9e1e5e22b0d56b023946119da8aacbae54b68c26ff0a1a82dc890c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ea232ad87393685aa323ce6393ff3869f8f8d423c2406a501c69c355535adfa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc6c7eff9e17b8a8ecc88bbfe703b5a01dbf1e02d342cede4f6a263a6526af43\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:48Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.089815 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:48Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.101869 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06f56d86-ec6f-4b57-b314-f508c0dd4d9f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d25c94f1ae0bbabb733f0e1ea1d1ba39562a92af73047a39aacc03013e239681\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca06234486b59cf990c69d9501e23f935328e2ba21013cd71ed64831ea749bcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b63e45d66bcd11ff3479cc3ec6b450017da8d21d36e967b8d496ae0f24b0ee9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52758bcd31f2ba64edf03adfd4d1be6428015abff5b64c7c029fa787085f9c29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2736e73342d2adb175b0721dd4f40cc42d26fca06d5eb3334927518531c3b107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b35ad7dd0570c191eb0080739304f3cd43f17b3d1dddf895549441ebe50deaa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b35ad7dd0570c191eb0080739304f3cd43f17b3d1dddf895549441ebe50deaa6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://80cd5ea5e6fc19a88c1aab5c61f827f103ab798a5c84da8dc4d45cea86d5aacb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80cd5ea5e6fc19a88c1aab5c61f827f103ab798a5c84da8dc4d45cea86d5aacb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://234d8b914e6a88480021647ae65d74e9b21c55b6788312a094d00c8a42677d48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://234d8b914e6a88480021647ae65d74e9b21c55b6788312a094d00c8a42677d48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:48Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.109531 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:48Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.116006 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-9rmnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f9bd12b-22ee-44ed-a8b9-c100d043f691\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2fgn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2fgn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:28Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-9rmnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:48Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.123720 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86d3005cd61c0973659e1c1c0d3ef1ddd6fcf95bede5465f4f2099e64a74bb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:48Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.129715 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.129736 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.129744 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.129757 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.129768 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:48Z","lastTransitionTime":"2026-01-21T11:11:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.131189 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:48Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.139776 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x94nb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75d52224135b75133fc000611cd179240b9174f6492d3263341716c7e42ccb15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgr8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x94nb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:48Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.146824 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33f3d922-4ffe-409b-a49a-d88c85898260\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a9e57b39eed8300107e944b9fcf6a2ff1e75991f8f37387a3de615b14f9bc78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpwst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b1b37c6d2994021cfc30f85e92c40ae35677113b17aa066d52bc7a76ae5c153\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpwst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zqjg9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:48Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.155872 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gjtnh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a8d2cac-2d0f-4894-9bc0-2abd81699c47\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67c122a26811deaa9b70d19a1a65daee318ca3af2c626712482179d72fe67875\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2239186f1de46481cebfcafb1cab85cd320a205b5a0dfc6dd1b321d9baedeba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b2239186f1de46481cebfcafb1cab85cd320a205b5a0dfc6dd1b321d9baedeba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0db5ecd7759d974ae1b0d8065b3b8bd4d7957fae16eb5bab2d71c2fe69243816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0db5ecd7759d974ae1b0d8065b3b8bd4d7957fae16eb5bab2d71c2fe69243816\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba9b3278c208fc7aac630ca7a0daa3de0823f85d0799d6f1f278da80ae14c862\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba9b3278c208fc7aac630ca7a0daa3de0823f85d0799d6f1f278da80ae14c862\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79b8f6c016b43c3ac3ba7b22cca9a6b72c262cb17ddd6236c5d4970ecc09680c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79b8f6c016b43c3ac3ba7b22cca9a6b72c262cb17ddd6236c5d4970ecc09680c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38c4f218eb2ca7ea0c92bd0ae2ae9efeaa40746751608fbcf7f6c3f8d9e3c758\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38c4f218eb2ca7ea0c92bd0ae2ae9efeaa40746751608fbcf7f6c3f8d9e3c758\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d66ad9831d6029c8c350054525a00069e869d39c1b5638e06e9b772666bd8630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d66ad9831d6029c8c350054525a00069e869d39c1b5638e06e9b772666bd8630\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gjtnh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:48Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.163125 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d87422473838f5218c9fb646ab2515f2f42592754ffdf07b19f7a8e093fa48a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:48Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.171810 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://516c009d4720ab5b393c7482872f12fed22d4f4414bd1f3400411c0e8da88ce4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee18d293036971832c8d9c3f7b9eca84ab779ff6768a2db3c6f7bdcf3433cdef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:48Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.183649 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"26b1108a-4b98-4e39-a3cd-e0d055089fd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc319bbc06568cc811ded2b418b04b986a31f321fed492e8be50104af79bb30b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cee9cc9eec3fc815f1621a22a37805691288dbb60dd3ff64194a84be62abcddc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc4c96e33e069f7d08d33caf92cd0987668506f931fd134b246d5ce5e934f95c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ddbb0e5ae02e9d5d82a80ead7efc91f1b8e557a9fb6f39940e5530fae8de2d38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://64a9065254f889b1c43d63759283335bf78fc4fa1531a4baae472c06c075dae3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e94ef8a8f47dc75ecfbab2e3fb8e7097c1b73eed32d7ca947f11ea198cea880\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c655fe698872c3a7804fed5e2e75f300110a816eae4412891f3433762ed0161\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2c655fe698872c3a7804fed5e2e75f300110a816eae4412891f3433762ed0161\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-21T11:11:40Z\\\",\\\"message\\\":\\\"ps://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:40Z is after 2025-08-24T17:21:41Z]\\\\nI0121 11:11:40.696180 6500 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf\\\\nI0121 11:11:40.696191 6500 obj_retry.go:303] Retry object setup: *v1.Pod openshift-kube-apiserver/kube-apiserver-crc\\\\nI0121 11:11:40.696193 6500 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf in node crc\\\\nI0121 11:11:40.696185 6500 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-apiserver-operator/metrics]} name:Service_openshift-apiserver-operator/metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.38:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {8b82f026-5975-4a1b-bb18-08d5d51147ec}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0121 11:11:40.696197 6500 obj_retry.go:365] Adding new object: *v1.Pod openshift-kube-apiserver/kube\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:40Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-htkvk_openshift-ovn-kubernetes(26b1108a-4b98-4e39-a3cd-e0d055089fd8)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948e0497eaffa7b0d84b4085c9ed8d987eb2c6527c9a82c10e576ed9c9c1e21b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-htkvk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:48Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.190501 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-np2v8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4698abe4-0f33-4349-a2a4-614de80ae21b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://191328471bb6f2b34cde2eb927da697619826bf0b6174ecb583c8c25d27a6654\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n9qbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://050bd5bdded2162d5e9dac81713f4ed63cb575d1776e828b2f98e50d89f11b04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n9qbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-np2v8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:48Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.221080 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.221114 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.221140 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.221166 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.221174 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:48Z","lastTransitionTime":"2026-01-21T11:11:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:48 crc kubenswrapper[4824]: E0121 11:11:48.229111 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:48Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:48Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"695c676c-720c-4921-9e71-56b96f2fb46b\\\",\\\"systemUUID\\\":\\\"d00624f5-786f-41dc-bdc1-a1875f167d65\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:48Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.231207 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.231239 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.231249 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.231260 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.231282 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:48Z","lastTransitionTime":"2026-01-21T11:11:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:48 crc kubenswrapper[4824]: E0121 11:11:48.238793 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:48Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:48Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"695c676c-720c-4921-9e71-56b96f2fb46b\\\",\\\"systemUUID\\\":\\\"d00624f5-786f-41dc-bdc1-a1875f167d65\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:48Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.240650 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.240683 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.240692 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.240705 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.240714 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:48Z","lastTransitionTime":"2026-01-21T11:11:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:48 crc kubenswrapper[4824]: E0121 11:11:48.249735 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:48Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:48Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"695c676c-720c-4921-9e71-56b96f2fb46b\\\",\\\"systemUUID\\\":\\\"d00624f5-786f-41dc-bdc1-a1875f167d65\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:48Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.251886 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.251907 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.251914 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.251923 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.251930 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:48Z","lastTransitionTime":"2026-01-21T11:11:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:48 crc kubenswrapper[4824]: E0121 11:11:48.259503 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:48Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:48Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"695c676c-720c-4921-9e71-56b96f2fb46b\\\",\\\"systemUUID\\\":\\\"d00624f5-786f-41dc-bdc1-a1875f167d65\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:48Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.261508 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.261592 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.261602 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.261614 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.261623 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:48Z","lastTransitionTime":"2026-01-21T11:11:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:48 crc kubenswrapper[4824]: E0121 11:11:48.269903 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:48Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:48Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"695c676c-720c-4921-9e71-56b96f2fb46b\\\",\\\"systemUUID\\\":\\\"d00624f5-786f-41dc-bdc1-a1875f167d65\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:48Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:48 crc kubenswrapper[4824]: E0121 11:11:48.270023 4824 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.270838 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.270862 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.270872 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.270880 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.270887 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:48Z","lastTransitionTime":"2026-01-21T11:11:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.372614 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.372642 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.372651 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.372662 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.372669 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:48Z","lastTransitionTime":"2026-01-21T11:11:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.473971 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.474003 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.474012 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.474024 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.474031 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:48Z","lastTransitionTime":"2026-01-21T11:11:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.576262 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.576309 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.576320 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.576333 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.576342 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:48Z","lastTransitionTime":"2026-01-21T11:11:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.678087 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.678123 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.678131 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.678143 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.678152 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:48Z","lastTransitionTime":"2026-01-21T11:11:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.779895 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.779926 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.779934 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.779944 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.779951 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:48Z","lastTransitionTime":"2026-01-21T11:11:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.882140 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.882178 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.882187 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.882198 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.882205 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:48Z","lastTransitionTime":"2026-01-21T11:11:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.984026 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.984073 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.984085 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.984099 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:48 crc kubenswrapper[4824]: I0121 11:11:48.984109 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:48Z","lastTransitionTime":"2026-01-21T11:11:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:49 crc kubenswrapper[4824]: I0121 11:11:49.008648 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-08 10:30:56.229177316 +0000 UTC Jan 21 11:11:49 crc kubenswrapper[4824]: I0121 11:11:49.049194 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 11:11:49 crc kubenswrapper[4824]: I0121 11:11:49.049203 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 11:11:49 crc kubenswrapper[4824]: E0121 11:11:49.049302 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 21 11:11:49 crc kubenswrapper[4824]: E0121 11:11:49.049379 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 21 11:11:49 crc kubenswrapper[4824]: I0121 11:11:49.085680 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:49 crc kubenswrapper[4824]: I0121 11:11:49.085715 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:49 crc kubenswrapper[4824]: I0121 11:11:49.085727 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:49 crc kubenswrapper[4824]: I0121 11:11:49.085738 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:49 crc kubenswrapper[4824]: I0121 11:11:49.085746 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:49Z","lastTransitionTime":"2026-01-21T11:11:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:49 crc kubenswrapper[4824]: I0121 11:11:49.187680 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:49 crc kubenswrapper[4824]: I0121 11:11:49.187709 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:49 crc kubenswrapper[4824]: I0121 11:11:49.187718 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:49 crc kubenswrapper[4824]: I0121 11:11:49.187729 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:49 crc kubenswrapper[4824]: I0121 11:11:49.187738 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:49Z","lastTransitionTime":"2026-01-21T11:11:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:49 crc kubenswrapper[4824]: I0121 11:11:49.290207 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:49 crc kubenswrapper[4824]: I0121 11:11:49.290251 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:49 crc kubenswrapper[4824]: I0121 11:11:49.290260 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:49 crc kubenswrapper[4824]: I0121 11:11:49.290272 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:49 crc kubenswrapper[4824]: I0121 11:11:49.290295 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:49Z","lastTransitionTime":"2026-01-21T11:11:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:49 crc kubenswrapper[4824]: I0121 11:11:49.392001 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:49 crc kubenswrapper[4824]: I0121 11:11:49.392043 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:49 crc kubenswrapper[4824]: I0121 11:11:49.392075 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:49 crc kubenswrapper[4824]: I0121 11:11:49.392087 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:49 crc kubenswrapper[4824]: I0121 11:11:49.392095 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:49Z","lastTransitionTime":"2026-01-21T11:11:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:49 crc kubenswrapper[4824]: I0121 11:11:49.493784 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:49 crc kubenswrapper[4824]: I0121 11:11:49.493823 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:49 crc kubenswrapper[4824]: I0121 11:11:49.493831 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:49 crc kubenswrapper[4824]: I0121 11:11:49.493843 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:49 crc kubenswrapper[4824]: I0121 11:11:49.493854 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:49Z","lastTransitionTime":"2026-01-21T11:11:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:49 crc kubenswrapper[4824]: I0121 11:11:49.596161 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:49 crc kubenswrapper[4824]: I0121 11:11:49.596200 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:49 crc kubenswrapper[4824]: I0121 11:11:49.596210 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:49 crc kubenswrapper[4824]: I0121 11:11:49.596225 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:49 crc kubenswrapper[4824]: I0121 11:11:49.596234 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:49Z","lastTransitionTime":"2026-01-21T11:11:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:49 crc kubenswrapper[4824]: I0121 11:11:49.698556 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:49 crc kubenswrapper[4824]: I0121 11:11:49.698591 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:49 crc kubenswrapper[4824]: I0121 11:11:49.698600 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:49 crc kubenswrapper[4824]: I0121 11:11:49.698613 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:49 crc kubenswrapper[4824]: I0121 11:11:49.698623 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:49Z","lastTransitionTime":"2026-01-21T11:11:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:49 crc kubenswrapper[4824]: I0121 11:11:49.800502 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:49 crc kubenswrapper[4824]: I0121 11:11:49.800544 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:49 crc kubenswrapper[4824]: I0121 11:11:49.800554 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:49 crc kubenswrapper[4824]: I0121 11:11:49.800571 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:49 crc kubenswrapper[4824]: I0121 11:11:49.800580 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:49Z","lastTransitionTime":"2026-01-21T11:11:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:49 crc kubenswrapper[4824]: I0121 11:11:49.902575 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:49 crc kubenswrapper[4824]: I0121 11:11:49.902630 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:49 crc kubenswrapper[4824]: I0121 11:11:49.902641 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:49 crc kubenswrapper[4824]: I0121 11:11:49.902665 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:49 crc kubenswrapper[4824]: I0121 11:11:49.902679 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:49Z","lastTransitionTime":"2026-01-21T11:11:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:50 crc kubenswrapper[4824]: I0121 11:11:50.005190 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:50 crc kubenswrapper[4824]: I0121 11:11:50.005242 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:50 crc kubenswrapper[4824]: I0121 11:11:50.005253 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:50 crc kubenswrapper[4824]: I0121 11:11:50.005266 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:50 crc kubenswrapper[4824]: I0121 11:11:50.005295 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:50Z","lastTransitionTime":"2026-01-21T11:11:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:50 crc kubenswrapper[4824]: I0121 11:11:50.009415 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-26 01:38:54.221692997 +0000 UTC Jan 21 11:11:50 crc kubenswrapper[4824]: I0121 11:11:50.048587 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 11:11:50 crc kubenswrapper[4824]: E0121 11:11:50.048703 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 21 11:11:50 crc kubenswrapper[4824]: I0121 11:11:50.048588 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9rmnw" Jan 21 11:11:50 crc kubenswrapper[4824]: E0121 11:11:50.048892 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9rmnw" podUID="8f9bd12b-22ee-44ed-a8b9-c100d043f691" Jan 21 11:11:50 crc kubenswrapper[4824]: I0121 11:11:50.107403 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:50 crc kubenswrapper[4824]: I0121 11:11:50.107437 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:50 crc kubenswrapper[4824]: I0121 11:11:50.107448 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:50 crc kubenswrapper[4824]: I0121 11:11:50.107462 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:50 crc kubenswrapper[4824]: I0121 11:11:50.107473 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:50Z","lastTransitionTime":"2026-01-21T11:11:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:50 crc kubenswrapper[4824]: I0121 11:11:50.208891 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:50 crc kubenswrapper[4824]: I0121 11:11:50.208922 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:50 crc kubenswrapper[4824]: I0121 11:11:50.208931 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:50 crc kubenswrapper[4824]: I0121 11:11:50.208947 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:50 crc kubenswrapper[4824]: I0121 11:11:50.209018 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:50Z","lastTransitionTime":"2026-01-21T11:11:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:50 crc kubenswrapper[4824]: I0121 11:11:50.311562 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:50 crc kubenswrapper[4824]: I0121 11:11:50.311598 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:50 crc kubenswrapper[4824]: I0121 11:11:50.311608 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:50 crc kubenswrapper[4824]: I0121 11:11:50.311625 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:50 crc kubenswrapper[4824]: I0121 11:11:50.311635 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:50Z","lastTransitionTime":"2026-01-21T11:11:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:50 crc kubenswrapper[4824]: I0121 11:11:50.413808 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:50 crc kubenswrapper[4824]: I0121 11:11:50.414095 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:50 crc kubenswrapper[4824]: I0121 11:11:50.414105 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:50 crc kubenswrapper[4824]: I0121 11:11:50.414121 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:50 crc kubenswrapper[4824]: I0121 11:11:50.414133 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:50Z","lastTransitionTime":"2026-01-21T11:11:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:50 crc kubenswrapper[4824]: I0121 11:11:50.515451 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:50 crc kubenswrapper[4824]: I0121 11:11:50.515485 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:50 crc kubenswrapper[4824]: I0121 11:11:50.515493 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:50 crc kubenswrapper[4824]: I0121 11:11:50.515505 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:50 crc kubenswrapper[4824]: I0121 11:11:50.515514 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:50Z","lastTransitionTime":"2026-01-21T11:11:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:50 crc kubenswrapper[4824]: I0121 11:11:50.617118 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:50 crc kubenswrapper[4824]: I0121 11:11:50.617164 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:50 crc kubenswrapper[4824]: I0121 11:11:50.617173 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:50 crc kubenswrapper[4824]: I0121 11:11:50.617185 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:50 crc kubenswrapper[4824]: I0121 11:11:50.617195 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:50Z","lastTransitionTime":"2026-01-21T11:11:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:50 crc kubenswrapper[4824]: I0121 11:11:50.718414 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:50 crc kubenswrapper[4824]: I0121 11:11:50.718452 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:50 crc kubenswrapper[4824]: I0121 11:11:50.718463 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:50 crc kubenswrapper[4824]: I0121 11:11:50.718476 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:50 crc kubenswrapper[4824]: I0121 11:11:50.718485 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:50Z","lastTransitionTime":"2026-01-21T11:11:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:50 crc kubenswrapper[4824]: I0121 11:11:50.820124 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:50 crc kubenswrapper[4824]: I0121 11:11:50.820151 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:50 crc kubenswrapper[4824]: I0121 11:11:50.820184 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:50 crc kubenswrapper[4824]: I0121 11:11:50.820199 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:50 crc kubenswrapper[4824]: I0121 11:11:50.820208 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:50Z","lastTransitionTime":"2026-01-21T11:11:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:50 crc kubenswrapper[4824]: I0121 11:11:50.921704 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:50 crc kubenswrapper[4824]: I0121 11:11:50.921740 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:50 crc kubenswrapper[4824]: I0121 11:11:50.921749 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:50 crc kubenswrapper[4824]: I0121 11:11:50.921759 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:50 crc kubenswrapper[4824]: I0121 11:11:50.921767 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:50Z","lastTransitionTime":"2026-01-21T11:11:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:51 crc kubenswrapper[4824]: I0121 11:11:51.009751 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-17 01:44:52.968773125 +0000 UTC Jan 21 11:11:51 crc kubenswrapper[4824]: I0121 11:11:51.023329 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:51 crc kubenswrapper[4824]: I0121 11:11:51.023370 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:51 crc kubenswrapper[4824]: I0121 11:11:51.023380 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:51 crc kubenswrapper[4824]: I0121 11:11:51.023394 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:51 crc kubenswrapper[4824]: I0121 11:11:51.023403 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:51Z","lastTransitionTime":"2026-01-21T11:11:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:51 crc kubenswrapper[4824]: I0121 11:11:51.048579 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 11:11:51 crc kubenswrapper[4824]: I0121 11:11:51.048606 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 11:11:51 crc kubenswrapper[4824]: E0121 11:11:51.048667 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 21 11:11:51 crc kubenswrapper[4824]: E0121 11:11:51.048737 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 21 11:11:51 crc kubenswrapper[4824]: I0121 11:11:51.125596 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:51 crc kubenswrapper[4824]: I0121 11:11:51.125626 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:51 crc kubenswrapper[4824]: I0121 11:11:51.125634 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:51 crc kubenswrapper[4824]: I0121 11:11:51.125646 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:51 crc kubenswrapper[4824]: I0121 11:11:51.125654 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:51Z","lastTransitionTime":"2026-01-21T11:11:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:51 crc kubenswrapper[4824]: I0121 11:11:51.227391 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:51 crc kubenswrapper[4824]: I0121 11:11:51.227422 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:51 crc kubenswrapper[4824]: I0121 11:11:51.227432 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:51 crc kubenswrapper[4824]: I0121 11:11:51.227445 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:51 crc kubenswrapper[4824]: I0121 11:11:51.227456 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:51Z","lastTransitionTime":"2026-01-21T11:11:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:51 crc kubenswrapper[4824]: I0121 11:11:51.329630 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:51 crc kubenswrapper[4824]: I0121 11:11:51.329668 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:51 crc kubenswrapper[4824]: I0121 11:11:51.329677 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:51 crc kubenswrapper[4824]: I0121 11:11:51.329690 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:51 crc kubenswrapper[4824]: I0121 11:11:51.329700 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:51Z","lastTransitionTime":"2026-01-21T11:11:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:51 crc kubenswrapper[4824]: I0121 11:11:51.432242 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:51 crc kubenswrapper[4824]: I0121 11:11:51.432287 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:51 crc kubenswrapper[4824]: I0121 11:11:51.432311 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:51 crc kubenswrapper[4824]: I0121 11:11:51.432327 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:51 crc kubenswrapper[4824]: I0121 11:11:51.432337 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:51Z","lastTransitionTime":"2026-01-21T11:11:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:51 crc kubenswrapper[4824]: I0121 11:11:51.534950 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:51 crc kubenswrapper[4824]: I0121 11:11:51.535028 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:51 crc kubenswrapper[4824]: I0121 11:11:51.535038 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:51 crc kubenswrapper[4824]: I0121 11:11:51.535055 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:51 crc kubenswrapper[4824]: I0121 11:11:51.535063 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:51Z","lastTransitionTime":"2026-01-21T11:11:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:51 crc kubenswrapper[4824]: I0121 11:11:51.637149 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:51 crc kubenswrapper[4824]: I0121 11:11:51.637195 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:51 crc kubenswrapper[4824]: I0121 11:11:51.637206 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:51 crc kubenswrapper[4824]: I0121 11:11:51.637223 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:51 crc kubenswrapper[4824]: I0121 11:11:51.637231 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:51Z","lastTransitionTime":"2026-01-21T11:11:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:51 crc kubenswrapper[4824]: I0121 11:11:51.739261 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:51 crc kubenswrapper[4824]: I0121 11:11:51.739351 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:51 crc kubenswrapper[4824]: I0121 11:11:51.739372 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:51 crc kubenswrapper[4824]: I0121 11:11:51.739399 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:51 crc kubenswrapper[4824]: I0121 11:11:51.739412 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:51Z","lastTransitionTime":"2026-01-21T11:11:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:51 crc kubenswrapper[4824]: I0121 11:11:51.841927 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:51 crc kubenswrapper[4824]: I0121 11:11:51.841974 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:51 crc kubenswrapper[4824]: I0121 11:11:51.841985 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:51 crc kubenswrapper[4824]: I0121 11:11:51.841999 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:51 crc kubenswrapper[4824]: I0121 11:11:51.842008 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:51Z","lastTransitionTime":"2026-01-21T11:11:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:51 crc kubenswrapper[4824]: I0121 11:11:51.944223 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:51 crc kubenswrapper[4824]: I0121 11:11:51.944292 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:51 crc kubenswrapper[4824]: I0121 11:11:51.944317 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:51 crc kubenswrapper[4824]: I0121 11:11:51.944340 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:51 crc kubenswrapper[4824]: I0121 11:11:51.944354 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:51Z","lastTransitionTime":"2026-01-21T11:11:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:52 crc kubenswrapper[4824]: I0121 11:11:52.010930 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-06 03:06:13.665639909 +0000 UTC Jan 21 11:11:52 crc kubenswrapper[4824]: I0121 11:11:52.046597 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:52 crc kubenswrapper[4824]: I0121 11:11:52.046675 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:52 crc kubenswrapper[4824]: I0121 11:11:52.046686 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:52 crc kubenswrapper[4824]: I0121 11:11:52.046704 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:52 crc kubenswrapper[4824]: I0121 11:11:52.046714 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:52Z","lastTransitionTime":"2026-01-21T11:11:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:52 crc kubenswrapper[4824]: I0121 11:11:52.048927 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9rmnw" Jan 21 11:11:52 crc kubenswrapper[4824]: I0121 11:11:52.048980 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 11:11:52 crc kubenswrapper[4824]: E0121 11:11:52.049046 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9rmnw" podUID="8f9bd12b-22ee-44ed-a8b9-c100d043f691" Jan 21 11:11:52 crc kubenswrapper[4824]: E0121 11:11:52.049131 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 21 11:11:52 crc kubenswrapper[4824]: I0121 11:11:52.149020 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:52 crc kubenswrapper[4824]: I0121 11:11:52.149090 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:52 crc kubenswrapper[4824]: I0121 11:11:52.149103 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:52 crc kubenswrapper[4824]: I0121 11:11:52.149123 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:52 crc kubenswrapper[4824]: I0121 11:11:52.149134 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:52Z","lastTransitionTime":"2026-01-21T11:11:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:52 crc kubenswrapper[4824]: I0121 11:11:52.253516 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:52 crc kubenswrapper[4824]: I0121 11:11:52.253554 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:52 crc kubenswrapper[4824]: I0121 11:11:52.253563 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:52 crc kubenswrapper[4824]: I0121 11:11:52.253575 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:52 crc kubenswrapper[4824]: I0121 11:11:52.253583 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:52Z","lastTransitionTime":"2026-01-21T11:11:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:52 crc kubenswrapper[4824]: I0121 11:11:52.355732 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:52 crc kubenswrapper[4824]: I0121 11:11:52.355764 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:52 crc kubenswrapper[4824]: I0121 11:11:52.355773 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:52 crc kubenswrapper[4824]: I0121 11:11:52.355785 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:52 crc kubenswrapper[4824]: I0121 11:11:52.355792 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:52Z","lastTransitionTime":"2026-01-21T11:11:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:52 crc kubenswrapper[4824]: I0121 11:11:52.457420 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:52 crc kubenswrapper[4824]: I0121 11:11:52.457452 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:52 crc kubenswrapper[4824]: I0121 11:11:52.457460 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:52 crc kubenswrapper[4824]: I0121 11:11:52.457471 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:52 crc kubenswrapper[4824]: I0121 11:11:52.457479 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:52Z","lastTransitionTime":"2026-01-21T11:11:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:52 crc kubenswrapper[4824]: I0121 11:11:52.559765 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:52 crc kubenswrapper[4824]: I0121 11:11:52.559794 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:52 crc kubenswrapper[4824]: I0121 11:11:52.559801 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:52 crc kubenswrapper[4824]: I0121 11:11:52.559811 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:52 crc kubenswrapper[4824]: I0121 11:11:52.559818 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:52Z","lastTransitionTime":"2026-01-21T11:11:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:52 crc kubenswrapper[4824]: I0121 11:11:52.662158 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:52 crc kubenswrapper[4824]: I0121 11:11:52.662192 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:52 crc kubenswrapper[4824]: I0121 11:11:52.662202 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:52 crc kubenswrapper[4824]: I0121 11:11:52.662213 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:52 crc kubenswrapper[4824]: I0121 11:11:52.662221 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:52Z","lastTransitionTime":"2026-01-21T11:11:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:52 crc kubenswrapper[4824]: I0121 11:11:52.763614 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:52 crc kubenswrapper[4824]: I0121 11:11:52.763636 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:52 crc kubenswrapper[4824]: I0121 11:11:52.763645 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:52 crc kubenswrapper[4824]: I0121 11:11:52.763655 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:52 crc kubenswrapper[4824]: I0121 11:11:52.763663 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:52Z","lastTransitionTime":"2026-01-21T11:11:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:52 crc kubenswrapper[4824]: I0121 11:11:52.865604 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:52 crc kubenswrapper[4824]: I0121 11:11:52.865636 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:52 crc kubenswrapper[4824]: I0121 11:11:52.865644 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:52 crc kubenswrapper[4824]: I0121 11:11:52.865654 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:52 crc kubenswrapper[4824]: I0121 11:11:52.865662 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:52Z","lastTransitionTime":"2026-01-21T11:11:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:52 crc kubenswrapper[4824]: I0121 11:11:52.967484 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:52 crc kubenswrapper[4824]: I0121 11:11:52.967513 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:52 crc kubenswrapper[4824]: I0121 11:11:52.967521 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:52 crc kubenswrapper[4824]: I0121 11:11:52.967531 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:52 crc kubenswrapper[4824]: I0121 11:11:52.967541 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:52Z","lastTransitionTime":"2026-01-21T11:11:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:53 crc kubenswrapper[4824]: I0121 11:11:53.012151 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-29 19:26:48.282007098 +0000 UTC Jan 21 11:11:53 crc kubenswrapper[4824]: I0121 11:11:53.048836 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 11:11:53 crc kubenswrapper[4824]: I0121 11:11:53.048862 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 11:11:53 crc kubenswrapper[4824]: E0121 11:11:53.048939 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 21 11:11:53 crc kubenswrapper[4824]: E0121 11:11:53.049023 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 21 11:11:53 crc kubenswrapper[4824]: I0121 11:11:53.049829 4824 scope.go:117] "RemoveContainer" containerID="2c655fe698872c3a7804fed5e2e75f300110a816eae4412891f3433762ed0161" Jan 21 11:11:53 crc kubenswrapper[4824]: E0121 11:11:53.050163 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-htkvk_openshift-ovn-kubernetes(26b1108a-4b98-4e39-a3cd-e0d055089fd8)\"" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" podUID="26b1108a-4b98-4e39-a3cd-e0d055089fd8" Jan 21 11:11:53 crc kubenswrapper[4824]: I0121 11:11:53.069484 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:53 crc kubenswrapper[4824]: I0121 11:11:53.069524 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:53 crc kubenswrapper[4824]: I0121 11:11:53.069538 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:53 crc kubenswrapper[4824]: I0121 11:11:53.069553 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:53 crc kubenswrapper[4824]: I0121 11:11:53.069566 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:53Z","lastTransitionTime":"2026-01-21T11:11:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:53 crc kubenswrapper[4824]: I0121 11:11:53.171720 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:53 crc kubenswrapper[4824]: I0121 11:11:53.171792 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:53 crc kubenswrapper[4824]: I0121 11:11:53.171832 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:53 crc kubenswrapper[4824]: I0121 11:11:53.171861 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:53 crc kubenswrapper[4824]: I0121 11:11:53.171874 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:53Z","lastTransitionTime":"2026-01-21T11:11:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:53 crc kubenswrapper[4824]: I0121 11:11:53.273868 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:53 crc kubenswrapper[4824]: I0121 11:11:53.273899 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:53 crc kubenswrapper[4824]: I0121 11:11:53.273909 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:53 crc kubenswrapper[4824]: I0121 11:11:53.273923 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:53 crc kubenswrapper[4824]: I0121 11:11:53.273950 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:53Z","lastTransitionTime":"2026-01-21T11:11:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:53 crc kubenswrapper[4824]: I0121 11:11:53.375779 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:53 crc kubenswrapper[4824]: I0121 11:11:53.375826 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:53 crc kubenswrapper[4824]: I0121 11:11:53.375836 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:53 crc kubenswrapper[4824]: I0121 11:11:53.375850 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:53 crc kubenswrapper[4824]: I0121 11:11:53.375860 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:53Z","lastTransitionTime":"2026-01-21T11:11:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:53 crc kubenswrapper[4824]: I0121 11:11:53.477202 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:53 crc kubenswrapper[4824]: I0121 11:11:53.477235 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:53 crc kubenswrapper[4824]: I0121 11:11:53.477243 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:53 crc kubenswrapper[4824]: I0121 11:11:53.477256 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:53 crc kubenswrapper[4824]: I0121 11:11:53.477266 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:53Z","lastTransitionTime":"2026-01-21T11:11:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:53 crc kubenswrapper[4824]: I0121 11:11:53.579574 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:53 crc kubenswrapper[4824]: I0121 11:11:53.579634 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:53 crc kubenswrapper[4824]: I0121 11:11:53.579647 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:53 crc kubenswrapper[4824]: I0121 11:11:53.579675 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:53 crc kubenswrapper[4824]: I0121 11:11:53.579688 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:53Z","lastTransitionTime":"2026-01-21T11:11:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:53 crc kubenswrapper[4824]: I0121 11:11:53.682010 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:53 crc kubenswrapper[4824]: I0121 11:11:53.682081 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:53 crc kubenswrapper[4824]: I0121 11:11:53.682093 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:53 crc kubenswrapper[4824]: I0121 11:11:53.682117 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:53 crc kubenswrapper[4824]: I0121 11:11:53.682133 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:53Z","lastTransitionTime":"2026-01-21T11:11:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:53 crc kubenswrapper[4824]: I0121 11:11:53.718426 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Jan 21 11:11:53 crc kubenswrapper[4824]: I0121 11:11:53.727647 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d87422473838f5218c9fb646ab2515f2f42592754ffdf07b19f7a8e093fa48a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:53Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:53 crc kubenswrapper[4824]: I0121 11:11:53.727670 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Jan 21 11:11:53 crc kubenswrapper[4824]: I0121 11:11:53.737167 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://516c009d4720ab5b393c7482872f12fed22d4f4414bd1f3400411c0e8da88ce4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee18d293036971832c8d9c3f7b9eca84ab779ff6768a2db3c6f7bdcf3433cdef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:53Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:53 crc kubenswrapper[4824]: I0121 11:11:53.753081 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"26b1108a-4b98-4e39-a3cd-e0d055089fd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc319bbc06568cc811ded2b418b04b986a31f321fed492e8be50104af79bb30b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cee9cc9eec3fc815f1621a22a37805691288dbb60dd3ff64194a84be62abcddc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc4c96e33e069f7d08d33caf92cd0987668506f931fd134b246d5ce5e934f95c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ddbb0e5ae02e9d5d82a80ead7efc91f1b8e557a9fb6f39940e5530fae8de2d38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://64a9065254f889b1c43d63759283335bf78fc4fa1531a4baae472c06c075dae3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e94ef8a8f47dc75ecfbab2e3fb8e7097c1b73eed32d7ca947f11ea198cea880\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c655fe698872c3a7804fed5e2e75f300110a816eae4412891f3433762ed0161\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2c655fe698872c3a7804fed5e2e75f300110a816eae4412891f3433762ed0161\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-21T11:11:40Z\\\",\\\"message\\\":\\\"ps://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:40Z is after 2025-08-24T17:21:41Z]\\\\nI0121 11:11:40.696180 6500 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf\\\\nI0121 11:11:40.696191 6500 obj_retry.go:303] Retry object setup: *v1.Pod openshift-kube-apiserver/kube-apiserver-crc\\\\nI0121 11:11:40.696193 6500 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf in node crc\\\\nI0121 11:11:40.696185 6500 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-apiserver-operator/metrics]} name:Service_openshift-apiserver-operator/metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.38:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {8b82f026-5975-4a1b-bb18-08d5d51147ec}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0121 11:11:40.696197 6500 obj_retry.go:365] Adding new object: *v1.Pod openshift-kube-apiserver/kube\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:40Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-htkvk_openshift-ovn-kubernetes(26b1108a-4b98-4e39-a3cd-e0d055089fd8)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948e0497eaffa7b0d84b4085c9ed8d987eb2c6527c9a82c10e576ed9c9c1e21b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-htkvk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:53Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:53 crc kubenswrapper[4824]: I0121 11:11:53.760468 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-np2v8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4698abe4-0f33-4349-a2a4-614de80ae21b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://191328471bb6f2b34cde2eb927da697619826bf0b6174ecb583c8c25d27a6654\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n9qbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://050bd5bdded2162d5e9dac81713f4ed63cb575d1776e828b2f98e50d89f11b04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n9qbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-np2v8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:53Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:53 crc kubenswrapper[4824]: I0121 11:11:53.768748 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:53Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:53 crc kubenswrapper[4824]: I0121 11:11:53.776270 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lv4rn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1267539-9adf-466f-8027-8ede009d19a0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eed1b884aa53e1ee318169a0cc2c9ad10130aa363ae4ac55e4bf065237756f53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbmfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lv4rn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:53Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:53 crc kubenswrapper[4824]: I0121 11:11:53.783310 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jr64s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9768c60d-5bee-4286-bced-076e584b04e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91158f14d2e98ac9f875a5e7101df3107ef28aa5fc5d62218f460af98ff920fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flrnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:17Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jr64s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:53Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:53 crc kubenswrapper[4824]: I0121 11:11:53.784031 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:53 crc kubenswrapper[4824]: I0121 11:11:53.784095 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:53 crc kubenswrapper[4824]: I0121 11:11:53.784110 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:53 crc kubenswrapper[4824]: I0121 11:11:53.784125 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:53 crc kubenswrapper[4824]: I0121 11:11:53.784137 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:53Z","lastTransitionTime":"2026-01-21T11:11:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:53 crc kubenswrapper[4824]: I0121 11:11:53.792480 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d31ab61-f94c-4f49-85d0-c9e91b83069b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a66b5c33884af00408dc35cf02f08ebd8570bd695f9f1fe558bf54fd06930431\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7514786602746dabfa690da48d48e233fe68d6dd6674aa5ecd489c0b3185f08\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61b14b6846f94cdeba8f1f61b3754e7663041f92c70356c4ee6cbb988a2e680f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a7023cfe4c098bcdd9c0477a8187f4ed7a0d2abdbe1e143436863948f441b4f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af11ad95bcf0c6bbd5fa4ab99bb592ce1c87f40d29c5ade7832414315820a569\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW0121 11:11:14.604290 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0121 11:11:14.604572 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0121 11:11:14.605875 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3284452394/tls.crt::/tmp/serving-cert-3284452394/tls.key\\\\\\\"\\\\nI0121 11:11:14.923427 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0121 11:11:14.928141 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0121 11:11:14.928162 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0121 11:11:14.928179 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0121 11:11:14.928184 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0121 11:11:14.933154 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0121 11:11:14.933230 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 11:11:14.933259 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 11:11:14.933282 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0121 11:11:14.933299 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0121 11:11:14.933315 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0121 11:11:14.933331 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0121 11:11:14.933177 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0121 11:11:14.934666 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48070c09fad511108ff22e44211cf26911509b8406d66948985cbfc3009d8233\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:53Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:53 crc kubenswrapper[4824]: I0121 11:11:53.801515 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9bb99c0a-683b-44da-9c0e-e8d284e87d53\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05e076febd697ed802ce5c7ee92a8c2c636a1a2c6609ad7eb01f17dace247c4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43529d1b9e1e5e22b0d56b023946119da8aacbae54b68c26ff0a1a82dc890c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ea232ad87393685aa323ce6393ff3869f8f8d423c2406a501c69c355535adfa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc6c7eff9e17b8a8ecc88bbfe703b5a01dbf1e02d342cede4f6a263a6526af43\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:53Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:53 crc kubenswrapper[4824]: I0121 11:11:53.815660 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06f56d86-ec6f-4b57-b314-f508c0dd4d9f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d25c94f1ae0bbabb733f0e1ea1d1ba39562a92af73047a39aacc03013e239681\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca06234486b59cf990c69d9501e23f935328e2ba21013cd71ed64831ea749bcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b63e45d66bcd11ff3479cc3ec6b450017da8d21d36e967b8d496ae0f24b0ee9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52758bcd31f2ba64edf03adfd4d1be6428015abff5b64c7c029fa787085f9c29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2736e73342d2adb175b0721dd4f40cc42d26fca06d5eb3334927518531c3b107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b35ad7dd0570c191eb0080739304f3cd43f17b3d1dddf895549441ebe50deaa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b35ad7dd0570c191eb0080739304f3cd43f17b3d1dddf895549441ebe50deaa6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://80cd5ea5e6fc19a88c1aab5c61f827f103ab798a5c84da8dc4d45cea86d5aacb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80cd5ea5e6fc19a88c1aab5c61f827f103ab798a5c84da8dc4d45cea86d5aacb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://234d8b914e6a88480021647ae65d74e9b21c55b6788312a094d00c8a42677d48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://234d8b914e6a88480021647ae65d74e9b21c55b6788312a094d00c8a42677d48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:53Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:53 crc kubenswrapper[4824]: I0121 11:11:53.824377 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:53Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:53 crc kubenswrapper[4824]: I0121 11:11:53.834757 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gjtnh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a8d2cac-2d0f-4894-9bc0-2abd81699c47\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67c122a26811deaa9b70d19a1a65daee318ca3af2c626712482179d72fe67875\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2239186f1de46481cebfcafb1cab85cd320a205b5a0dfc6dd1b321d9baedeba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b2239186f1de46481cebfcafb1cab85cd320a205b5a0dfc6dd1b321d9baedeba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0db5ecd7759d974ae1b0d8065b3b8bd4d7957fae16eb5bab2d71c2fe69243816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0db5ecd7759d974ae1b0d8065b3b8bd4d7957fae16eb5bab2d71c2fe69243816\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba9b3278c208fc7aac630ca7a0daa3de0823f85d0799d6f1f278da80ae14c862\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba9b3278c208fc7aac630ca7a0daa3de0823f85d0799d6f1f278da80ae14c862\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79b8f6c016b43c3ac3ba7b22cca9a6b72c262cb17ddd6236c5d4970ecc09680c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79b8f6c016b43c3ac3ba7b22cca9a6b72c262cb17ddd6236c5d4970ecc09680c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38c4f218eb2ca7ea0c92bd0ae2ae9efeaa40746751608fbcf7f6c3f8d9e3c758\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38c4f218eb2ca7ea0c92bd0ae2ae9efeaa40746751608fbcf7f6c3f8d9e3c758\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d66ad9831d6029c8c350054525a00069e869d39c1b5638e06e9b772666bd8630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d66ad9831d6029c8c350054525a00069e869d39c1b5638e06e9b772666bd8630\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gjtnh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:53Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:53 crc kubenswrapper[4824]: I0121 11:11:53.842152 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-9rmnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f9bd12b-22ee-44ed-a8b9-c100d043f691\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2fgn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2fgn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:28Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-9rmnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:53Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:53 crc kubenswrapper[4824]: I0121 11:11:53.850429 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86d3005cd61c0973659e1c1c0d3ef1ddd6fcf95bede5465f4f2099e64a74bb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:53Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:53 crc kubenswrapper[4824]: I0121 11:11:53.860572 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:53Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:53 crc kubenswrapper[4824]: I0121 11:11:53.870064 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x94nb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75d52224135b75133fc000611cd179240b9174f6492d3263341716c7e42ccb15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgr8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x94nb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:53Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:53 crc kubenswrapper[4824]: I0121 11:11:53.877879 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33f3d922-4ffe-409b-a49a-d88c85898260\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a9e57b39eed8300107e944b9fcf6a2ff1e75991f8f37387a3de615b14f9bc78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpwst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b1b37c6d2994021cfc30f85e92c40ae35677113b17aa066d52bc7a76ae5c153\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpwst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zqjg9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:53Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:53 crc kubenswrapper[4824]: I0121 11:11:53.886277 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:53 crc kubenswrapper[4824]: I0121 11:11:53.886354 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:53 crc kubenswrapper[4824]: I0121 11:11:53.886363 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:53 crc kubenswrapper[4824]: I0121 11:11:53.886394 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:53 crc kubenswrapper[4824]: I0121 11:11:53.886405 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:53Z","lastTransitionTime":"2026-01-21T11:11:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:53 crc kubenswrapper[4824]: I0121 11:11:53.988156 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:53 crc kubenswrapper[4824]: I0121 11:11:53.988189 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:53 crc kubenswrapper[4824]: I0121 11:11:53.988198 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:53 crc kubenswrapper[4824]: I0121 11:11:53.988211 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:53 crc kubenswrapper[4824]: I0121 11:11:53.988219 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:53Z","lastTransitionTime":"2026-01-21T11:11:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:54 crc kubenswrapper[4824]: I0121 11:11:54.012613 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-11 22:26:05.602798767 +0000 UTC Jan 21 11:11:54 crc kubenswrapper[4824]: I0121 11:11:54.048395 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 11:11:54 crc kubenswrapper[4824]: I0121 11:11:54.048433 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9rmnw" Jan 21 11:11:54 crc kubenswrapper[4824]: E0121 11:11:54.048527 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 21 11:11:54 crc kubenswrapper[4824]: E0121 11:11:54.048598 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9rmnw" podUID="8f9bd12b-22ee-44ed-a8b9-c100d043f691" Jan 21 11:11:54 crc kubenswrapper[4824]: I0121 11:11:54.090445 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:54 crc kubenswrapper[4824]: I0121 11:11:54.090478 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:54 crc kubenswrapper[4824]: I0121 11:11:54.090486 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:54 crc kubenswrapper[4824]: I0121 11:11:54.090498 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:54 crc kubenswrapper[4824]: I0121 11:11:54.090506 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:54Z","lastTransitionTime":"2026-01-21T11:11:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:54 crc kubenswrapper[4824]: I0121 11:11:54.192454 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:54 crc kubenswrapper[4824]: I0121 11:11:54.192498 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:54 crc kubenswrapper[4824]: I0121 11:11:54.192507 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:54 crc kubenswrapper[4824]: I0121 11:11:54.192520 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:54 crc kubenswrapper[4824]: I0121 11:11:54.192528 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:54Z","lastTransitionTime":"2026-01-21T11:11:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:54 crc kubenswrapper[4824]: I0121 11:11:54.293952 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:54 crc kubenswrapper[4824]: I0121 11:11:54.294000 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:54 crc kubenswrapper[4824]: I0121 11:11:54.294008 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:54 crc kubenswrapper[4824]: I0121 11:11:54.294020 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:54 crc kubenswrapper[4824]: I0121 11:11:54.294028 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:54Z","lastTransitionTime":"2026-01-21T11:11:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:54 crc kubenswrapper[4824]: I0121 11:11:54.395276 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:54 crc kubenswrapper[4824]: I0121 11:11:54.395310 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:54 crc kubenswrapper[4824]: I0121 11:11:54.395329 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:54 crc kubenswrapper[4824]: I0121 11:11:54.395342 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:54 crc kubenswrapper[4824]: I0121 11:11:54.395350 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:54Z","lastTransitionTime":"2026-01-21T11:11:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:54 crc kubenswrapper[4824]: I0121 11:11:54.496504 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:54 crc kubenswrapper[4824]: I0121 11:11:54.496527 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:54 crc kubenswrapper[4824]: I0121 11:11:54.496534 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:54 crc kubenswrapper[4824]: I0121 11:11:54.496544 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:54 crc kubenswrapper[4824]: I0121 11:11:54.496551 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:54Z","lastTransitionTime":"2026-01-21T11:11:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:54 crc kubenswrapper[4824]: I0121 11:11:54.598189 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:54 crc kubenswrapper[4824]: I0121 11:11:54.598241 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:54 crc kubenswrapper[4824]: I0121 11:11:54.598250 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:54 crc kubenswrapper[4824]: I0121 11:11:54.598260 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:54 crc kubenswrapper[4824]: I0121 11:11:54.598268 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:54Z","lastTransitionTime":"2026-01-21T11:11:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:54 crc kubenswrapper[4824]: I0121 11:11:54.700655 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:54 crc kubenswrapper[4824]: I0121 11:11:54.700722 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:54 crc kubenswrapper[4824]: I0121 11:11:54.700737 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:54 crc kubenswrapper[4824]: I0121 11:11:54.700761 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:54 crc kubenswrapper[4824]: I0121 11:11:54.700775 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:54Z","lastTransitionTime":"2026-01-21T11:11:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:54 crc kubenswrapper[4824]: I0121 11:11:54.802527 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:54 crc kubenswrapper[4824]: I0121 11:11:54.802559 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:54 crc kubenswrapper[4824]: I0121 11:11:54.802569 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:54 crc kubenswrapper[4824]: I0121 11:11:54.802581 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:54 crc kubenswrapper[4824]: I0121 11:11:54.802589 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:54Z","lastTransitionTime":"2026-01-21T11:11:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:54 crc kubenswrapper[4824]: I0121 11:11:54.904479 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:54 crc kubenswrapper[4824]: I0121 11:11:54.904500 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:54 crc kubenswrapper[4824]: I0121 11:11:54.904514 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:54 crc kubenswrapper[4824]: I0121 11:11:54.904523 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:54 crc kubenswrapper[4824]: I0121 11:11:54.904529 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:54Z","lastTransitionTime":"2026-01-21T11:11:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:55 crc kubenswrapper[4824]: I0121 11:11:55.006384 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:55 crc kubenswrapper[4824]: I0121 11:11:55.006479 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:55 crc kubenswrapper[4824]: I0121 11:11:55.006489 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:55 crc kubenswrapper[4824]: I0121 11:11:55.006524 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:55 crc kubenswrapper[4824]: I0121 11:11:55.006533 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:55Z","lastTransitionTime":"2026-01-21T11:11:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:55 crc kubenswrapper[4824]: I0121 11:11:55.012736 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-31 19:19:19.242227906 +0000 UTC Jan 21 11:11:55 crc kubenswrapper[4824]: I0121 11:11:55.049123 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 11:11:55 crc kubenswrapper[4824]: E0121 11:11:55.049201 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 21 11:11:55 crc kubenswrapper[4824]: I0121 11:11:55.049123 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 11:11:55 crc kubenswrapper[4824]: E0121 11:11:55.049283 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 21 11:11:55 crc kubenswrapper[4824]: I0121 11:11:55.108066 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:55 crc kubenswrapper[4824]: I0121 11:11:55.108104 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:55 crc kubenswrapper[4824]: I0121 11:11:55.108113 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:55 crc kubenswrapper[4824]: I0121 11:11:55.108122 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:55 crc kubenswrapper[4824]: I0121 11:11:55.108130 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:55Z","lastTransitionTime":"2026-01-21T11:11:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:55 crc kubenswrapper[4824]: I0121 11:11:55.209068 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:55 crc kubenswrapper[4824]: I0121 11:11:55.209096 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:55 crc kubenswrapper[4824]: I0121 11:11:55.209104 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:55 crc kubenswrapper[4824]: I0121 11:11:55.209114 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:55 crc kubenswrapper[4824]: I0121 11:11:55.209126 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:55Z","lastTransitionTime":"2026-01-21T11:11:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:55 crc kubenswrapper[4824]: I0121 11:11:55.310682 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:55 crc kubenswrapper[4824]: I0121 11:11:55.310708 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:55 crc kubenswrapper[4824]: I0121 11:11:55.310716 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:55 crc kubenswrapper[4824]: I0121 11:11:55.310726 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:55 crc kubenswrapper[4824]: I0121 11:11:55.310734 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:55Z","lastTransitionTime":"2026-01-21T11:11:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:55 crc kubenswrapper[4824]: I0121 11:11:55.413086 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:55 crc kubenswrapper[4824]: I0121 11:11:55.413134 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:55 crc kubenswrapper[4824]: I0121 11:11:55.413144 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:55 crc kubenswrapper[4824]: I0121 11:11:55.413155 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:55 crc kubenswrapper[4824]: I0121 11:11:55.413163 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:55Z","lastTransitionTime":"2026-01-21T11:11:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:55 crc kubenswrapper[4824]: I0121 11:11:55.515041 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:55 crc kubenswrapper[4824]: I0121 11:11:55.515070 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:55 crc kubenswrapper[4824]: I0121 11:11:55.515079 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:55 crc kubenswrapper[4824]: I0121 11:11:55.515089 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:55 crc kubenswrapper[4824]: I0121 11:11:55.515096 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:55Z","lastTransitionTime":"2026-01-21T11:11:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:55 crc kubenswrapper[4824]: I0121 11:11:55.616887 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:55 crc kubenswrapper[4824]: I0121 11:11:55.616916 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:55 crc kubenswrapper[4824]: I0121 11:11:55.616925 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:55 crc kubenswrapper[4824]: I0121 11:11:55.616935 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:55 crc kubenswrapper[4824]: I0121 11:11:55.616942 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:55Z","lastTransitionTime":"2026-01-21T11:11:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:55 crc kubenswrapper[4824]: I0121 11:11:55.726220 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:55 crc kubenswrapper[4824]: I0121 11:11:55.727058 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:55 crc kubenswrapper[4824]: I0121 11:11:55.727066 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:55 crc kubenswrapper[4824]: I0121 11:11:55.727078 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:55 crc kubenswrapper[4824]: I0121 11:11:55.727086 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:55Z","lastTransitionTime":"2026-01-21T11:11:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:55 crc kubenswrapper[4824]: I0121 11:11:55.828815 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:55 crc kubenswrapper[4824]: I0121 11:11:55.828845 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:55 crc kubenswrapper[4824]: I0121 11:11:55.828853 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:55 crc kubenswrapper[4824]: I0121 11:11:55.828864 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:55 crc kubenswrapper[4824]: I0121 11:11:55.828871 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:55Z","lastTransitionTime":"2026-01-21T11:11:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:55 crc kubenswrapper[4824]: I0121 11:11:55.930780 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:55 crc kubenswrapper[4824]: I0121 11:11:55.930814 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:55 crc kubenswrapper[4824]: I0121 11:11:55.930823 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:55 crc kubenswrapper[4824]: I0121 11:11:55.930847 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:55 crc kubenswrapper[4824]: I0121 11:11:55.930855 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:55Z","lastTransitionTime":"2026-01-21T11:11:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:56 crc kubenswrapper[4824]: I0121 11:11:56.013691 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-14 14:58:39.769297053 +0000 UTC Jan 21 11:11:56 crc kubenswrapper[4824]: I0121 11:11:56.033090 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:56 crc kubenswrapper[4824]: I0121 11:11:56.033117 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:56 crc kubenswrapper[4824]: I0121 11:11:56.033142 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:56 crc kubenswrapper[4824]: I0121 11:11:56.033154 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:56 crc kubenswrapper[4824]: I0121 11:11:56.033162 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:56Z","lastTransitionTime":"2026-01-21T11:11:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:56 crc kubenswrapper[4824]: I0121 11:11:56.048283 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9rmnw" Jan 21 11:11:56 crc kubenswrapper[4824]: I0121 11:11:56.048305 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 11:11:56 crc kubenswrapper[4824]: E0121 11:11:56.048398 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9rmnw" podUID="8f9bd12b-22ee-44ed-a8b9-c100d043f691" Jan 21 11:11:56 crc kubenswrapper[4824]: E0121 11:11:56.048449 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 21 11:11:56 crc kubenswrapper[4824]: I0121 11:11:56.135244 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:56 crc kubenswrapper[4824]: I0121 11:11:56.135279 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:56 crc kubenswrapper[4824]: I0121 11:11:56.135288 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:56 crc kubenswrapper[4824]: I0121 11:11:56.135302 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:56 crc kubenswrapper[4824]: I0121 11:11:56.135312 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:56Z","lastTransitionTime":"2026-01-21T11:11:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:56 crc kubenswrapper[4824]: I0121 11:11:56.237601 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:56 crc kubenswrapper[4824]: I0121 11:11:56.237636 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:56 crc kubenswrapper[4824]: I0121 11:11:56.237647 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:56 crc kubenswrapper[4824]: I0121 11:11:56.237660 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:56 crc kubenswrapper[4824]: I0121 11:11:56.237669 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:56Z","lastTransitionTime":"2026-01-21T11:11:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:56 crc kubenswrapper[4824]: I0121 11:11:56.339897 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:56 crc kubenswrapper[4824]: I0121 11:11:56.339947 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:56 crc kubenswrapper[4824]: I0121 11:11:56.339976 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:56 crc kubenswrapper[4824]: I0121 11:11:56.339990 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:56 crc kubenswrapper[4824]: I0121 11:11:56.339998 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:56Z","lastTransitionTime":"2026-01-21T11:11:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:56 crc kubenswrapper[4824]: I0121 11:11:56.442297 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:56 crc kubenswrapper[4824]: I0121 11:11:56.442348 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:56 crc kubenswrapper[4824]: I0121 11:11:56.442361 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:56 crc kubenswrapper[4824]: I0121 11:11:56.442377 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:56 crc kubenswrapper[4824]: I0121 11:11:56.442387 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:56Z","lastTransitionTime":"2026-01-21T11:11:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:56 crc kubenswrapper[4824]: I0121 11:11:56.544272 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:56 crc kubenswrapper[4824]: I0121 11:11:56.544303 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:56 crc kubenswrapper[4824]: I0121 11:11:56.544311 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:56 crc kubenswrapper[4824]: I0121 11:11:56.544321 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:56 crc kubenswrapper[4824]: I0121 11:11:56.544329 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:56Z","lastTransitionTime":"2026-01-21T11:11:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:56 crc kubenswrapper[4824]: I0121 11:11:56.646136 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:56 crc kubenswrapper[4824]: I0121 11:11:56.646168 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:56 crc kubenswrapper[4824]: I0121 11:11:56.646177 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:56 crc kubenswrapper[4824]: I0121 11:11:56.646188 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:56 crc kubenswrapper[4824]: I0121 11:11:56.646196 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:56Z","lastTransitionTime":"2026-01-21T11:11:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:56 crc kubenswrapper[4824]: I0121 11:11:56.748358 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:56 crc kubenswrapper[4824]: I0121 11:11:56.748391 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:56 crc kubenswrapper[4824]: I0121 11:11:56.748399 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:56 crc kubenswrapper[4824]: I0121 11:11:56.748409 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:56 crc kubenswrapper[4824]: I0121 11:11:56.748420 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:56Z","lastTransitionTime":"2026-01-21T11:11:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:56 crc kubenswrapper[4824]: I0121 11:11:56.850776 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:56 crc kubenswrapper[4824]: I0121 11:11:56.850865 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:56 crc kubenswrapper[4824]: I0121 11:11:56.850876 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:56 crc kubenswrapper[4824]: I0121 11:11:56.850904 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:56 crc kubenswrapper[4824]: I0121 11:11:56.850918 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:56Z","lastTransitionTime":"2026-01-21T11:11:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:56 crc kubenswrapper[4824]: I0121 11:11:56.953765 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:56 crc kubenswrapper[4824]: I0121 11:11:56.953807 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:56 crc kubenswrapper[4824]: I0121 11:11:56.953816 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:56 crc kubenswrapper[4824]: I0121 11:11:56.953832 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:56 crc kubenswrapper[4824]: I0121 11:11:56.953842 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:56Z","lastTransitionTime":"2026-01-21T11:11:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:57 crc kubenswrapper[4824]: I0121 11:11:57.014545 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-12 15:14:16.766274193 +0000 UTC Jan 21 11:11:57 crc kubenswrapper[4824]: I0121 11:11:57.048813 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 11:11:57 crc kubenswrapper[4824]: E0121 11:11:57.048933 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 21 11:11:57 crc kubenswrapper[4824]: I0121 11:11:57.049024 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 11:11:57 crc kubenswrapper[4824]: E0121 11:11:57.049109 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 21 11:11:57 crc kubenswrapper[4824]: I0121 11:11:57.056567 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:57 crc kubenswrapper[4824]: I0121 11:11:57.056602 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:57 crc kubenswrapper[4824]: I0121 11:11:57.056612 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:57 crc kubenswrapper[4824]: I0121 11:11:57.056626 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:57 crc kubenswrapper[4824]: I0121 11:11:57.056635 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:57Z","lastTransitionTime":"2026-01-21T11:11:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:57 crc kubenswrapper[4824]: I0121 11:11:57.158769 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:57 crc kubenswrapper[4824]: I0121 11:11:57.158816 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:57 crc kubenswrapper[4824]: I0121 11:11:57.158824 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:57 crc kubenswrapper[4824]: I0121 11:11:57.158836 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:57 crc kubenswrapper[4824]: I0121 11:11:57.158844 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:57Z","lastTransitionTime":"2026-01-21T11:11:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:57 crc kubenswrapper[4824]: I0121 11:11:57.261158 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:57 crc kubenswrapper[4824]: I0121 11:11:57.261195 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:57 crc kubenswrapper[4824]: I0121 11:11:57.261204 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:57 crc kubenswrapper[4824]: I0121 11:11:57.261217 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:57 crc kubenswrapper[4824]: I0121 11:11:57.261228 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:57Z","lastTransitionTime":"2026-01-21T11:11:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:57 crc kubenswrapper[4824]: I0121 11:11:57.363984 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:57 crc kubenswrapper[4824]: I0121 11:11:57.364023 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:57 crc kubenswrapper[4824]: I0121 11:11:57.364031 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:57 crc kubenswrapper[4824]: I0121 11:11:57.364044 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:57 crc kubenswrapper[4824]: I0121 11:11:57.364052 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:57Z","lastTransitionTime":"2026-01-21T11:11:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:57 crc kubenswrapper[4824]: I0121 11:11:57.466188 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:57 crc kubenswrapper[4824]: I0121 11:11:57.466217 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:57 crc kubenswrapper[4824]: I0121 11:11:57.466227 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:57 crc kubenswrapper[4824]: I0121 11:11:57.466239 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:57 crc kubenswrapper[4824]: I0121 11:11:57.466246 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:57Z","lastTransitionTime":"2026-01-21T11:11:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:57 crc kubenswrapper[4824]: I0121 11:11:57.568559 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:57 crc kubenswrapper[4824]: I0121 11:11:57.568601 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:57 crc kubenswrapper[4824]: I0121 11:11:57.568610 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:57 crc kubenswrapper[4824]: I0121 11:11:57.568623 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:57 crc kubenswrapper[4824]: I0121 11:11:57.568632 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:57Z","lastTransitionTime":"2026-01-21T11:11:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:57 crc kubenswrapper[4824]: I0121 11:11:57.670691 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:57 crc kubenswrapper[4824]: I0121 11:11:57.670726 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:57 crc kubenswrapper[4824]: I0121 11:11:57.670735 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:57 crc kubenswrapper[4824]: I0121 11:11:57.670747 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:57 crc kubenswrapper[4824]: I0121 11:11:57.670756 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:57Z","lastTransitionTime":"2026-01-21T11:11:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:57 crc kubenswrapper[4824]: I0121 11:11:57.773262 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:57 crc kubenswrapper[4824]: I0121 11:11:57.773297 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:57 crc kubenswrapper[4824]: I0121 11:11:57.773307 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:57 crc kubenswrapper[4824]: I0121 11:11:57.773321 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:57 crc kubenswrapper[4824]: I0121 11:11:57.773329 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:57Z","lastTransitionTime":"2026-01-21T11:11:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:57 crc kubenswrapper[4824]: I0121 11:11:57.875315 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:57 crc kubenswrapper[4824]: I0121 11:11:57.875369 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:57 crc kubenswrapper[4824]: I0121 11:11:57.875377 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:57 crc kubenswrapper[4824]: I0121 11:11:57.875388 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:57 crc kubenswrapper[4824]: I0121 11:11:57.875397 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:57Z","lastTransitionTime":"2026-01-21T11:11:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:57 crc kubenswrapper[4824]: I0121 11:11:57.976942 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:57 crc kubenswrapper[4824]: I0121 11:11:57.976992 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:57 crc kubenswrapper[4824]: I0121 11:11:57.977001 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:57 crc kubenswrapper[4824]: I0121 11:11:57.977016 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:57 crc kubenswrapper[4824]: I0121 11:11:57.977024 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:57Z","lastTransitionTime":"2026-01-21T11:11:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.015075 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-23 04:16:54.091968613 +0000 UTC Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.049237 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9rmnw" Jan 21 11:11:58 crc kubenswrapper[4824]: E0121 11:11:58.049334 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9rmnw" podUID="8f9bd12b-22ee-44ed-a8b9-c100d043f691" Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.049552 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 11:11:58 crc kubenswrapper[4824]: E0121 11:11:58.049809 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.058792 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x94nb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75d52224135b75133fc000611cd179240b9174f6492d3263341716c7e42ccb15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgr8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x94nb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:58Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.066451 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33f3d922-4ffe-409b-a49a-d88c85898260\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a9e57b39eed8300107e944b9fcf6a2ff1e75991f8f37387a3de615b14f9bc78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpwst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b1b37c6d2994021cfc30f85e92c40ae35677113b17aa066d52bc7a76ae5c153\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpwst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zqjg9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:58Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.077780 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gjtnh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a8d2cac-2d0f-4894-9bc0-2abd81699c47\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67c122a26811deaa9b70d19a1a65daee318ca3af2c626712482179d72fe67875\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2239186f1de46481cebfcafb1cab85cd320a205b5a0dfc6dd1b321d9baedeba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b2239186f1de46481cebfcafb1cab85cd320a205b5a0dfc6dd1b321d9baedeba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0db5ecd7759d974ae1b0d8065b3b8bd4d7957fae16eb5bab2d71c2fe69243816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0db5ecd7759d974ae1b0d8065b3b8bd4d7957fae16eb5bab2d71c2fe69243816\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba9b3278c208fc7aac630ca7a0daa3de0823f85d0799d6f1f278da80ae14c862\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba9b3278c208fc7aac630ca7a0daa3de0823f85d0799d6f1f278da80ae14c862\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79b8f6c016b43c3ac3ba7b22cca9a6b72c262cb17ddd6236c5d4970ecc09680c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79b8f6c016b43c3ac3ba7b22cca9a6b72c262cb17ddd6236c5d4970ecc09680c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38c4f218eb2ca7ea0c92bd0ae2ae9efeaa40746751608fbcf7f6c3f8d9e3c758\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38c4f218eb2ca7ea0c92bd0ae2ae9efeaa40746751608fbcf7f6c3f8d9e3c758\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d66ad9831d6029c8c350054525a00069e869d39c1b5638e06e9b772666bd8630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d66ad9831d6029c8c350054525a00069e869d39c1b5638e06e9b772666bd8630\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gjtnh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:58Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.078632 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.078660 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.078669 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.078682 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.078693 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:58Z","lastTransitionTime":"2026-01-21T11:11:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.090343 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-9rmnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f9bd12b-22ee-44ed-a8b9-c100d043f691\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2fgn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2fgn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:28Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-9rmnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:58Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.099560 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86d3005cd61c0973659e1c1c0d3ef1ddd6fcf95bede5465f4f2099e64a74bb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:58Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.107427 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:58Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.119879 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"26b1108a-4b98-4e39-a3cd-e0d055089fd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc319bbc06568cc811ded2b418b04b986a31f321fed492e8be50104af79bb30b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cee9cc9eec3fc815f1621a22a37805691288dbb60dd3ff64194a84be62abcddc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc4c96e33e069f7d08d33caf92cd0987668506f931fd134b246d5ce5e934f95c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ddbb0e5ae02e9d5d82a80ead7efc91f1b8e557a9fb6f39940e5530fae8de2d38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://64a9065254f889b1c43d63759283335bf78fc4fa1531a4baae472c06c075dae3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e94ef8a8f47dc75ecfbab2e3fb8e7097c1b73eed32d7ca947f11ea198cea880\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c655fe698872c3a7804fed5e2e75f300110a816eae4412891f3433762ed0161\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2c655fe698872c3a7804fed5e2e75f300110a816eae4412891f3433762ed0161\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-21T11:11:40Z\\\",\\\"message\\\":\\\"ps://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:40Z is after 2025-08-24T17:21:41Z]\\\\nI0121 11:11:40.696180 6500 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf\\\\nI0121 11:11:40.696191 6500 obj_retry.go:303] Retry object setup: *v1.Pod openshift-kube-apiserver/kube-apiserver-crc\\\\nI0121 11:11:40.696193 6500 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf in node crc\\\\nI0121 11:11:40.696185 6500 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-apiserver-operator/metrics]} name:Service_openshift-apiserver-operator/metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.38:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {8b82f026-5975-4a1b-bb18-08d5d51147ec}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0121 11:11:40.696197 6500 obj_retry.go:365] Adding new object: *v1.Pod openshift-kube-apiserver/kube\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:40Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-htkvk_openshift-ovn-kubernetes(26b1108a-4b98-4e39-a3cd-e0d055089fd8)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948e0497eaffa7b0d84b4085c9ed8d987eb2c6527c9a82c10e576ed9c9c1e21b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-htkvk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:58Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.129507 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-np2v8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4698abe4-0f33-4349-a2a4-614de80ae21b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://191328471bb6f2b34cde2eb927da697619826bf0b6174ecb583c8c25d27a6654\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n9qbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://050bd5bdded2162d5e9dac81713f4ed63cb575d1776e828b2f98e50d89f11b04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n9qbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-np2v8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:58Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.137389 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d87422473838f5218c9fb646ab2515f2f42592754ffdf07b19f7a8e093fa48a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:58Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.145733 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://516c009d4720ab5b393c7482872f12fed22d4f4414bd1f3400411c0e8da88ce4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee18d293036971832c8d9c3f7b9eca84ab779ff6768a2db3c6f7bdcf3433cdef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:58Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.154895 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9bb99c0a-683b-44da-9c0e-e8d284e87d53\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05e076febd697ed802ce5c7ee92a8c2c636a1a2c6609ad7eb01f17dace247c4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43529d1b9e1e5e22b0d56b023946119da8aacbae54b68c26ff0a1a82dc890c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ea232ad87393685aa323ce6393ff3869f8f8d423c2406a501c69c355535adfa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc6c7eff9e17b8a8ecc88bbfe703b5a01dbf1e02d342cede4f6a263a6526af43\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:58Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.162707 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e570fff1-6227-44d6-b71f-ee377b1099ae\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4d41492c3802a8168fffd249c15aeb9c902a5c7e01550e1aa6f9b769adfdc4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2690a1ecbc8fe1290e68bba2695a3edd833b5da933fb71e8ddc926a1340a9a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b85db961e1e078e1f54323c79a99c9413365256ac2fce411db7932bd0580ab6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://595783a1454735bd33f4862180b60d99a4e1b6fae1e735f2b7328391951bf422\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://595783a1454735bd33f4862180b60d99a4e1b6fae1e735f2b7328391951bf422\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:58Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.170863 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:58Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.177688 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lv4rn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1267539-9adf-466f-8027-8ede009d19a0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eed1b884aa53e1ee318169a0cc2c9ad10130aa363ae4ac55e4bf065237756f53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbmfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lv4rn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:58Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.180284 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.180336 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.180346 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.180371 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.180380 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:58Z","lastTransitionTime":"2026-01-21T11:11:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.185057 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jr64s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9768c60d-5bee-4286-bced-076e584b04e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91158f14d2e98ac9f875a5e7101df3107ef28aa5fc5d62218f460af98ff920fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flrnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:17Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jr64s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:58Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.194062 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d31ab61-f94c-4f49-85d0-c9e91b83069b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a66b5c33884af00408dc35cf02f08ebd8570bd695f9f1fe558bf54fd06930431\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7514786602746dabfa690da48d48e233fe68d6dd6674aa5ecd489c0b3185f08\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61b14b6846f94cdeba8f1f61b3754e7663041f92c70356c4ee6cbb988a2e680f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a7023cfe4c098bcdd9c0477a8187f4ed7a0d2abdbe1e143436863948f441b4f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af11ad95bcf0c6bbd5fa4ab99bb592ce1c87f40d29c5ade7832414315820a569\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW0121 11:11:14.604290 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0121 11:11:14.604572 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0121 11:11:14.605875 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3284452394/tls.crt::/tmp/serving-cert-3284452394/tls.key\\\\\\\"\\\\nI0121 11:11:14.923427 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0121 11:11:14.928141 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0121 11:11:14.928162 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0121 11:11:14.928179 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0121 11:11:14.928184 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0121 11:11:14.933154 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0121 11:11:14.933230 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 11:11:14.933259 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 11:11:14.933282 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0121 11:11:14.933299 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0121 11:11:14.933315 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0121 11:11:14.933331 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0121 11:11:14.933177 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0121 11:11:14.934666 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48070c09fad511108ff22e44211cf26911509b8406d66948985cbfc3009d8233\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:58Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.206661 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06f56d86-ec6f-4b57-b314-f508c0dd4d9f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d25c94f1ae0bbabb733f0e1ea1d1ba39562a92af73047a39aacc03013e239681\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca06234486b59cf990c69d9501e23f935328e2ba21013cd71ed64831ea749bcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b63e45d66bcd11ff3479cc3ec6b450017da8d21d36e967b8d496ae0f24b0ee9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52758bcd31f2ba64edf03adfd4d1be6428015abff5b64c7c029fa787085f9c29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2736e73342d2adb175b0721dd4f40cc42d26fca06d5eb3334927518531c3b107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b35ad7dd0570c191eb0080739304f3cd43f17b3d1dddf895549441ebe50deaa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b35ad7dd0570c191eb0080739304f3cd43f17b3d1dddf895549441ebe50deaa6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://80cd5ea5e6fc19a88c1aab5c61f827f103ab798a5c84da8dc4d45cea86d5aacb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80cd5ea5e6fc19a88c1aab5c61f827f103ab798a5c84da8dc4d45cea86d5aacb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://234d8b914e6a88480021647ae65d74e9b21c55b6788312a094d00c8a42677d48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://234d8b914e6a88480021647ae65d74e9b21c55b6788312a094d00c8a42677d48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:58Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.214620 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:58Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.282092 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.282128 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.282140 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.282165 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.282177 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:58Z","lastTransitionTime":"2026-01-21T11:11:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.383982 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.384116 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.384182 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.384250 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.384307 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:58Z","lastTransitionTime":"2026-01-21T11:11:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.486204 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.486237 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.486245 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.486257 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.486264 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:58Z","lastTransitionTime":"2026-01-21T11:11:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.588066 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.588104 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.588113 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.588126 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.588135 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:58Z","lastTransitionTime":"2026-01-21T11:11:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.590639 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.590669 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.590677 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.590690 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.590698 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:58Z","lastTransitionTime":"2026-01-21T11:11:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:58 crc kubenswrapper[4824]: E0121 11:11:58.600209 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:58Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:58Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"695c676c-720c-4921-9e71-56b96f2fb46b\\\",\\\"systemUUID\\\":\\\"d00624f5-786f-41dc-bdc1-a1875f167d65\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:58Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.602438 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.602460 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.602467 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.602477 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.602484 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:58Z","lastTransitionTime":"2026-01-21T11:11:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:58 crc kubenswrapper[4824]: E0121 11:11:58.611275 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:58Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:58Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"695c676c-720c-4921-9e71-56b96f2fb46b\\\",\\\"systemUUID\\\":\\\"d00624f5-786f-41dc-bdc1-a1875f167d65\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:58Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.614000 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.614092 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.614158 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.614220 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.614284 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:58Z","lastTransitionTime":"2026-01-21T11:11:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:58 crc kubenswrapper[4824]: E0121 11:11:58.622421 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:58Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:58Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"695c676c-720c-4921-9e71-56b96f2fb46b\\\",\\\"systemUUID\\\":\\\"d00624f5-786f-41dc-bdc1-a1875f167d65\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:58Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.624706 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.624784 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.624793 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.624804 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.624811 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:58Z","lastTransitionTime":"2026-01-21T11:11:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:58 crc kubenswrapper[4824]: E0121 11:11:58.633265 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:58Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:58Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"695c676c-720c-4921-9e71-56b96f2fb46b\\\",\\\"systemUUID\\\":\\\"d00624f5-786f-41dc-bdc1-a1875f167d65\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:58Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.635292 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.635417 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.635478 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.635542 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.635599 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:58Z","lastTransitionTime":"2026-01-21T11:11:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:58 crc kubenswrapper[4824]: E0121 11:11:58.643433 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:58Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:11:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:58Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"695c676c-720c-4921-9e71-56b96f2fb46b\\\",\\\"systemUUID\\\":\\\"d00624f5-786f-41dc-bdc1-a1875f167d65\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:58Z is after 2025-08-24T17:21:41Z" Jan 21 11:11:58 crc kubenswrapper[4824]: E0121 11:11:58.643927 4824 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.690254 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.690303 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.690314 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.690328 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.690336 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:58Z","lastTransitionTime":"2026-01-21T11:11:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.792110 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.792166 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.792176 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.792210 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.792223 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:58Z","lastTransitionTime":"2026-01-21T11:11:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.895051 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.895106 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.895116 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.895131 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.895139 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:58Z","lastTransitionTime":"2026-01-21T11:11:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.996892 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.996934 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.996943 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.996972 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:58 crc kubenswrapper[4824]: I0121 11:11:58.996982 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:58Z","lastTransitionTime":"2026-01-21T11:11:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:59 crc kubenswrapper[4824]: I0121 11:11:59.015599 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-14 04:06:58.95960934 +0000 UTC Jan 21 11:11:59 crc kubenswrapper[4824]: I0121 11:11:59.049038 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 11:11:59 crc kubenswrapper[4824]: I0121 11:11:59.049078 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 11:11:59 crc kubenswrapper[4824]: E0121 11:11:59.049344 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 21 11:11:59 crc kubenswrapper[4824]: E0121 11:11:59.049233 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 21 11:11:59 crc kubenswrapper[4824]: I0121 11:11:59.099016 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:59 crc kubenswrapper[4824]: I0121 11:11:59.099053 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:59 crc kubenswrapper[4824]: I0121 11:11:59.099063 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:59 crc kubenswrapper[4824]: I0121 11:11:59.099078 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:59 crc kubenswrapper[4824]: I0121 11:11:59.099087 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:59Z","lastTransitionTime":"2026-01-21T11:11:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:59 crc kubenswrapper[4824]: I0121 11:11:59.201103 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:59 crc kubenswrapper[4824]: I0121 11:11:59.201286 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:59 crc kubenswrapper[4824]: I0121 11:11:59.201374 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:59 crc kubenswrapper[4824]: I0121 11:11:59.201453 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:59 crc kubenswrapper[4824]: I0121 11:11:59.201520 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:59Z","lastTransitionTime":"2026-01-21T11:11:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:59 crc kubenswrapper[4824]: I0121 11:11:59.303898 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:59 crc kubenswrapper[4824]: I0121 11:11:59.304144 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:59 crc kubenswrapper[4824]: I0121 11:11:59.304206 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:59 crc kubenswrapper[4824]: I0121 11:11:59.304270 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:59 crc kubenswrapper[4824]: I0121 11:11:59.304320 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:59Z","lastTransitionTime":"2026-01-21T11:11:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:59 crc kubenswrapper[4824]: I0121 11:11:59.406665 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:59 crc kubenswrapper[4824]: I0121 11:11:59.406742 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:59 crc kubenswrapper[4824]: I0121 11:11:59.406754 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:59 crc kubenswrapper[4824]: I0121 11:11:59.406779 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:59 crc kubenswrapper[4824]: I0121 11:11:59.406793 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:59Z","lastTransitionTime":"2026-01-21T11:11:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:59 crc kubenswrapper[4824]: I0121 11:11:59.509467 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:59 crc kubenswrapper[4824]: I0121 11:11:59.509526 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:59 crc kubenswrapper[4824]: I0121 11:11:59.509541 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:59 crc kubenswrapper[4824]: I0121 11:11:59.509565 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:59 crc kubenswrapper[4824]: I0121 11:11:59.509577 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:59Z","lastTransitionTime":"2026-01-21T11:11:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:59 crc kubenswrapper[4824]: I0121 11:11:59.611567 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:59 crc kubenswrapper[4824]: I0121 11:11:59.611611 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:59 crc kubenswrapper[4824]: I0121 11:11:59.611620 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:59 crc kubenswrapper[4824]: I0121 11:11:59.611634 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:59 crc kubenswrapper[4824]: I0121 11:11:59.611642 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:59Z","lastTransitionTime":"2026-01-21T11:11:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:59 crc kubenswrapper[4824]: I0121 11:11:59.713702 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:59 crc kubenswrapper[4824]: I0121 11:11:59.713741 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:59 crc kubenswrapper[4824]: I0121 11:11:59.713749 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:59 crc kubenswrapper[4824]: I0121 11:11:59.713763 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:59 crc kubenswrapper[4824]: I0121 11:11:59.713774 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:59Z","lastTransitionTime":"2026-01-21T11:11:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:59 crc kubenswrapper[4824]: I0121 11:11:59.816061 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:59 crc kubenswrapper[4824]: I0121 11:11:59.816108 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:59 crc kubenswrapper[4824]: I0121 11:11:59.816118 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:59 crc kubenswrapper[4824]: I0121 11:11:59.816133 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:59 crc kubenswrapper[4824]: I0121 11:11:59.816142 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:59Z","lastTransitionTime":"2026-01-21T11:11:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:11:59 crc kubenswrapper[4824]: I0121 11:11:59.917782 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:11:59 crc kubenswrapper[4824]: I0121 11:11:59.917816 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:11:59 crc kubenswrapper[4824]: I0121 11:11:59.917827 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:11:59 crc kubenswrapper[4824]: I0121 11:11:59.917843 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:11:59 crc kubenswrapper[4824]: I0121 11:11:59.917852 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:11:59Z","lastTransitionTime":"2026-01-21T11:11:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:00 crc kubenswrapper[4824]: I0121 11:12:00.016432 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-10 01:48:51.567181656 +0000 UTC Jan 21 11:12:00 crc kubenswrapper[4824]: I0121 11:12:00.020003 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:00 crc kubenswrapper[4824]: I0121 11:12:00.020046 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:00 crc kubenswrapper[4824]: I0121 11:12:00.020058 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:00 crc kubenswrapper[4824]: I0121 11:12:00.020076 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:00 crc kubenswrapper[4824]: I0121 11:12:00.020096 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:00Z","lastTransitionTime":"2026-01-21T11:12:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:00 crc kubenswrapper[4824]: I0121 11:12:00.049115 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9rmnw" Jan 21 11:12:00 crc kubenswrapper[4824]: E0121 11:12:00.049220 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9rmnw" podUID="8f9bd12b-22ee-44ed-a8b9-c100d043f691" Jan 21 11:12:00 crc kubenswrapper[4824]: I0121 11:12:00.049253 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 11:12:00 crc kubenswrapper[4824]: E0121 11:12:00.049316 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 21 11:12:00 crc kubenswrapper[4824]: I0121 11:12:00.104415 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8f9bd12b-22ee-44ed-a8b9-c100d043f691-metrics-certs\") pod \"network-metrics-daemon-9rmnw\" (UID: \"8f9bd12b-22ee-44ed-a8b9-c100d043f691\") " pod="openshift-multus/network-metrics-daemon-9rmnw" Jan 21 11:12:00 crc kubenswrapper[4824]: E0121 11:12:00.104508 4824 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Jan 21 11:12:00 crc kubenswrapper[4824]: E0121 11:12:00.104558 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8f9bd12b-22ee-44ed-a8b9-c100d043f691-metrics-certs podName:8f9bd12b-22ee-44ed-a8b9-c100d043f691 nodeName:}" failed. No retries permitted until 2026-01-21 11:12:32.104543511 +0000 UTC m=+94.397572813 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/8f9bd12b-22ee-44ed-a8b9-c100d043f691-metrics-certs") pod "network-metrics-daemon-9rmnw" (UID: "8f9bd12b-22ee-44ed-a8b9-c100d043f691") : object "openshift-multus"/"metrics-daemon-secret" not registered Jan 21 11:12:00 crc kubenswrapper[4824]: I0121 11:12:00.121747 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:00 crc kubenswrapper[4824]: I0121 11:12:00.121775 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:00 crc kubenswrapper[4824]: I0121 11:12:00.121786 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:00 crc kubenswrapper[4824]: I0121 11:12:00.121798 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:00 crc kubenswrapper[4824]: I0121 11:12:00.121807 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:00Z","lastTransitionTime":"2026-01-21T11:12:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:00 crc kubenswrapper[4824]: I0121 11:12:00.224309 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:00 crc kubenswrapper[4824]: I0121 11:12:00.224353 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:00 crc kubenswrapper[4824]: I0121 11:12:00.224365 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:00 crc kubenswrapper[4824]: I0121 11:12:00.224395 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:00 crc kubenswrapper[4824]: I0121 11:12:00.224405 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:00Z","lastTransitionTime":"2026-01-21T11:12:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:00 crc kubenswrapper[4824]: I0121 11:12:00.326599 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:00 crc kubenswrapper[4824]: I0121 11:12:00.326643 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:00 crc kubenswrapper[4824]: I0121 11:12:00.326653 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:00 crc kubenswrapper[4824]: I0121 11:12:00.326669 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:00 crc kubenswrapper[4824]: I0121 11:12:00.326683 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:00Z","lastTransitionTime":"2026-01-21T11:12:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:00 crc kubenswrapper[4824]: I0121 11:12:00.429192 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:00 crc kubenswrapper[4824]: I0121 11:12:00.429242 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:00 crc kubenswrapper[4824]: I0121 11:12:00.429252 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:00 crc kubenswrapper[4824]: I0121 11:12:00.429266 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:00 crc kubenswrapper[4824]: I0121 11:12:00.429301 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:00Z","lastTransitionTime":"2026-01-21T11:12:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:00 crc kubenswrapper[4824]: I0121 11:12:00.531751 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:00 crc kubenswrapper[4824]: I0121 11:12:00.531786 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:00 crc kubenswrapper[4824]: I0121 11:12:00.531793 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:00 crc kubenswrapper[4824]: I0121 11:12:00.531804 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:00 crc kubenswrapper[4824]: I0121 11:12:00.531811 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:00Z","lastTransitionTime":"2026-01-21T11:12:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:00 crc kubenswrapper[4824]: I0121 11:12:00.633017 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:00 crc kubenswrapper[4824]: I0121 11:12:00.633051 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:00 crc kubenswrapper[4824]: I0121 11:12:00.633062 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:00 crc kubenswrapper[4824]: I0121 11:12:00.633074 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:00 crc kubenswrapper[4824]: I0121 11:12:00.633082 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:00Z","lastTransitionTime":"2026-01-21T11:12:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:00 crc kubenswrapper[4824]: I0121 11:12:00.735028 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:00 crc kubenswrapper[4824]: I0121 11:12:00.735050 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:00 crc kubenswrapper[4824]: I0121 11:12:00.735059 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:00 crc kubenswrapper[4824]: I0121 11:12:00.735070 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:00 crc kubenswrapper[4824]: I0121 11:12:00.735077 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:00Z","lastTransitionTime":"2026-01-21T11:12:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:00 crc kubenswrapper[4824]: I0121 11:12:00.836798 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:00 crc kubenswrapper[4824]: I0121 11:12:00.836840 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:00 crc kubenswrapper[4824]: I0121 11:12:00.836848 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:00 crc kubenswrapper[4824]: I0121 11:12:00.836862 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:00 crc kubenswrapper[4824]: I0121 11:12:00.836870 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:00Z","lastTransitionTime":"2026-01-21T11:12:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:00 crc kubenswrapper[4824]: I0121 11:12:00.939220 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:00 crc kubenswrapper[4824]: I0121 11:12:00.939264 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:00 crc kubenswrapper[4824]: I0121 11:12:00.939274 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:00 crc kubenswrapper[4824]: I0121 11:12:00.939289 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:00 crc kubenswrapper[4824]: I0121 11:12:00.939301 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:00Z","lastTransitionTime":"2026-01-21T11:12:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:01 crc kubenswrapper[4824]: I0121 11:12:01.017236 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-07 02:45:04.660901158 +0000 UTC Jan 21 11:12:01 crc kubenswrapper[4824]: I0121 11:12:01.041317 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:01 crc kubenswrapper[4824]: I0121 11:12:01.041475 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:01 crc kubenswrapper[4824]: I0121 11:12:01.041536 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:01 crc kubenswrapper[4824]: I0121 11:12:01.041601 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:01 crc kubenswrapper[4824]: I0121 11:12:01.041671 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:01Z","lastTransitionTime":"2026-01-21T11:12:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:01 crc kubenswrapper[4824]: I0121 11:12:01.048680 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 11:12:01 crc kubenswrapper[4824]: I0121 11:12:01.048841 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 11:12:01 crc kubenswrapper[4824]: E0121 11:12:01.048862 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 21 11:12:01 crc kubenswrapper[4824]: E0121 11:12:01.049075 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 21 11:12:01 crc kubenswrapper[4824]: I0121 11:12:01.143641 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:01 crc kubenswrapper[4824]: I0121 11:12:01.143686 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:01 crc kubenswrapper[4824]: I0121 11:12:01.143695 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:01 crc kubenswrapper[4824]: I0121 11:12:01.143711 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:01 crc kubenswrapper[4824]: I0121 11:12:01.143720 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:01Z","lastTransitionTime":"2026-01-21T11:12:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:01 crc kubenswrapper[4824]: I0121 11:12:01.245450 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:01 crc kubenswrapper[4824]: I0121 11:12:01.245497 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:01 crc kubenswrapper[4824]: I0121 11:12:01.245507 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:01 crc kubenswrapper[4824]: I0121 11:12:01.245522 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:01 crc kubenswrapper[4824]: I0121 11:12:01.245531 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:01Z","lastTransitionTime":"2026-01-21T11:12:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:01 crc kubenswrapper[4824]: I0121 11:12:01.347879 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:01 crc kubenswrapper[4824]: I0121 11:12:01.347939 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:01 crc kubenswrapper[4824]: I0121 11:12:01.347950 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:01 crc kubenswrapper[4824]: I0121 11:12:01.347991 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:01 crc kubenswrapper[4824]: I0121 11:12:01.348005 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:01Z","lastTransitionTime":"2026-01-21T11:12:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:01 crc kubenswrapper[4824]: I0121 11:12:01.450633 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:01 crc kubenswrapper[4824]: I0121 11:12:01.450705 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:01 crc kubenswrapper[4824]: I0121 11:12:01.450718 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:01 crc kubenswrapper[4824]: I0121 11:12:01.450768 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:01 crc kubenswrapper[4824]: I0121 11:12:01.450783 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:01Z","lastTransitionTime":"2026-01-21T11:12:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:01 crc kubenswrapper[4824]: I0121 11:12:01.555809 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:01 crc kubenswrapper[4824]: I0121 11:12:01.555936 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:01 crc kubenswrapper[4824]: I0121 11:12:01.556051 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:01 crc kubenswrapper[4824]: I0121 11:12:01.556129 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:01 crc kubenswrapper[4824]: I0121 11:12:01.556188 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:01Z","lastTransitionTime":"2026-01-21T11:12:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:01 crc kubenswrapper[4824]: I0121 11:12:01.658093 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:01 crc kubenswrapper[4824]: I0121 11:12:01.658206 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:01 crc kubenswrapper[4824]: I0121 11:12:01.658271 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:01 crc kubenswrapper[4824]: I0121 11:12:01.658331 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:01 crc kubenswrapper[4824]: I0121 11:12:01.658411 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:01Z","lastTransitionTime":"2026-01-21T11:12:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:01 crc kubenswrapper[4824]: I0121 11:12:01.760235 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:01 crc kubenswrapper[4824]: I0121 11:12:01.760296 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:01 crc kubenswrapper[4824]: I0121 11:12:01.760306 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:01 crc kubenswrapper[4824]: I0121 11:12:01.760325 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:01 crc kubenswrapper[4824]: I0121 11:12:01.760336 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:01Z","lastTransitionTime":"2026-01-21T11:12:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:01 crc kubenswrapper[4824]: I0121 11:12:01.862711 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:01 crc kubenswrapper[4824]: I0121 11:12:01.862774 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:01 crc kubenswrapper[4824]: I0121 11:12:01.862787 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:01 crc kubenswrapper[4824]: I0121 11:12:01.862814 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:01 crc kubenswrapper[4824]: I0121 11:12:01.862827 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:01Z","lastTransitionTime":"2026-01-21T11:12:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:01 crc kubenswrapper[4824]: I0121 11:12:01.964913 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:01 crc kubenswrapper[4824]: I0121 11:12:01.964977 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:01 crc kubenswrapper[4824]: I0121 11:12:01.964989 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:01 crc kubenswrapper[4824]: I0121 11:12:01.965006 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:01 crc kubenswrapper[4824]: I0121 11:12:01.965018 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:01Z","lastTransitionTime":"2026-01-21T11:12:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:02 crc kubenswrapper[4824]: I0121 11:12:02.018277 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-11 22:51:06.852216222 +0000 UTC Jan 21 11:12:02 crc kubenswrapper[4824]: I0121 11:12:02.048831 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 11:12:02 crc kubenswrapper[4824]: I0121 11:12:02.048862 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9rmnw" Jan 21 11:12:02 crc kubenswrapper[4824]: E0121 11:12:02.048999 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 21 11:12:02 crc kubenswrapper[4824]: E0121 11:12:02.049137 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9rmnw" podUID="8f9bd12b-22ee-44ed-a8b9-c100d043f691" Jan 21 11:12:02 crc kubenswrapper[4824]: I0121 11:12:02.067080 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:02 crc kubenswrapper[4824]: I0121 11:12:02.067116 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:02 crc kubenswrapper[4824]: I0121 11:12:02.067127 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:02 crc kubenswrapper[4824]: I0121 11:12:02.067141 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:02 crc kubenswrapper[4824]: I0121 11:12:02.067152 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:02Z","lastTransitionTime":"2026-01-21T11:12:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:02 crc kubenswrapper[4824]: I0121 11:12:02.169796 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:02 crc kubenswrapper[4824]: I0121 11:12:02.169860 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:02 crc kubenswrapper[4824]: I0121 11:12:02.169874 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:02 crc kubenswrapper[4824]: I0121 11:12:02.169901 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:02 crc kubenswrapper[4824]: I0121 11:12:02.169917 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:02Z","lastTransitionTime":"2026-01-21T11:12:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:02 crc kubenswrapper[4824]: I0121 11:12:02.272084 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:02 crc kubenswrapper[4824]: I0121 11:12:02.272117 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:02 crc kubenswrapper[4824]: I0121 11:12:02.272126 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:02 crc kubenswrapper[4824]: I0121 11:12:02.272137 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:02 crc kubenswrapper[4824]: I0121 11:12:02.272146 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:02Z","lastTransitionTime":"2026-01-21T11:12:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:02 crc kubenswrapper[4824]: I0121 11:12:02.373967 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:02 crc kubenswrapper[4824]: I0121 11:12:02.374001 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:02 crc kubenswrapper[4824]: I0121 11:12:02.374011 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:02 crc kubenswrapper[4824]: I0121 11:12:02.374022 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:02 crc kubenswrapper[4824]: I0121 11:12:02.374031 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:02Z","lastTransitionTime":"2026-01-21T11:12:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:02 crc kubenswrapper[4824]: I0121 11:12:02.475194 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:02 crc kubenswrapper[4824]: I0121 11:12:02.475223 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:02 crc kubenswrapper[4824]: I0121 11:12:02.475232 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:02 crc kubenswrapper[4824]: I0121 11:12:02.475242 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:02 crc kubenswrapper[4824]: I0121 11:12:02.475250 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:02Z","lastTransitionTime":"2026-01-21T11:12:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:02 crc kubenswrapper[4824]: I0121 11:12:02.577336 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:02 crc kubenswrapper[4824]: I0121 11:12:02.577443 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:02 crc kubenswrapper[4824]: I0121 11:12:02.577458 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:02 crc kubenswrapper[4824]: I0121 11:12:02.577482 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:02 crc kubenswrapper[4824]: I0121 11:12:02.577494 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:02Z","lastTransitionTime":"2026-01-21T11:12:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:02 crc kubenswrapper[4824]: I0121 11:12:02.679000 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:02 crc kubenswrapper[4824]: I0121 11:12:02.679033 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:02 crc kubenswrapper[4824]: I0121 11:12:02.679043 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:02 crc kubenswrapper[4824]: I0121 11:12:02.679054 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:02 crc kubenswrapper[4824]: I0121 11:12:02.679062 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:02Z","lastTransitionTime":"2026-01-21T11:12:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:02 crc kubenswrapper[4824]: I0121 11:12:02.780554 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:02 crc kubenswrapper[4824]: I0121 11:12:02.780651 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:02 crc kubenswrapper[4824]: I0121 11:12:02.780724 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:02 crc kubenswrapper[4824]: I0121 11:12:02.780791 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:02 crc kubenswrapper[4824]: I0121 11:12:02.780846 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:02Z","lastTransitionTime":"2026-01-21T11:12:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:02 crc kubenswrapper[4824]: I0121 11:12:02.882403 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:02 crc kubenswrapper[4824]: I0121 11:12:02.882436 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:02 crc kubenswrapper[4824]: I0121 11:12:02.882447 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:02 crc kubenswrapper[4824]: I0121 11:12:02.882459 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:02 crc kubenswrapper[4824]: I0121 11:12:02.882471 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:02Z","lastTransitionTime":"2026-01-21T11:12:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:02 crc kubenswrapper[4824]: I0121 11:12:02.984590 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:02 crc kubenswrapper[4824]: I0121 11:12:02.984618 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:02 crc kubenswrapper[4824]: I0121 11:12:02.984626 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:02 crc kubenswrapper[4824]: I0121 11:12:02.984637 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:02 crc kubenswrapper[4824]: I0121 11:12:02.984645 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:02Z","lastTransitionTime":"2026-01-21T11:12:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:03 crc kubenswrapper[4824]: I0121 11:12:03.018856 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-26 17:33:40.855594493 +0000 UTC Jan 21 11:12:03 crc kubenswrapper[4824]: I0121 11:12:03.048848 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 11:12:03 crc kubenswrapper[4824]: I0121 11:12:03.048863 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 11:12:03 crc kubenswrapper[4824]: E0121 11:12:03.048946 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 21 11:12:03 crc kubenswrapper[4824]: E0121 11:12:03.049037 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 21 11:12:03 crc kubenswrapper[4824]: I0121 11:12:03.086109 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:03 crc kubenswrapper[4824]: I0121 11:12:03.086134 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:03 crc kubenswrapper[4824]: I0121 11:12:03.086143 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:03 crc kubenswrapper[4824]: I0121 11:12:03.086154 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:03 crc kubenswrapper[4824]: I0121 11:12:03.086163 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:03Z","lastTransitionTime":"2026-01-21T11:12:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:03 crc kubenswrapper[4824]: I0121 11:12:03.188077 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:03 crc kubenswrapper[4824]: I0121 11:12:03.188103 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:03 crc kubenswrapper[4824]: I0121 11:12:03.188114 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:03 crc kubenswrapper[4824]: I0121 11:12:03.188124 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:03 crc kubenswrapper[4824]: I0121 11:12:03.188133 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:03Z","lastTransitionTime":"2026-01-21T11:12:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:03 crc kubenswrapper[4824]: I0121 11:12:03.289682 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:03 crc kubenswrapper[4824]: I0121 11:12:03.289713 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:03 crc kubenswrapper[4824]: I0121 11:12:03.289723 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:03 crc kubenswrapper[4824]: I0121 11:12:03.289739 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:03 crc kubenswrapper[4824]: I0121 11:12:03.289751 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:03Z","lastTransitionTime":"2026-01-21T11:12:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:03 crc kubenswrapper[4824]: I0121 11:12:03.296601 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-x94nb_6b9588d1-5876-4c2c-8de5-e5c0ad46f04e/kube-multus/0.log" Jan 21 11:12:03 crc kubenswrapper[4824]: I0121 11:12:03.296637 4824 generic.go:334] "Generic (PLEG): container finished" podID="6b9588d1-5876-4c2c-8de5-e5c0ad46f04e" containerID="75d52224135b75133fc000611cd179240b9174f6492d3263341716c7e42ccb15" exitCode=1 Jan 21 11:12:03 crc kubenswrapper[4824]: I0121 11:12:03.296666 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-x94nb" event={"ID":"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e","Type":"ContainerDied","Data":"75d52224135b75133fc000611cd179240b9174f6492d3263341716c7e42ccb15"} Jan 21 11:12:03 crc kubenswrapper[4824]: I0121 11:12:03.296944 4824 scope.go:117] "RemoveContainer" containerID="75d52224135b75133fc000611cd179240b9174f6492d3263341716c7e42ccb15" Jan 21 11:12:03 crc kubenswrapper[4824]: I0121 11:12:03.307030 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jr64s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9768c60d-5bee-4286-bced-076e584b04e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91158f14d2e98ac9f875a5e7101df3107ef28aa5fc5d62218f460af98ff920fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flrnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:17Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jr64s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:03Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:03 crc kubenswrapper[4824]: I0121 11:12:03.321752 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d31ab61-f94c-4f49-85d0-c9e91b83069b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a66b5c33884af00408dc35cf02f08ebd8570bd695f9f1fe558bf54fd06930431\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7514786602746dabfa690da48d48e233fe68d6dd6674aa5ecd489c0b3185f08\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61b14b6846f94cdeba8f1f61b3754e7663041f92c70356c4ee6cbb988a2e680f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a7023cfe4c098bcdd9c0477a8187f4ed7a0d2abdbe1e143436863948f441b4f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af11ad95bcf0c6bbd5fa4ab99bb592ce1c87f40d29c5ade7832414315820a569\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW0121 11:11:14.604290 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0121 11:11:14.604572 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0121 11:11:14.605875 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3284452394/tls.crt::/tmp/serving-cert-3284452394/tls.key\\\\\\\"\\\\nI0121 11:11:14.923427 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0121 11:11:14.928141 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0121 11:11:14.928162 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0121 11:11:14.928179 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0121 11:11:14.928184 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0121 11:11:14.933154 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0121 11:11:14.933230 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 11:11:14.933259 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 11:11:14.933282 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0121 11:11:14.933299 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0121 11:11:14.933315 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0121 11:11:14.933331 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0121 11:11:14.933177 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0121 11:11:14.934666 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48070c09fad511108ff22e44211cf26911509b8406d66948985cbfc3009d8233\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:03Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:03 crc kubenswrapper[4824]: I0121 11:12:03.336058 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9bb99c0a-683b-44da-9c0e-e8d284e87d53\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05e076febd697ed802ce5c7ee92a8c2c636a1a2c6609ad7eb01f17dace247c4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43529d1b9e1e5e22b0d56b023946119da8aacbae54b68c26ff0a1a82dc890c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ea232ad87393685aa323ce6393ff3869f8f8d423c2406a501c69c355535adfa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc6c7eff9e17b8a8ecc88bbfe703b5a01dbf1e02d342cede4f6a263a6526af43\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:03Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:03 crc kubenswrapper[4824]: I0121 11:12:03.346591 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e570fff1-6227-44d6-b71f-ee377b1099ae\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4d41492c3802a8168fffd249c15aeb9c902a5c7e01550e1aa6f9b769adfdc4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2690a1ecbc8fe1290e68bba2695a3edd833b5da933fb71e8ddc926a1340a9a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b85db961e1e078e1f54323c79a99c9413365256ac2fce411db7932bd0580ab6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://595783a1454735bd33f4862180b60d99a4e1b6fae1e735f2b7328391951bf422\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://595783a1454735bd33f4862180b60d99a4e1b6fae1e735f2b7328391951bf422\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:03Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:03 crc kubenswrapper[4824]: I0121 11:12:03.357517 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:03Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:03 crc kubenswrapper[4824]: I0121 11:12:03.365444 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lv4rn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1267539-9adf-466f-8027-8ede009d19a0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eed1b884aa53e1ee318169a0cc2c9ad10130aa363ae4ac55e4bf065237756f53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbmfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lv4rn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:03Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:03 crc kubenswrapper[4824]: I0121 11:12:03.378922 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06f56d86-ec6f-4b57-b314-f508c0dd4d9f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d25c94f1ae0bbabb733f0e1ea1d1ba39562a92af73047a39aacc03013e239681\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca06234486b59cf990c69d9501e23f935328e2ba21013cd71ed64831ea749bcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b63e45d66bcd11ff3479cc3ec6b450017da8d21d36e967b8d496ae0f24b0ee9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52758bcd31f2ba64edf03adfd4d1be6428015abff5b64c7c029fa787085f9c29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2736e73342d2adb175b0721dd4f40cc42d26fca06d5eb3334927518531c3b107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b35ad7dd0570c191eb0080739304f3cd43f17b3d1dddf895549441ebe50deaa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b35ad7dd0570c191eb0080739304f3cd43f17b3d1dddf895549441ebe50deaa6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://80cd5ea5e6fc19a88c1aab5c61f827f103ab798a5c84da8dc4d45cea86d5aacb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80cd5ea5e6fc19a88c1aab5c61f827f103ab798a5c84da8dc4d45cea86d5aacb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://234d8b914e6a88480021647ae65d74e9b21c55b6788312a094d00c8a42677d48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://234d8b914e6a88480021647ae65d74e9b21c55b6788312a094d00c8a42677d48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:03Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:03 crc kubenswrapper[4824]: I0121 11:12:03.388567 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:03Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:03 crc kubenswrapper[4824]: I0121 11:12:03.391926 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:03 crc kubenswrapper[4824]: I0121 11:12:03.391979 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:03 crc kubenswrapper[4824]: I0121 11:12:03.391991 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:03 crc kubenswrapper[4824]: I0121 11:12:03.392010 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:03 crc kubenswrapper[4824]: I0121 11:12:03.392024 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:03Z","lastTransitionTime":"2026-01-21T11:12:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:03 crc kubenswrapper[4824]: I0121 11:12:03.398570 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86d3005cd61c0973659e1c1c0d3ef1ddd6fcf95bede5465f4f2099e64a74bb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:03Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:03 crc kubenswrapper[4824]: I0121 11:12:03.408676 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:03Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:03 crc kubenswrapper[4824]: I0121 11:12:03.418424 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x94nb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:12:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:12:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75d52224135b75133fc000611cd179240b9174f6492d3263341716c7e42ccb15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75d52224135b75133fc000611cd179240b9174f6492d3263341716c7e42ccb15\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-21T11:12:02Z\\\",\\\"message\\\":\\\"2026-01-21T11:11:17+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_c9e2a0c0-e545-4ddc-b22f-97601d5900ca\\\\n2026-01-21T11:11:17+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_c9e2a0c0-e545-4ddc-b22f-97601d5900ca to /host/opt/cni/bin/\\\\n2026-01-21T11:11:17Z [verbose] multus-daemon started\\\\n2026-01-21T11:11:17Z [verbose] Readiness Indicator file check\\\\n2026-01-21T11:12:02Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgr8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x94nb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:03Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:03 crc kubenswrapper[4824]: I0121 11:12:03.426673 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33f3d922-4ffe-409b-a49a-d88c85898260\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a9e57b39eed8300107e944b9fcf6a2ff1e75991f8f37387a3de615b14f9bc78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpwst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b1b37c6d2994021cfc30f85e92c40ae35677113b17aa066d52bc7a76ae5c153\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpwst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zqjg9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:03Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:03 crc kubenswrapper[4824]: I0121 11:12:03.436921 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gjtnh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a8d2cac-2d0f-4894-9bc0-2abd81699c47\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67c122a26811deaa9b70d19a1a65daee318ca3af2c626712482179d72fe67875\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2239186f1de46481cebfcafb1cab85cd320a205b5a0dfc6dd1b321d9baedeba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b2239186f1de46481cebfcafb1cab85cd320a205b5a0dfc6dd1b321d9baedeba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0db5ecd7759d974ae1b0d8065b3b8bd4d7957fae16eb5bab2d71c2fe69243816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0db5ecd7759d974ae1b0d8065b3b8bd4d7957fae16eb5bab2d71c2fe69243816\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba9b3278c208fc7aac630ca7a0daa3de0823f85d0799d6f1f278da80ae14c862\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba9b3278c208fc7aac630ca7a0daa3de0823f85d0799d6f1f278da80ae14c862\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79b8f6c016b43c3ac3ba7b22cca9a6b72c262cb17ddd6236c5d4970ecc09680c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79b8f6c016b43c3ac3ba7b22cca9a6b72c262cb17ddd6236c5d4970ecc09680c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38c4f218eb2ca7ea0c92bd0ae2ae9efeaa40746751608fbcf7f6c3f8d9e3c758\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38c4f218eb2ca7ea0c92bd0ae2ae9efeaa40746751608fbcf7f6c3f8d9e3c758\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d66ad9831d6029c8c350054525a00069e869d39c1b5638e06e9b772666bd8630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d66ad9831d6029c8c350054525a00069e869d39c1b5638e06e9b772666bd8630\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gjtnh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:03Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:03 crc kubenswrapper[4824]: I0121 11:12:03.444709 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-9rmnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f9bd12b-22ee-44ed-a8b9-c100d043f691\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2fgn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2fgn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:28Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-9rmnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:03Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:03 crc kubenswrapper[4824]: I0121 11:12:03.452886 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d87422473838f5218c9fb646ab2515f2f42592754ffdf07b19f7a8e093fa48a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:03Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:03 crc kubenswrapper[4824]: I0121 11:12:03.461530 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://516c009d4720ab5b393c7482872f12fed22d4f4414bd1f3400411c0e8da88ce4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee18d293036971832c8d9c3f7b9eca84ab779ff6768a2db3c6f7bdcf3433cdef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:03Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:03 crc kubenswrapper[4824]: I0121 11:12:03.478360 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"26b1108a-4b98-4e39-a3cd-e0d055089fd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc319bbc06568cc811ded2b418b04b986a31f321fed492e8be50104af79bb30b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cee9cc9eec3fc815f1621a22a37805691288dbb60dd3ff64194a84be62abcddc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc4c96e33e069f7d08d33caf92cd0987668506f931fd134b246d5ce5e934f95c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ddbb0e5ae02e9d5d82a80ead7efc91f1b8e557a9fb6f39940e5530fae8de2d38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://64a9065254f889b1c43d63759283335bf78fc4fa1531a4baae472c06c075dae3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e94ef8a8f47dc75ecfbab2e3fb8e7097c1b73eed32d7ca947f11ea198cea880\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c655fe698872c3a7804fed5e2e75f300110a816eae4412891f3433762ed0161\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2c655fe698872c3a7804fed5e2e75f300110a816eae4412891f3433762ed0161\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-21T11:11:40Z\\\",\\\"message\\\":\\\"ps://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:40Z is after 2025-08-24T17:21:41Z]\\\\nI0121 11:11:40.696180 6500 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf\\\\nI0121 11:11:40.696191 6500 obj_retry.go:303] Retry object setup: *v1.Pod openshift-kube-apiserver/kube-apiserver-crc\\\\nI0121 11:11:40.696193 6500 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf in node crc\\\\nI0121 11:11:40.696185 6500 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-apiserver-operator/metrics]} name:Service_openshift-apiserver-operator/metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.38:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {8b82f026-5975-4a1b-bb18-08d5d51147ec}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0121 11:11:40.696197 6500 obj_retry.go:365] Adding new object: *v1.Pod openshift-kube-apiserver/kube\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:40Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-htkvk_openshift-ovn-kubernetes(26b1108a-4b98-4e39-a3cd-e0d055089fd8)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948e0497eaffa7b0d84b4085c9ed8d987eb2c6527c9a82c10e576ed9c9c1e21b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-htkvk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:03Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:03 crc kubenswrapper[4824]: I0121 11:12:03.488158 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-np2v8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4698abe4-0f33-4349-a2a4-614de80ae21b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://191328471bb6f2b34cde2eb927da697619826bf0b6174ecb583c8c25d27a6654\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n9qbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://050bd5bdded2162d5e9dac81713f4ed63cb575d1776e828b2f98e50d89f11b04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n9qbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-np2v8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:03Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:03 crc kubenswrapper[4824]: I0121 11:12:03.494839 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:03 crc kubenswrapper[4824]: I0121 11:12:03.494879 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:03 crc kubenswrapper[4824]: I0121 11:12:03.494889 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:03 crc kubenswrapper[4824]: I0121 11:12:03.494906 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:03 crc kubenswrapper[4824]: I0121 11:12:03.494918 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:03Z","lastTransitionTime":"2026-01-21T11:12:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:03 crc kubenswrapper[4824]: I0121 11:12:03.597179 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:03 crc kubenswrapper[4824]: I0121 11:12:03.597213 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:03 crc kubenswrapper[4824]: I0121 11:12:03.597432 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:03 crc kubenswrapper[4824]: I0121 11:12:03.597450 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:03 crc kubenswrapper[4824]: I0121 11:12:03.597459 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:03Z","lastTransitionTime":"2026-01-21T11:12:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:03 crc kubenswrapper[4824]: I0121 11:12:03.699334 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:03 crc kubenswrapper[4824]: I0121 11:12:03.699376 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:03 crc kubenswrapper[4824]: I0121 11:12:03.699389 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:03 crc kubenswrapper[4824]: I0121 11:12:03.699421 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:03 crc kubenswrapper[4824]: I0121 11:12:03.699434 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:03Z","lastTransitionTime":"2026-01-21T11:12:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:03 crc kubenswrapper[4824]: I0121 11:12:03.801819 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:03 crc kubenswrapper[4824]: I0121 11:12:03.801868 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:03 crc kubenswrapper[4824]: I0121 11:12:03.801879 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:03 crc kubenswrapper[4824]: I0121 11:12:03.801894 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:03 crc kubenswrapper[4824]: I0121 11:12:03.801903 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:03Z","lastTransitionTime":"2026-01-21T11:12:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:03 crc kubenswrapper[4824]: I0121 11:12:03.903472 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:03 crc kubenswrapper[4824]: I0121 11:12:03.903505 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:03 crc kubenswrapper[4824]: I0121 11:12:03.903518 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:03 crc kubenswrapper[4824]: I0121 11:12:03.903532 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:03 crc kubenswrapper[4824]: I0121 11:12:03.903540 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:03Z","lastTransitionTime":"2026-01-21T11:12:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:04 crc kubenswrapper[4824]: I0121 11:12:04.005802 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:04 crc kubenswrapper[4824]: I0121 11:12:04.005830 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:04 crc kubenswrapper[4824]: I0121 11:12:04.005839 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:04 crc kubenswrapper[4824]: I0121 11:12:04.005852 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:04 crc kubenswrapper[4824]: I0121 11:12:04.005860 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:04Z","lastTransitionTime":"2026-01-21T11:12:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:04 crc kubenswrapper[4824]: I0121 11:12:04.019278 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-14 22:44:58.881872006 +0000 UTC Jan 21 11:12:04 crc kubenswrapper[4824]: I0121 11:12:04.048634 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9rmnw" Jan 21 11:12:04 crc kubenswrapper[4824]: I0121 11:12:04.048634 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 11:12:04 crc kubenswrapper[4824]: E0121 11:12:04.048789 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9rmnw" podUID="8f9bd12b-22ee-44ed-a8b9-c100d043f691" Jan 21 11:12:04 crc kubenswrapper[4824]: E0121 11:12:04.048837 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 21 11:12:04 crc kubenswrapper[4824]: I0121 11:12:04.107618 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:04 crc kubenswrapper[4824]: I0121 11:12:04.107653 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:04 crc kubenswrapper[4824]: I0121 11:12:04.107661 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:04 crc kubenswrapper[4824]: I0121 11:12:04.107686 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:04 crc kubenswrapper[4824]: I0121 11:12:04.107697 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:04Z","lastTransitionTime":"2026-01-21T11:12:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:04 crc kubenswrapper[4824]: I0121 11:12:04.210075 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:04 crc kubenswrapper[4824]: I0121 11:12:04.210128 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:04 crc kubenswrapper[4824]: I0121 11:12:04.210137 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:04 crc kubenswrapper[4824]: I0121 11:12:04.210150 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:04 crc kubenswrapper[4824]: I0121 11:12:04.210159 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:04Z","lastTransitionTime":"2026-01-21T11:12:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:04 crc kubenswrapper[4824]: I0121 11:12:04.302234 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-x94nb_6b9588d1-5876-4c2c-8de5-e5c0ad46f04e/kube-multus/0.log" Jan 21 11:12:04 crc kubenswrapper[4824]: I0121 11:12:04.302296 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-x94nb" event={"ID":"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e","Type":"ContainerStarted","Data":"49ce055ec8090d6957cb7e23fe75edb82c0c2d7f87e7203d280606368dd3179f"} Jan 21 11:12:04 crc kubenswrapper[4824]: I0121 11:12:04.311804 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:04 crc kubenswrapper[4824]: I0121 11:12:04.311890 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:04 crc kubenswrapper[4824]: I0121 11:12:04.311908 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:04 crc kubenswrapper[4824]: I0121 11:12:04.311933 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:04 crc kubenswrapper[4824]: I0121 11:12:04.311947 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:04Z","lastTransitionTime":"2026-01-21T11:12:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:04 crc kubenswrapper[4824]: I0121 11:12:04.316503 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06f56d86-ec6f-4b57-b314-f508c0dd4d9f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d25c94f1ae0bbabb733f0e1ea1d1ba39562a92af73047a39aacc03013e239681\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca06234486b59cf990c69d9501e23f935328e2ba21013cd71ed64831ea749bcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b63e45d66bcd11ff3479cc3ec6b450017da8d21d36e967b8d496ae0f24b0ee9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52758bcd31f2ba64edf03adfd4d1be6428015abff5b64c7c029fa787085f9c29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2736e73342d2adb175b0721dd4f40cc42d26fca06d5eb3334927518531c3b107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b35ad7dd0570c191eb0080739304f3cd43f17b3d1dddf895549441ebe50deaa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b35ad7dd0570c191eb0080739304f3cd43f17b3d1dddf895549441ebe50deaa6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://80cd5ea5e6fc19a88c1aab5c61f827f103ab798a5c84da8dc4d45cea86d5aacb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80cd5ea5e6fc19a88c1aab5c61f827f103ab798a5c84da8dc4d45cea86d5aacb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://234d8b914e6a88480021647ae65d74e9b21c55b6788312a094d00c8a42677d48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://234d8b914e6a88480021647ae65d74e9b21c55b6788312a094d00c8a42677d48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:04Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:04 crc kubenswrapper[4824]: I0121 11:12:04.326825 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:04Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:04 crc kubenswrapper[4824]: I0121 11:12:04.335171 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33f3d922-4ffe-409b-a49a-d88c85898260\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a9e57b39eed8300107e944b9fcf6a2ff1e75991f8f37387a3de615b14f9bc78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpwst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b1b37c6d2994021cfc30f85e92c40ae35677113b17aa066d52bc7a76ae5c153\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpwst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zqjg9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:04Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:04 crc kubenswrapper[4824]: I0121 11:12:04.345471 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gjtnh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a8d2cac-2d0f-4894-9bc0-2abd81699c47\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67c122a26811deaa9b70d19a1a65daee318ca3af2c626712482179d72fe67875\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2239186f1de46481cebfcafb1cab85cd320a205b5a0dfc6dd1b321d9baedeba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b2239186f1de46481cebfcafb1cab85cd320a205b5a0dfc6dd1b321d9baedeba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0db5ecd7759d974ae1b0d8065b3b8bd4d7957fae16eb5bab2d71c2fe69243816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0db5ecd7759d974ae1b0d8065b3b8bd4d7957fae16eb5bab2d71c2fe69243816\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba9b3278c208fc7aac630ca7a0daa3de0823f85d0799d6f1f278da80ae14c862\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba9b3278c208fc7aac630ca7a0daa3de0823f85d0799d6f1f278da80ae14c862\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79b8f6c016b43c3ac3ba7b22cca9a6b72c262cb17ddd6236c5d4970ecc09680c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79b8f6c016b43c3ac3ba7b22cca9a6b72c262cb17ddd6236c5d4970ecc09680c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38c4f218eb2ca7ea0c92bd0ae2ae9efeaa40746751608fbcf7f6c3f8d9e3c758\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38c4f218eb2ca7ea0c92bd0ae2ae9efeaa40746751608fbcf7f6c3f8d9e3c758\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d66ad9831d6029c8c350054525a00069e869d39c1b5638e06e9b772666bd8630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d66ad9831d6029c8c350054525a00069e869d39c1b5638e06e9b772666bd8630\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gjtnh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:04Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:04 crc kubenswrapper[4824]: I0121 11:12:04.353425 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-9rmnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f9bd12b-22ee-44ed-a8b9-c100d043f691\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2fgn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2fgn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:28Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-9rmnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:04Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:04 crc kubenswrapper[4824]: I0121 11:12:04.363988 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86d3005cd61c0973659e1c1c0d3ef1ddd6fcf95bede5465f4f2099e64a74bb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:04Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:04 crc kubenswrapper[4824]: I0121 11:12:04.375853 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:04Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:04 crc kubenswrapper[4824]: I0121 11:12:04.386973 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x94nb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:12:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:12:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49ce055ec8090d6957cb7e23fe75edb82c0c2d7f87e7203d280606368dd3179f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75d52224135b75133fc000611cd179240b9174f6492d3263341716c7e42ccb15\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-21T11:12:02Z\\\",\\\"message\\\":\\\"2026-01-21T11:11:17+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_c9e2a0c0-e545-4ddc-b22f-97601d5900ca\\\\n2026-01-21T11:11:17+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_c9e2a0c0-e545-4ddc-b22f-97601d5900ca to /host/opt/cni/bin/\\\\n2026-01-21T11:11:17Z [verbose] multus-daemon started\\\\n2026-01-21T11:11:17Z [verbose] Readiness Indicator file check\\\\n2026-01-21T11:12:02Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:12:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgr8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x94nb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:04Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:04 crc kubenswrapper[4824]: I0121 11:12:04.395551 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-np2v8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4698abe4-0f33-4349-a2a4-614de80ae21b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://191328471bb6f2b34cde2eb927da697619826bf0b6174ecb583c8c25d27a6654\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n9qbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://050bd5bdded2162d5e9dac81713f4ed63cb575d1776e828b2f98e50d89f11b04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n9qbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-np2v8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:04Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:04 crc kubenswrapper[4824]: I0121 11:12:04.404611 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d87422473838f5218c9fb646ab2515f2f42592754ffdf07b19f7a8e093fa48a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:04Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:04 crc kubenswrapper[4824]: I0121 11:12:04.413698 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:04 crc kubenswrapper[4824]: I0121 11:12:04.413748 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:04 crc kubenswrapper[4824]: I0121 11:12:04.413761 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:04 crc kubenswrapper[4824]: I0121 11:12:04.413775 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:04 crc kubenswrapper[4824]: I0121 11:12:04.413782 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:04Z","lastTransitionTime":"2026-01-21T11:12:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:04 crc kubenswrapper[4824]: I0121 11:12:04.414089 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://516c009d4720ab5b393c7482872f12fed22d4f4414bd1f3400411c0e8da88ce4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee18d293036971832c8d9c3f7b9eca84ab779ff6768a2db3c6f7bdcf3433cdef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:04Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:04 crc kubenswrapper[4824]: I0121 11:12:04.430725 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"26b1108a-4b98-4e39-a3cd-e0d055089fd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc319bbc06568cc811ded2b418b04b986a31f321fed492e8be50104af79bb30b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cee9cc9eec3fc815f1621a22a37805691288dbb60dd3ff64194a84be62abcddc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc4c96e33e069f7d08d33caf92cd0987668506f931fd134b246d5ce5e934f95c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ddbb0e5ae02e9d5d82a80ead7efc91f1b8e557a9fb6f39940e5530fae8de2d38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://64a9065254f889b1c43d63759283335bf78fc4fa1531a4baae472c06c075dae3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e94ef8a8f47dc75ecfbab2e3fb8e7097c1b73eed32d7ca947f11ea198cea880\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c655fe698872c3a7804fed5e2e75f300110a816eae4412891f3433762ed0161\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2c655fe698872c3a7804fed5e2e75f300110a816eae4412891f3433762ed0161\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-21T11:11:40Z\\\",\\\"message\\\":\\\"ps://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:40Z is after 2025-08-24T17:21:41Z]\\\\nI0121 11:11:40.696180 6500 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf\\\\nI0121 11:11:40.696191 6500 obj_retry.go:303] Retry object setup: *v1.Pod openshift-kube-apiserver/kube-apiserver-crc\\\\nI0121 11:11:40.696193 6500 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf in node crc\\\\nI0121 11:11:40.696185 6500 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-apiserver-operator/metrics]} name:Service_openshift-apiserver-operator/metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.38:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {8b82f026-5975-4a1b-bb18-08d5d51147ec}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0121 11:11:40.696197 6500 obj_retry.go:365] Adding new object: *v1.Pod openshift-kube-apiserver/kube\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:40Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-htkvk_openshift-ovn-kubernetes(26b1108a-4b98-4e39-a3cd-e0d055089fd8)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948e0497eaffa7b0d84b4085c9ed8d987eb2c6527c9a82c10e576ed9c9c1e21b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-htkvk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:04Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:04 crc kubenswrapper[4824]: I0121 11:12:04.440791 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e570fff1-6227-44d6-b71f-ee377b1099ae\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4d41492c3802a8168fffd249c15aeb9c902a5c7e01550e1aa6f9b769adfdc4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2690a1ecbc8fe1290e68bba2695a3edd833b5da933fb71e8ddc926a1340a9a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b85db961e1e078e1f54323c79a99c9413365256ac2fce411db7932bd0580ab6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://595783a1454735bd33f4862180b60d99a4e1b6fae1e735f2b7328391951bf422\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://595783a1454735bd33f4862180b60d99a4e1b6fae1e735f2b7328391951bf422\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:04Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:04 crc kubenswrapper[4824]: I0121 11:12:04.452012 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:04Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:04 crc kubenswrapper[4824]: I0121 11:12:04.460485 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lv4rn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1267539-9adf-466f-8027-8ede009d19a0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eed1b884aa53e1ee318169a0cc2c9ad10130aa363ae4ac55e4bf065237756f53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbmfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lv4rn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:04Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:04 crc kubenswrapper[4824]: I0121 11:12:04.469890 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jr64s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9768c60d-5bee-4286-bced-076e584b04e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91158f14d2e98ac9f875a5e7101df3107ef28aa5fc5d62218f460af98ff920fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flrnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:17Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jr64s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:04Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:04 crc kubenswrapper[4824]: I0121 11:12:04.481149 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d31ab61-f94c-4f49-85d0-c9e91b83069b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a66b5c33884af00408dc35cf02f08ebd8570bd695f9f1fe558bf54fd06930431\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7514786602746dabfa690da48d48e233fe68d6dd6674aa5ecd489c0b3185f08\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61b14b6846f94cdeba8f1f61b3754e7663041f92c70356c4ee6cbb988a2e680f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a7023cfe4c098bcdd9c0477a8187f4ed7a0d2abdbe1e143436863948f441b4f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af11ad95bcf0c6bbd5fa4ab99bb592ce1c87f40d29c5ade7832414315820a569\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW0121 11:11:14.604290 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0121 11:11:14.604572 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0121 11:11:14.605875 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3284452394/tls.crt::/tmp/serving-cert-3284452394/tls.key\\\\\\\"\\\\nI0121 11:11:14.923427 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0121 11:11:14.928141 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0121 11:11:14.928162 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0121 11:11:14.928179 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0121 11:11:14.928184 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0121 11:11:14.933154 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0121 11:11:14.933230 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 11:11:14.933259 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 11:11:14.933282 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0121 11:11:14.933299 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0121 11:11:14.933315 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0121 11:11:14.933331 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0121 11:11:14.933177 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0121 11:11:14.934666 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48070c09fad511108ff22e44211cf26911509b8406d66948985cbfc3009d8233\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:04Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:04 crc kubenswrapper[4824]: I0121 11:12:04.491911 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9bb99c0a-683b-44da-9c0e-e8d284e87d53\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05e076febd697ed802ce5c7ee92a8c2c636a1a2c6609ad7eb01f17dace247c4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43529d1b9e1e5e22b0d56b023946119da8aacbae54b68c26ff0a1a82dc890c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ea232ad87393685aa323ce6393ff3869f8f8d423c2406a501c69c355535adfa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc6c7eff9e17b8a8ecc88bbfe703b5a01dbf1e02d342cede4f6a263a6526af43\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:04Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:04 crc kubenswrapper[4824]: I0121 11:12:04.517020 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:04 crc kubenswrapper[4824]: I0121 11:12:04.517085 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:04 crc kubenswrapper[4824]: I0121 11:12:04.517100 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:04 crc kubenswrapper[4824]: I0121 11:12:04.517119 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:04 crc kubenswrapper[4824]: I0121 11:12:04.517129 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:04Z","lastTransitionTime":"2026-01-21T11:12:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:04 crc kubenswrapper[4824]: I0121 11:12:04.620006 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:04 crc kubenswrapper[4824]: I0121 11:12:04.620073 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:04 crc kubenswrapper[4824]: I0121 11:12:04.620087 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:04 crc kubenswrapper[4824]: I0121 11:12:04.620111 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:04 crc kubenswrapper[4824]: I0121 11:12:04.620124 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:04Z","lastTransitionTime":"2026-01-21T11:12:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:04 crc kubenswrapper[4824]: I0121 11:12:04.721799 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:04 crc kubenswrapper[4824]: I0121 11:12:04.721935 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:04 crc kubenswrapper[4824]: I0121 11:12:04.722036 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:04 crc kubenswrapper[4824]: I0121 11:12:04.722116 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:04 crc kubenswrapper[4824]: I0121 11:12:04.722174 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:04Z","lastTransitionTime":"2026-01-21T11:12:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:04 crc kubenswrapper[4824]: I0121 11:12:04.824680 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:04 crc kubenswrapper[4824]: I0121 11:12:04.824729 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:04 crc kubenswrapper[4824]: I0121 11:12:04.824740 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:04 crc kubenswrapper[4824]: I0121 11:12:04.824758 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:04 crc kubenswrapper[4824]: I0121 11:12:04.824767 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:04Z","lastTransitionTime":"2026-01-21T11:12:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:04 crc kubenswrapper[4824]: I0121 11:12:04.926504 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:04 crc kubenswrapper[4824]: I0121 11:12:04.926556 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:04 crc kubenswrapper[4824]: I0121 11:12:04.926565 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:04 crc kubenswrapper[4824]: I0121 11:12:04.926579 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:04 crc kubenswrapper[4824]: I0121 11:12:04.926590 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:04Z","lastTransitionTime":"2026-01-21T11:12:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:05 crc kubenswrapper[4824]: I0121 11:12:05.019837 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-01 06:58:50.036087652 +0000 UTC Jan 21 11:12:05 crc kubenswrapper[4824]: I0121 11:12:05.028077 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:05 crc kubenswrapper[4824]: I0121 11:12:05.028111 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:05 crc kubenswrapper[4824]: I0121 11:12:05.028121 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:05 crc kubenswrapper[4824]: I0121 11:12:05.028135 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:05 crc kubenswrapper[4824]: I0121 11:12:05.028143 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:05Z","lastTransitionTime":"2026-01-21T11:12:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:05 crc kubenswrapper[4824]: I0121 11:12:05.048225 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 11:12:05 crc kubenswrapper[4824]: E0121 11:12:05.048419 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 21 11:12:05 crc kubenswrapper[4824]: I0121 11:12:05.048270 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 11:12:05 crc kubenswrapper[4824]: E0121 11:12:05.048634 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 21 11:12:05 crc kubenswrapper[4824]: I0121 11:12:05.130552 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:05 crc kubenswrapper[4824]: I0121 11:12:05.130589 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:05 crc kubenswrapper[4824]: I0121 11:12:05.130618 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:05 crc kubenswrapper[4824]: I0121 11:12:05.130632 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:05 crc kubenswrapper[4824]: I0121 11:12:05.130639 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:05Z","lastTransitionTime":"2026-01-21T11:12:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:05 crc kubenswrapper[4824]: I0121 11:12:05.232644 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:05 crc kubenswrapper[4824]: I0121 11:12:05.232690 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:05 crc kubenswrapper[4824]: I0121 11:12:05.232702 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:05 crc kubenswrapper[4824]: I0121 11:12:05.232722 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:05 crc kubenswrapper[4824]: I0121 11:12:05.232738 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:05Z","lastTransitionTime":"2026-01-21T11:12:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:05 crc kubenswrapper[4824]: I0121 11:12:05.335558 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:05 crc kubenswrapper[4824]: I0121 11:12:05.335594 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:05 crc kubenswrapper[4824]: I0121 11:12:05.335612 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:05 crc kubenswrapper[4824]: I0121 11:12:05.335626 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:05 crc kubenswrapper[4824]: I0121 11:12:05.335637 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:05Z","lastTransitionTime":"2026-01-21T11:12:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:05 crc kubenswrapper[4824]: I0121 11:12:05.438525 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:05 crc kubenswrapper[4824]: I0121 11:12:05.438564 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:05 crc kubenswrapper[4824]: I0121 11:12:05.438574 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:05 crc kubenswrapper[4824]: I0121 11:12:05.438588 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:05 crc kubenswrapper[4824]: I0121 11:12:05.438598 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:05Z","lastTransitionTime":"2026-01-21T11:12:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:05 crc kubenswrapper[4824]: I0121 11:12:05.540988 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:05 crc kubenswrapper[4824]: I0121 11:12:05.541055 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:05 crc kubenswrapper[4824]: I0121 11:12:05.541068 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:05 crc kubenswrapper[4824]: I0121 11:12:05.541096 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:05 crc kubenswrapper[4824]: I0121 11:12:05.541109 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:05Z","lastTransitionTime":"2026-01-21T11:12:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:05 crc kubenswrapper[4824]: I0121 11:12:05.644007 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:05 crc kubenswrapper[4824]: I0121 11:12:05.644058 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:05 crc kubenswrapper[4824]: I0121 11:12:05.644069 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:05 crc kubenswrapper[4824]: I0121 11:12:05.644091 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:05 crc kubenswrapper[4824]: I0121 11:12:05.644106 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:05Z","lastTransitionTime":"2026-01-21T11:12:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:05 crc kubenswrapper[4824]: I0121 11:12:05.746503 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:05 crc kubenswrapper[4824]: I0121 11:12:05.746556 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:05 crc kubenswrapper[4824]: I0121 11:12:05.746567 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:05 crc kubenswrapper[4824]: I0121 11:12:05.746584 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:05 crc kubenswrapper[4824]: I0121 11:12:05.746593 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:05Z","lastTransitionTime":"2026-01-21T11:12:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:05 crc kubenswrapper[4824]: I0121 11:12:05.848593 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:05 crc kubenswrapper[4824]: I0121 11:12:05.848642 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:05 crc kubenswrapper[4824]: I0121 11:12:05.848654 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:05 crc kubenswrapper[4824]: I0121 11:12:05.848669 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:05 crc kubenswrapper[4824]: I0121 11:12:05.848681 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:05Z","lastTransitionTime":"2026-01-21T11:12:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:05 crc kubenswrapper[4824]: I0121 11:12:05.949832 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:05 crc kubenswrapper[4824]: I0121 11:12:05.950055 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:05 crc kubenswrapper[4824]: I0121 11:12:05.950127 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:05 crc kubenswrapper[4824]: I0121 11:12:05.950235 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:05 crc kubenswrapper[4824]: I0121 11:12:05.950315 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:05Z","lastTransitionTime":"2026-01-21T11:12:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:06 crc kubenswrapper[4824]: I0121 11:12:06.020750 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-17 17:24:25.064947226 +0000 UTC Jan 21 11:12:06 crc kubenswrapper[4824]: I0121 11:12:06.049125 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9rmnw" Jan 21 11:12:06 crc kubenswrapper[4824]: E0121 11:12:06.049236 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9rmnw" podUID="8f9bd12b-22ee-44ed-a8b9-c100d043f691" Jan 21 11:12:06 crc kubenswrapper[4824]: I0121 11:12:06.049324 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 11:12:06 crc kubenswrapper[4824]: E0121 11:12:06.049599 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 21 11:12:06 crc kubenswrapper[4824]: I0121 11:12:06.052276 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:06 crc kubenswrapper[4824]: I0121 11:12:06.052317 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:06 crc kubenswrapper[4824]: I0121 11:12:06.052327 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:06 crc kubenswrapper[4824]: I0121 11:12:06.052340 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:06 crc kubenswrapper[4824]: I0121 11:12:06.052349 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:06Z","lastTransitionTime":"2026-01-21T11:12:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:06 crc kubenswrapper[4824]: I0121 11:12:06.154738 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:06 crc kubenswrapper[4824]: I0121 11:12:06.154788 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:06 crc kubenswrapper[4824]: I0121 11:12:06.154798 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:06 crc kubenswrapper[4824]: I0121 11:12:06.154815 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:06 crc kubenswrapper[4824]: I0121 11:12:06.154825 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:06Z","lastTransitionTime":"2026-01-21T11:12:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:06 crc kubenswrapper[4824]: I0121 11:12:06.256768 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:06 crc kubenswrapper[4824]: I0121 11:12:06.256801 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:06 crc kubenswrapper[4824]: I0121 11:12:06.256813 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:06 crc kubenswrapper[4824]: I0121 11:12:06.256826 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:06 crc kubenswrapper[4824]: I0121 11:12:06.256837 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:06Z","lastTransitionTime":"2026-01-21T11:12:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:06 crc kubenswrapper[4824]: I0121 11:12:06.359063 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:06 crc kubenswrapper[4824]: I0121 11:12:06.359100 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:06 crc kubenswrapper[4824]: I0121 11:12:06.359109 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:06 crc kubenswrapper[4824]: I0121 11:12:06.359124 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:06 crc kubenswrapper[4824]: I0121 11:12:06.359134 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:06Z","lastTransitionTime":"2026-01-21T11:12:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:06 crc kubenswrapper[4824]: I0121 11:12:06.461160 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:06 crc kubenswrapper[4824]: I0121 11:12:06.461199 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:06 crc kubenswrapper[4824]: I0121 11:12:06.461209 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:06 crc kubenswrapper[4824]: I0121 11:12:06.461223 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:06 crc kubenswrapper[4824]: I0121 11:12:06.461233 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:06Z","lastTransitionTime":"2026-01-21T11:12:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:06 crc kubenswrapper[4824]: I0121 11:12:06.563267 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:06 crc kubenswrapper[4824]: I0121 11:12:06.563307 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:06 crc kubenswrapper[4824]: I0121 11:12:06.563318 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:06 crc kubenswrapper[4824]: I0121 11:12:06.563330 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:06 crc kubenswrapper[4824]: I0121 11:12:06.563342 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:06Z","lastTransitionTime":"2026-01-21T11:12:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:06 crc kubenswrapper[4824]: I0121 11:12:06.665626 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:06 crc kubenswrapper[4824]: I0121 11:12:06.665663 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:06 crc kubenswrapper[4824]: I0121 11:12:06.665672 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:06 crc kubenswrapper[4824]: I0121 11:12:06.665687 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:06 crc kubenswrapper[4824]: I0121 11:12:06.665696 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:06Z","lastTransitionTime":"2026-01-21T11:12:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:06 crc kubenswrapper[4824]: I0121 11:12:06.767307 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:06 crc kubenswrapper[4824]: I0121 11:12:06.767347 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:06 crc kubenswrapper[4824]: I0121 11:12:06.767356 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:06 crc kubenswrapper[4824]: I0121 11:12:06.767374 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:06 crc kubenswrapper[4824]: I0121 11:12:06.767385 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:06Z","lastTransitionTime":"2026-01-21T11:12:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:06 crc kubenswrapper[4824]: I0121 11:12:06.869079 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:06 crc kubenswrapper[4824]: I0121 11:12:06.869119 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:06 crc kubenswrapper[4824]: I0121 11:12:06.869134 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:06 crc kubenswrapper[4824]: I0121 11:12:06.869151 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:06 crc kubenswrapper[4824]: I0121 11:12:06.869163 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:06Z","lastTransitionTime":"2026-01-21T11:12:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:06 crc kubenswrapper[4824]: I0121 11:12:06.970942 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:06 crc kubenswrapper[4824]: I0121 11:12:06.970999 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:06 crc kubenswrapper[4824]: I0121 11:12:06.971012 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:06 crc kubenswrapper[4824]: I0121 11:12:06.971029 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:06 crc kubenswrapper[4824]: I0121 11:12:06.971040 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:06Z","lastTransitionTime":"2026-01-21T11:12:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:07 crc kubenswrapper[4824]: I0121 11:12:07.021460 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-05 02:46:11.749001905 +0000 UTC Jan 21 11:12:07 crc kubenswrapper[4824]: I0121 11:12:07.048852 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 11:12:07 crc kubenswrapper[4824]: I0121 11:12:07.048920 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 11:12:07 crc kubenswrapper[4824]: E0121 11:12:07.049148 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 21 11:12:07 crc kubenswrapper[4824]: E0121 11:12:07.049258 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 21 11:12:07 crc kubenswrapper[4824]: I0121 11:12:07.049344 4824 scope.go:117] "RemoveContainer" containerID="2c655fe698872c3a7804fed5e2e75f300110a816eae4412891f3433762ed0161" Jan 21 11:12:07 crc kubenswrapper[4824]: I0121 11:12:07.073458 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:07 crc kubenswrapper[4824]: I0121 11:12:07.073495 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:07 crc kubenswrapper[4824]: I0121 11:12:07.073505 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:07 crc kubenswrapper[4824]: I0121 11:12:07.073520 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:07 crc kubenswrapper[4824]: I0121 11:12:07.073531 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:07Z","lastTransitionTime":"2026-01-21T11:12:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:07 crc kubenswrapper[4824]: I0121 11:12:07.175872 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:07 crc kubenswrapper[4824]: I0121 11:12:07.175904 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:07 crc kubenswrapper[4824]: I0121 11:12:07.175914 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:07 crc kubenswrapper[4824]: I0121 11:12:07.175928 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:07 crc kubenswrapper[4824]: I0121 11:12:07.175938 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:07Z","lastTransitionTime":"2026-01-21T11:12:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:07 crc kubenswrapper[4824]: I0121 11:12:07.278142 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:07 crc kubenswrapper[4824]: I0121 11:12:07.278190 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:07 crc kubenswrapper[4824]: I0121 11:12:07.278199 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:07 crc kubenswrapper[4824]: I0121 11:12:07.278219 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:07 crc kubenswrapper[4824]: I0121 11:12:07.278230 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:07Z","lastTransitionTime":"2026-01-21T11:12:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:07 crc kubenswrapper[4824]: I0121 11:12:07.311433 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-htkvk_26b1108a-4b98-4e39-a3cd-e0d055089fd8/ovnkube-controller/2.log" Jan 21 11:12:07 crc kubenswrapper[4824]: I0121 11:12:07.313518 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" event={"ID":"26b1108a-4b98-4e39-a3cd-e0d055089fd8","Type":"ContainerStarted","Data":"86a6881f109dab1f910fe505a21c7c800e587de4dded2635147997c51c05c483"} Jan 21 11:12:07 crc kubenswrapper[4824]: I0121 11:12:07.314524 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" Jan 21 11:12:07 crc kubenswrapper[4824]: I0121 11:12:07.333741 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06f56d86-ec6f-4b57-b314-f508c0dd4d9f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d25c94f1ae0bbabb733f0e1ea1d1ba39562a92af73047a39aacc03013e239681\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca06234486b59cf990c69d9501e23f935328e2ba21013cd71ed64831ea749bcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b63e45d66bcd11ff3479cc3ec6b450017da8d21d36e967b8d496ae0f24b0ee9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52758bcd31f2ba64edf03adfd4d1be6428015abff5b64c7c029fa787085f9c29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2736e73342d2adb175b0721dd4f40cc42d26fca06d5eb3334927518531c3b107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b35ad7dd0570c191eb0080739304f3cd43f17b3d1dddf895549441ebe50deaa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b35ad7dd0570c191eb0080739304f3cd43f17b3d1dddf895549441ebe50deaa6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://80cd5ea5e6fc19a88c1aab5c61f827f103ab798a5c84da8dc4d45cea86d5aacb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80cd5ea5e6fc19a88c1aab5c61f827f103ab798a5c84da8dc4d45cea86d5aacb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://234d8b914e6a88480021647ae65d74e9b21c55b6788312a094d00c8a42677d48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://234d8b914e6a88480021647ae65d74e9b21c55b6788312a094d00c8a42677d48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:07Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:07 crc kubenswrapper[4824]: I0121 11:12:07.343243 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:07Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:07 crc kubenswrapper[4824]: I0121 11:12:07.359894 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x94nb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:12:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:12:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49ce055ec8090d6957cb7e23fe75edb82c0c2d7f87e7203d280606368dd3179f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75d52224135b75133fc000611cd179240b9174f6492d3263341716c7e42ccb15\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-21T11:12:02Z\\\",\\\"message\\\":\\\"2026-01-21T11:11:17+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_c9e2a0c0-e545-4ddc-b22f-97601d5900ca\\\\n2026-01-21T11:11:17+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_c9e2a0c0-e545-4ddc-b22f-97601d5900ca to /host/opt/cni/bin/\\\\n2026-01-21T11:11:17Z [verbose] multus-daemon started\\\\n2026-01-21T11:11:17Z [verbose] Readiness Indicator file check\\\\n2026-01-21T11:12:02Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:12:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgr8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x94nb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:07Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:07 crc kubenswrapper[4824]: I0121 11:12:07.379714 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33f3d922-4ffe-409b-a49a-d88c85898260\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a9e57b39eed8300107e944b9fcf6a2ff1e75991f8f37387a3de615b14f9bc78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpwst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b1b37c6d2994021cfc30f85e92c40ae35677113b17aa066d52bc7a76ae5c153\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpwst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zqjg9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:07Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:07 crc kubenswrapper[4824]: I0121 11:12:07.381307 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:07 crc kubenswrapper[4824]: I0121 11:12:07.381343 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:07 crc kubenswrapper[4824]: I0121 11:12:07.381355 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:07 crc kubenswrapper[4824]: I0121 11:12:07.381373 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:07 crc kubenswrapper[4824]: I0121 11:12:07.381383 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:07Z","lastTransitionTime":"2026-01-21T11:12:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:07 crc kubenswrapper[4824]: I0121 11:12:07.397195 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gjtnh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a8d2cac-2d0f-4894-9bc0-2abd81699c47\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67c122a26811deaa9b70d19a1a65daee318ca3af2c626712482179d72fe67875\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2239186f1de46481cebfcafb1cab85cd320a205b5a0dfc6dd1b321d9baedeba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b2239186f1de46481cebfcafb1cab85cd320a205b5a0dfc6dd1b321d9baedeba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0db5ecd7759d974ae1b0d8065b3b8bd4d7957fae16eb5bab2d71c2fe69243816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0db5ecd7759d974ae1b0d8065b3b8bd4d7957fae16eb5bab2d71c2fe69243816\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba9b3278c208fc7aac630ca7a0daa3de0823f85d0799d6f1f278da80ae14c862\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba9b3278c208fc7aac630ca7a0daa3de0823f85d0799d6f1f278da80ae14c862\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79b8f6c016b43c3ac3ba7b22cca9a6b72c262cb17ddd6236c5d4970ecc09680c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79b8f6c016b43c3ac3ba7b22cca9a6b72c262cb17ddd6236c5d4970ecc09680c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38c4f218eb2ca7ea0c92bd0ae2ae9efeaa40746751608fbcf7f6c3f8d9e3c758\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38c4f218eb2ca7ea0c92bd0ae2ae9efeaa40746751608fbcf7f6c3f8d9e3c758\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d66ad9831d6029c8c350054525a00069e869d39c1b5638e06e9b772666bd8630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d66ad9831d6029c8c350054525a00069e869d39c1b5638e06e9b772666bd8630\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gjtnh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:07Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:07 crc kubenswrapper[4824]: I0121 11:12:07.404901 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-9rmnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f9bd12b-22ee-44ed-a8b9-c100d043f691\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2fgn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2fgn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:28Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-9rmnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:07Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:07 crc kubenswrapper[4824]: I0121 11:12:07.418592 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86d3005cd61c0973659e1c1c0d3ef1ddd6fcf95bede5465f4f2099e64a74bb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:07Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:07 crc kubenswrapper[4824]: I0121 11:12:07.427433 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:07Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:07 crc kubenswrapper[4824]: I0121 11:12:07.440932 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"26b1108a-4b98-4e39-a3cd-e0d055089fd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc319bbc06568cc811ded2b418b04b986a31f321fed492e8be50104af79bb30b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cee9cc9eec3fc815f1621a22a37805691288dbb60dd3ff64194a84be62abcddc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc4c96e33e069f7d08d33caf92cd0987668506f931fd134b246d5ce5e934f95c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ddbb0e5ae02e9d5d82a80ead7efc91f1b8e557a9fb6f39940e5530fae8de2d38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://64a9065254f889b1c43d63759283335bf78fc4fa1531a4baae472c06c075dae3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e94ef8a8f47dc75ecfbab2e3fb8e7097c1b73eed32d7ca947f11ea198cea880\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86a6881f109dab1f910fe505a21c7c800e587de4dded2635147997c51c05c483\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2c655fe698872c3a7804fed5e2e75f300110a816eae4412891f3433762ed0161\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-21T11:11:40Z\\\",\\\"message\\\":\\\"ps://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:40Z is after 2025-08-24T17:21:41Z]\\\\nI0121 11:11:40.696180 6500 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf\\\\nI0121 11:11:40.696191 6500 obj_retry.go:303] Retry object setup: *v1.Pod openshift-kube-apiserver/kube-apiserver-crc\\\\nI0121 11:11:40.696193 6500 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf in node crc\\\\nI0121 11:11:40.696185 6500 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-apiserver-operator/metrics]} name:Service_openshift-apiserver-operator/metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.38:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {8b82f026-5975-4a1b-bb18-08d5d51147ec}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0121 11:11:40.696197 6500 obj_retry.go:365] Adding new object: *v1.Pod openshift-kube-apiserver/kube\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:40Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:12:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948e0497eaffa7b0d84b4085c9ed8d987eb2c6527c9a82c10e576ed9c9c1e21b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-htkvk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:07Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:07 crc kubenswrapper[4824]: I0121 11:12:07.449564 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-np2v8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4698abe4-0f33-4349-a2a4-614de80ae21b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://191328471bb6f2b34cde2eb927da697619826bf0b6174ecb583c8c25d27a6654\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n9qbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://050bd5bdded2162d5e9dac81713f4ed63cb575d1776e828b2f98e50d89f11b04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n9qbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-np2v8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:07Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:07 crc kubenswrapper[4824]: I0121 11:12:07.457894 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d87422473838f5218c9fb646ab2515f2f42592754ffdf07b19f7a8e093fa48a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:07Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:07 crc kubenswrapper[4824]: I0121 11:12:07.466971 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://516c009d4720ab5b393c7482872f12fed22d4f4414bd1f3400411c0e8da88ce4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee18d293036971832c8d9c3f7b9eca84ab779ff6768a2db3c6f7bdcf3433cdef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:07Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:07 crc kubenswrapper[4824]: I0121 11:12:07.476594 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9bb99c0a-683b-44da-9c0e-e8d284e87d53\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05e076febd697ed802ce5c7ee92a8c2c636a1a2c6609ad7eb01f17dace247c4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43529d1b9e1e5e22b0d56b023946119da8aacbae54b68c26ff0a1a82dc890c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ea232ad87393685aa323ce6393ff3869f8f8d423c2406a501c69c355535adfa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc6c7eff9e17b8a8ecc88bbfe703b5a01dbf1e02d342cede4f6a263a6526af43\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:07Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:07 crc kubenswrapper[4824]: I0121 11:12:07.484122 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:07 crc kubenswrapper[4824]: I0121 11:12:07.484170 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:07 crc kubenswrapper[4824]: I0121 11:12:07.484181 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:07 crc kubenswrapper[4824]: I0121 11:12:07.484199 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:07 crc kubenswrapper[4824]: I0121 11:12:07.484210 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:07Z","lastTransitionTime":"2026-01-21T11:12:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:07 crc kubenswrapper[4824]: I0121 11:12:07.486543 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e570fff1-6227-44d6-b71f-ee377b1099ae\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4d41492c3802a8168fffd249c15aeb9c902a5c7e01550e1aa6f9b769adfdc4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2690a1ecbc8fe1290e68bba2695a3edd833b5da933fb71e8ddc926a1340a9a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b85db961e1e078e1f54323c79a99c9413365256ac2fce411db7932bd0580ab6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://595783a1454735bd33f4862180b60d99a4e1b6fae1e735f2b7328391951bf422\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://595783a1454735bd33f4862180b60d99a4e1b6fae1e735f2b7328391951bf422\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:07Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:07 crc kubenswrapper[4824]: I0121 11:12:07.495511 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:07Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:07 crc kubenswrapper[4824]: I0121 11:12:07.502777 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lv4rn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1267539-9adf-466f-8027-8ede009d19a0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eed1b884aa53e1ee318169a0cc2c9ad10130aa363ae4ac55e4bf065237756f53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbmfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lv4rn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:07Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:07 crc kubenswrapper[4824]: I0121 11:12:07.510132 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jr64s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9768c60d-5bee-4286-bced-076e584b04e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91158f14d2e98ac9f875a5e7101df3107ef28aa5fc5d62218f460af98ff920fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flrnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:17Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jr64s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:07Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:07 crc kubenswrapper[4824]: I0121 11:12:07.521010 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d31ab61-f94c-4f49-85d0-c9e91b83069b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a66b5c33884af00408dc35cf02f08ebd8570bd695f9f1fe558bf54fd06930431\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7514786602746dabfa690da48d48e233fe68d6dd6674aa5ecd489c0b3185f08\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61b14b6846f94cdeba8f1f61b3754e7663041f92c70356c4ee6cbb988a2e680f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a7023cfe4c098bcdd9c0477a8187f4ed7a0d2abdbe1e143436863948f441b4f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af11ad95bcf0c6bbd5fa4ab99bb592ce1c87f40d29c5ade7832414315820a569\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW0121 11:11:14.604290 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0121 11:11:14.604572 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0121 11:11:14.605875 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3284452394/tls.crt::/tmp/serving-cert-3284452394/tls.key\\\\\\\"\\\\nI0121 11:11:14.923427 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0121 11:11:14.928141 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0121 11:11:14.928162 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0121 11:11:14.928179 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0121 11:11:14.928184 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0121 11:11:14.933154 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0121 11:11:14.933230 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 11:11:14.933259 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 11:11:14.933282 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0121 11:11:14.933299 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0121 11:11:14.933315 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0121 11:11:14.933331 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0121 11:11:14.933177 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0121 11:11:14.934666 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48070c09fad511108ff22e44211cf26911509b8406d66948985cbfc3009d8233\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:07Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:07 crc kubenswrapper[4824]: I0121 11:12:07.586519 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:07 crc kubenswrapper[4824]: I0121 11:12:07.586565 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:07 crc kubenswrapper[4824]: I0121 11:12:07.586578 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:07 crc kubenswrapper[4824]: I0121 11:12:07.586597 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:07 crc kubenswrapper[4824]: I0121 11:12:07.586609 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:07Z","lastTransitionTime":"2026-01-21T11:12:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:07 crc kubenswrapper[4824]: I0121 11:12:07.688841 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:07 crc kubenswrapper[4824]: I0121 11:12:07.688900 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:07 crc kubenswrapper[4824]: I0121 11:12:07.688911 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:07 crc kubenswrapper[4824]: I0121 11:12:07.688929 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:07 crc kubenswrapper[4824]: I0121 11:12:07.688944 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:07Z","lastTransitionTime":"2026-01-21T11:12:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:07 crc kubenswrapper[4824]: I0121 11:12:07.790377 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:07 crc kubenswrapper[4824]: I0121 11:12:07.790413 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:07 crc kubenswrapper[4824]: I0121 11:12:07.790435 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:07 crc kubenswrapper[4824]: I0121 11:12:07.790449 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:07 crc kubenswrapper[4824]: I0121 11:12:07.790460 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:07Z","lastTransitionTime":"2026-01-21T11:12:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:07 crc kubenswrapper[4824]: I0121 11:12:07.892679 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:07 crc kubenswrapper[4824]: I0121 11:12:07.892713 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:07 crc kubenswrapper[4824]: I0121 11:12:07.892722 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:07 crc kubenswrapper[4824]: I0121 11:12:07.892736 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:07 crc kubenswrapper[4824]: I0121 11:12:07.892746 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:07Z","lastTransitionTime":"2026-01-21T11:12:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:07 crc kubenswrapper[4824]: I0121 11:12:07.994392 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:07 crc kubenswrapper[4824]: I0121 11:12:07.994435 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:07 crc kubenswrapper[4824]: I0121 11:12:07.994446 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:07 crc kubenswrapper[4824]: I0121 11:12:07.994463 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:07 crc kubenswrapper[4824]: I0121 11:12:07.994473 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:07Z","lastTransitionTime":"2026-01-21T11:12:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.022054 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-23 08:15:39.520093142 +0000 UTC Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.048274 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9rmnw" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.048281 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 11:12:08 crc kubenswrapper[4824]: E0121 11:12:08.048383 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9rmnw" podUID="8f9bd12b-22ee-44ed-a8b9-c100d043f691" Jan 21 11:12:08 crc kubenswrapper[4824]: E0121 11:12:08.048465 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.060564 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:08Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.069941 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lv4rn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1267539-9adf-466f-8027-8ede009d19a0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eed1b884aa53e1ee318169a0cc2c9ad10130aa363ae4ac55e4bf065237756f53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbmfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lv4rn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:08Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.076637 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jr64s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9768c60d-5bee-4286-bced-076e584b04e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91158f14d2e98ac9f875a5e7101df3107ef28aa5fc5d62218f460af98ff920fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flrnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:17Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jr64s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:08Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.086612 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d31ab61-f94c-4f49-85d0-c9e91b83069b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a66b5c33884af00408dc35cf02f08ebd8570bd695f9f1fe558bf54fd06930431\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7514786602746dabfa690da48d48e233fe68d6dd6674aa5ecd489c0b3185f08\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61b14b6846f94cdeba8f1f61b3754e7663041f92c70356c4ee6cbb988a2e680f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a7023cfe4c098bcdd9c0477a8187f4ed7a0d2abdbe1e143436863948f441b4f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af11ad95bcf0c6bbd5fa4ab99bb592ce1c87f40d29c5ade7832414315820a569\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW0121 11:11:14.604290 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0121 11:11:14.604572 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0121 11:11:14.605875 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3284452394/tls.crt::/tmp/serving-cert-3284452394/tls.key\\\\\\\"\\\\nI0121 11:11:14.923427 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0121 11:11:14.928141 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0121 11:11:14.928162 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0121 11:11:14.928179 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0121 11:11:14.928184 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0121 11:11:14.933154 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0121 11:11:14.933230 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 11:11:14.933259 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 11:11:14.933282 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0121 11:11:14.933299 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0121 11:11:14.933315 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0121 11:11:14.933331 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0121 11:11:14.933177 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0121 11:11:14.934666 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48070c09fad511108ff22e44211cf26911509b8406d66948985cbfc3009d8233\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:08Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.095522 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.095542 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.095550 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.095560 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.095568 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:08Z","lastTransitionTime":"2026-01-21T11:12:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.095918 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9bb99c0a-683b-44da-9c0e-e8d284e87d53\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05e076febd697ed802ce5c7ee92a8c2c636a1a2c6609ad7eb01f17dace247c4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43529d1b9e1e5e22b0d56b023946119da8aacbae54b68c26ff0a1a82dc890c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ea232ad87393685aa323ce6393ff3869f8f8d423c2406a501c69c355535adfa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc6c7eff9e17b8a8ecc88bbfe703b5a01dbf1e02d342cede4f6a263a6526af43\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:08Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.103697 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e570fff1-6227-44d6-b71f-ee377b1099ae\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4d41492c3802a8168fffd249c15aeb9c902a5c7e01550e1aa6f9b769adfdc4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2690a1ecbc8fe1290e68bba2695a3edd833b5da933fb71e8ddc926a1340a9a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b85db961e1e078e1f54323c79a99c9413365256ac2fce411db7932bd0580ab6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://595783a1454735bd33f4862180b60d99a4e1b6fae1e735f2b7328391951bf422\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://595783a1454735bd33f4862180b60d99a4e1b6fae1e735f2b7328391951bf422\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:08Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.116646 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06f56d86-ec6f-4b57-b314-f508c0dd4d9f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d25c94f1ae0bbabb733f0e1ea1d1ba39562a92af73047a39aacc03013e239681\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca06234486b59cf990c69d9501e23f935328e2ba21013cd71ed64831ea749bcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b63e45d66bcd11ff3479cc3ec6b450017da8d21d36e967b8d496ae0f24b0ee9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52758bcd31f2ba64edf03adfd4d1be6428015abff5b64c7c029fa787085f9c29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2736e73342d2adb175b0721dd4f40cc42d26fca06d5eb3334927518531c3b107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b35ad7dd0570c191eb0080739304f3cd43f17b3d1dddf895549441ebe50deaa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b35ad7dd0570c191eb0080739304f3cd43f17b3d1dddf895549441ebe50deaa6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://80cd5ea5e6fc19a88c1aab5c61f827f103ab798a5c84da8dc4d45cea86d5aacb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80cd5ea5e6fc19a88c1aab5c61f827f103ab798a5c84da8dc4d45cea86d5aacb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://234d8b914e6a88480021647ae65d74e9b21c55b6788312a094d00c8a42677d48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://234d8b914e6a88480021647ae65d74e9b21c55b6788312a094d00c8a42677d48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:08Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.127381 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:08Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.137147 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gjtnh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a8d2cac-2d0f-4894-9bc0-2abd81699c47\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67c122a26811deaa9b70d19a1a65daee318ca3af2c626712482179d72fe67875\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2239186f1de46481cebfcafb1cab85cd320a205b5a0dfc6dd1b321d9baedeba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b2239186f1de46481cebfcafb1cab85cd320a205b5a0dfc6dd1b321d9baedeba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0db5ecd7759d974ae1b0d8065b3b8bd4d7957fae16eb5bab2d71c2fe69243816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0db5ecd7759d974ae1b0d8065b3b8bd4d7957fae16eb5bab2d71c2fe69243816\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba9b3278c208fc7aac630ca7a0daa3de0823f85d0799d6f1f278da80ae14c862\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba9b3278c208fc7aac630ca7a0daa3de0823f85d0799d6f1f278da80ae14c862\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79b8f6c016b43c3ac3ba7b22cca9a6b72c262cb17ddd6236c5d4970ecc09680c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79b8f6c016b43c3ac3ba7b22cca9a6b72c262cb17ddd6236c5d4970ecc09680c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38c4f218eb2ca7ea0c92bd0ae2ae9efeaa40746751608fbcf7f6c3f8d9e3c758\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38c4f218eb2ca7ea0c92bd0ae2ae9efeaa40746751608fbcf7f6c3f8d9e3c758\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d66ad9831d6029c8c350054525a00069e869d39c1b5638e06e9b772666bd8630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d66ad9831d6029c8c350054525a00069e869d39c1b5638e06e9b772666bd8630\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gjtnh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:08Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.144289 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-9rmnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f9bd12b-22ee-44ed-a8b9-c100d043f691\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2fgn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2fgn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:28Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-9rmnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:08Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.152625 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86d3005cd61c0973659e1c1c0d3ef1ddd6fcf95bede5465f4f2099e64a74bb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:08Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.160620 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:08Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.169017 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x94nb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:12:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:12:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49ce055ec8090d6957cb7e23fe75edb82c0c2d7f87e7203d280606368dd3179f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75d52224135b75133fc000611cd179240b9174f6492d3263341716c7e42ccb15\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-21T11:12:02Z\\\",\\\"message\\\":\\\"2026-01-21T11:11:17+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_c9e2a0c0-e545-4ddc-b22f-97601d5900ca\\\\n2026-01-21T11:11:17+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_c9e2a0c0-e545-4ddc-b22f-97601d5900ca to /host/opt/cni/bin/\\\\n2026-01-21T11:11:17Z [verbose] multus-daemon started\\\\n2026-01-21T11:11:17Z [verbose] Readiness Indicator file check\\\\n2026-01-21T11:12:02Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:12:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgr8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x94nb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:08Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.176713 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33f3d922-4ffe-409b-a49a-d88c85898260\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a9e57b39eed8300107e944b9fcf6a2ff1e75991f8f37387a3de615b14f9bc78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpwst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b1b37c6d2994021cfc30f85e92c40ae35677113b17aa066d52bc7a76ae5c153\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpwst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zqjg9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:08Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.185071 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d87422473838f5218c9fb646ab2515f2f42592754ffdf07b19f7a8e093fa48a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:08Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.192857 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://516c009d4720ab5b393c7482872f12fed22d4f4414bd1f3400411c0e8da88ce4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee18d293036971832c8d9c3f7b9eca84ab779ff6768a2db3c6f7bdcf3433cdef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:08Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.197059 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.197081 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.197091 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.197103 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.197112 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:08Z","lastTransitionTime":"2026-01-21T11:12:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.205469 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"26b1108a-4b98-4e39-a3cd-e0d055089fd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc319bbc06568cc811ded2b418b04b986a31f321fed492e8be50104af79bb30b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cee9cc9eec3fc815f1621a22a37805691288dbb60dd3ff64194a84be62abcddc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc4c96e33e069f7d08d33caf92cd0987668506f931fd134b246d5ce5e934f95c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ddbb0e5ae02e9d5d82a80ead7efc91f1b8e557a9fb6f39940e5530fae8de2d38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://64a9065254f889b1c43d63759283335bf78fc4fa1531a4baae472c06c075dae3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e94ef8a8f47dc75ecfbab2e3fb8e7097c1b73eed32d7ca947f11ea198cea880\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86a6881f109dab1f910fe505a21c7c800e587de4dded2635147997c51c05c483\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2c655fe698872c3a7804fed5e2e75f300110a816eae4412891f3433762ed0161\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-21T11:11:40Z\\\",\\\"message\\\":\\\"ps://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:40Z is after 2025-08-24T17:21:41Z]\\\\nI0121 11:11:40.696180 6500 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf\\\\nI0121 11:11:40.696191 6500 obj_retry.go:303] Retry object setup: *v1.Pod openshift-kube-apiserver/kube-apiserver-crc\\\\nI0121 11:11:40.696193 6500 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf in node crc\\\\nI0121 11:11:40.696185 6500 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-apiserver-operator/metrics]} name:Service_openshift-apiserver-operator/metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.38:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {8b82f026-5975-4a1b-bb18-08d5d51147ec}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0121 11:11:40.696197 6500 obj_retry.go:365] Adding new object: *v1.Pod openshift-kube-apiserver/kube\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:40Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:12:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948e0497eaffa7b0d84b4085c9ed8d987eb2c6527c9a82c10e576ed9c9c1e21b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-htkvk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:08Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.213009 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-np2v8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4698abe4-0f33-4349-a2a4-614de80ae21b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://191328471bb6f2b34cde2eb927da697619826bf0b6174ecb583c8c25d27a6654\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n9qbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://050bd5bdded2162d5e9dac81713f4ed63cb575d1776e828b2f98e50d89f11b04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n9qbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-np2v8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:08Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.299045 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.299076 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.299087 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.299099 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.299107 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:08Z","lastTransitionTime":"2026-01-21T11:12:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.317309 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-htkvk_26b1108a-4b98-4e39-a3cd-e0d055089fd8/ovnkube-controller/3.log" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.317987 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-htkvk_26b1108a-4b98-4e39-a3cd-e0d055089fd8/ovnkube-controller/2.log" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.320054 4824 generic.go:334] "Generic (PLEG): container finished" podID="26b1108a-4b98-4e39-a3cd-e0d055089fd8" containerID="86a6881f109dab1f910fe505a21c7c800e587de4dded2635147997c51c05c483" exitCode=1 Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.320084 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" event={"ID":"26b1108a-4b98-4e39-a3cd-e0d055089fd8","Type":"ContainerDied","Data":"86a6881f109dab1f910fe505a21c7c800e587de4dded2635147997c51c05c483"} Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.320112 4824 scope.go:117] "RemoveContainer" containerID="2c655fe698872c3a7804fed5e2e75f300110a816eae4412891f3433762ed0161" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.320613 4824 scope.go:117] "RemoveContainer" containerID="86a6881f109dab1f910fe505a21c7c800e587de4dded2635147997c51c05c483" Jan 21 11:12:08 crc kubenswrapper[4824]: E0121 11:12:08.320779 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-htkvk_openshift-ovn-kubernetes(26b1108a-4b98-4e39-a3cd-e0d055089fd8)\"" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" podUID="26b1108a-4b98-4e39-a3cd-e0d055089fd8" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.332183 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86d3005cd61c0973659e1c1c0d3ef1ddd6fcf95bede5465f4f2099e64a74bb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:08Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.340632 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:08Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.349341 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x94nb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:12:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:12:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49ce055ec8090d6957cb7e23fe75edb82c0c2d7f87e7203d280606368dd3179f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75d52224135b75133fc000611cd179240b9174f6492d3263341716c7e42ccb15\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-21T11:12:02Z\\\",\\\"message\\\":\\\"2026-01-21T11:11:17+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_c9e2a0c0-e545-4ddc-b22f-97601d5900ca\\\\n2026-01-21T11:11:17+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_c9e2a0c0-e545-4ddc-b22f-97601d5900ca to /host/opt/cni/bin/\\\\n2026-01-21T11:11:17Z [verbose] multus-daemon started\\\\n2026-01-21T11:11:17Z [verbose] Readiness Indicator file check\\\\n2026-01-21T11:12:02Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:12:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgr8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x94nb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:08Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.357211 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33f3d922-4ffe-409b-a49a-d88c85898260\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a9e57b39eed8300107e944b9fcf6a2ff1e75991f8f37387a3de615b14f9bc78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpwst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b1b37c6d2994021cfc30f85e92c40ae35677113b17aa066d52bc7a76ae5c153\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpwst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zqjg9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:08Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.366624 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gjtnh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a8d2cac-2d0f-4894-9bc0-2abd81699c47\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67c122a26811deaa9b70d19a1a65daee318ca3af2c626712482179d72fe67875\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2239186f1de46481cebfcafb1cab85cd320a205b5a0dfc6dd1b321d9baedeba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b2239186f1de46481cebfcafb1cab85cd320a205b5a0dfc6dd1b321d9baedeba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0db5ecd7759d974ae1b0d8065b3b8bd4d7957fae16eb5bab2d71c2fe69243816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0db5ecd7759d974ae1b0d8065b3b8bd4d7957fae16eb5bab2d71c2fe69243816\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba9b3278c208fc7aac630ca7a0daa3de0823f85d0799d6f1f278da80ae14c862\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba9b3278c208fc7aac630ca7a0daa3de0823f85d0799d6f1f278da80ae14c862\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79b8f6c016b43c3ac3ba7b22cca9a6b72c262cb17ddd6236c5d4970ecc09680c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79b8f6c016b43c3ac3ba7b22cca9a6b72c262cb17ddd6236c5d4970ecc09680c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38c4f218eb2ca7ea0c92bd0ae2ae9efeaa40746751608fbcf7f6c3f8d9e3c758\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38c4f218eb2ca7ea0c92bd0ae2ae9efeaa40746751608fbcf7f6c3f8d9e3c758\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d66ad9831d6029c8c350054525a00069e869d39c1b5638e06e9b772666bd8630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d66ad9831d6029c8c350054525a00069e869d39c1b5638e06e9b772666bd8630\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gjtnh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:08Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.373335 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-9rmnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f9bd12b-22ee-44ed-a8b9-c100d043f691\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2fgn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2fgn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:28Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-9rmnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:08Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.380563 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d87422473838f5218c9fb646ab2515f2f42592754ffdf07b19f7a8e093fa48a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:08Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.388262 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://516c009d4720ab5b393c7482872f12fed22d4f4414bd1f3400411c0e8da88ce4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee18d293036971832c8d9c3f7b9eca84ab779ff6768a2db3c6f7bdcf3433cdef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:08Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.400464 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"26b1108a-4b98-4e39-a3cd-e0d055089fd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc319bbc06568cc811ded2b418b04b986a31f321fed492e8be50104af79bb30b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cee9cc9eec3fc815f1621a22a37805691288dbb60dd3ff64194a84be62abcddc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc4c96e33e069f7d08d33caf92cd0987668506f931fd134b246d5ce5e934f95c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ddbb0e5ae02e9d5d82a80ead7efc91f1b8e557a9fb6f39940e5530fae8de2d38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://64a9065254f889b1c43d63759283335bf78fc4fa1531a4baae472c06c075dae3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e94ef8a8f47dc75ecfbab2e3fb8e7097c1b73eed32d7ca947f11ea198cea880\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86a6881f109dab1f910fe505a21c7c800e587de4dded2635147997c51c05c483\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2c655fe698872c3a7804fed5e2e75f300110a816eae4412891f3433762ed0161\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-21T11:11:40Z\\\",\\\"message\\\":\\\"ps://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:11:40Z is after 2025-08-24T17:21:41Z]\\\\nI0121 11:11:40.696180 6500 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf\\\\nI0121 11:11:40.696191 6500 obj_retry.go:303] Retry object setup: *v1.Pod openshift-kube-apiserver/kube-apiserver-crc\\\\nI0121 11:11:40.696193 6500 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf in node crc\\\\nI0121 11:11:40.696185 6500 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-apiserver-operator/metrics]} name:Service_openshift-apiserver-operator/metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.38:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {8b82f026-5975-4a1b-bb18-08d5d51147ec}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0121 11:11:40.696197 6500 obj_retry.go:365] Adding new object: *v1.Pod openshift-kube-apiserver/kube\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:40Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86a6881f109dab1f910fe505a21c7c800e587de4dded2635147997c51c05c483\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-21T11:12:07Z\\\",\\\"message\\\":\\\"UUIDName:}]\\\\nI0121 11:12:07.749346 6924 services_controller.go:453] Built service openshift-console/console template LB for network=default: []services.LB{}\\\\nI0121 11:12:07.749353 6924 services_controller.go:454] Service openshift-console/console for network=default has 1 cluster-wide, 0 per-node configs, 0 template configs, making 1 (cluster) 0 (per node) and 0 (template) load balancers\\\\nI0121 11:12:07.747279 6924 services_controller.go:434] Service openshift-kube-storage-version-migrator-operator/metrics retrieved from lister for network=default: \\\\u0026Service{ObjectMeta:{metrics openshift-kube-storage-version-migrator-operator e1639a86-fb7f-46de-9d5e-4aee16dccea1 4372 0 2025-02-23 05:12:25 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[app:kube-storage-version-migrator-operator] map[include.release.openshift.io/hypershift:true include.release.openshift.io/ibm-cloud-managed:true include.release.openshift.io/self-managed-high-availability:true service.alpha.openshift.io/serving-cert-secret-name:serving-cert service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc007778417 \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:htt\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T11:12:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948e0497eaffa7b0d84b4085c9ed8d987eb2c6527c9a82c10e576ed9c9c1e21b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-htkvk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:08Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.401598 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.401624 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.401633 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.401645 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.401653 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:08Z","lastTransitionTime":"2026-01-21T11:12:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.407544 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-np2v8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4698abe4-0f33-4349-a2a4-614de80ae21b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://191328471bb6f2b34cde2eb927da697619826bf0b6174ecb583c8c25d27a6654\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n9qbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://050bd5bdded2162d5e9dac81713f4ed63cb575d1776e828b2f98e50d89f11b04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n9qbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-np2v8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:08Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.414315 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jr64s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9768c60d-5bee-4286-bced-076e584b04e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91158f14d2e98ac9f875a5e7101df3107ef28aa5fc5d62218f460af98ff920fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flrnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:17Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jr64s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:08Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.423354 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d31ab61-f94c-4f49-85d0-c9e91b83069b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a66b5c33884af00408dc35cf02f08ebd8570bd695f9f1fe558bf54fd06930431\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7514786602746dabfa690da48d48e233fe68d6dd6674aa5ecd489c0b3185f08\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61b14b6846f94cdeba8f1f61b3754e7663041f92c70356c4ee6cbb988a2e680f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a7023cfe4c098bcdd9c0477a8187f4ed7a0d2abdbe1e143436863948f441b4f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af11ad95bcf0c6bbd5fa4ab99bb592ce1c87f40d29c5ade7832414315820a569\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW0121 11:11:14.604290 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0121 11:11:14.604572 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0121 11:11:14.605875 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3284452394/tls.crt::/tmp/serving-cert-3284452394/tls.key\\\\\\\"\\\\nI0121 11:11:14.923427 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0121 11:11:14.928141 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0121 11:11:14.928162 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0121 11:11:14.928179 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0121 11:11:14.928184 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0121 11:11:14.933154 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0121 11:11:14.933230 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 11:11:14.933259 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 11:11:14.933282 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0121 11:11:14.933299 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0121 11:11:14.933315 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0121 11:11:14.933331 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0121 11:11:14.933177 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0121 11:11:14.934666 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48070c09fad511108ff22e44211cf26911509b8406d66948985cbfc3009d8233\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:08Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.430909 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9bb99c0a-683b-44da-9c0e-e8d284e87d53\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05e076febd697ed802ce5c7ee92a8c2c636a1a2c6609ad7eb01f17dace247c4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43529d1b9e1e5e22b0d56b023946119da8aacbae54b68c26ff0a1a82dc890c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ea232ad87393685aa323ce6393ff3869f8f8d423c2406a501c69c355535adfa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc6c7eff9e17b8a8ecc88bbfe703b5a01dbf1e02d342cede4f6a263a6526af43\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:08Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.438762 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e570fff1-6227-44d6-b71f-ee377b1099ae\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4d41492c3802a8168fffd249c15aeb9c902a5c7e01550e1aa6f9b769adfdc4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2690a1ecbc8fe1290e68bba2695a3edd833b5da933fb71e8ddc926a1340a9a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b85db961e1e078e1f54323c79a99c9413365256ac2fce411db7932bd0580ab6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://595783a1454735bd33f4862180b60d99a4e1b6fae1e735f2b7328391951bf422\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://595783a1454735bd33f4862180b60d99a4e1b6fae1e735f2b7328391951bf422\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:08Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.447244 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:08Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.453559 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lv4rn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1267539-9adf-466f-8027-8ede009d19a0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eed1b884aa53e1ee318169a0cc2c9ad10130aa363ae4ac55e4bf065237756f53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbmfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lv4rn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:08Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.466050 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06f56d86-ec6f-4b57-b314-f508c0dd4d9f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d25c94f1ae0bbabb733f0e1ea1d1ba39562a92af73047a39aacc03013e239681\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca06234486b59cf990c69d9501e23f935328e2ba21013cd71ed64831ea749bcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b63e45d66bcd11ff3479cc3ec6b450017da8d21d36e967b8d496ae0f24b0ee9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52758bcd31f2ba64edf03adfd4d1be6428015abff5b64c7c029fa787085f9c29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2736e73342d2adb175b0721dd4f40cc42d26fca06d5eb3334927518531c3b107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b35ad7dd0570c191eb0080739304f3cd43f17b3d1dddf895549441ebe50deaa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b35ad7dd0570c191eb0080739304f3cd43f17b3d1dddf895549441ebe50deaa6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://80cd5ea5e6fc19a88c1aab5c61f827f103ab798a5c84da8dc4d45cea86d5aacb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80cd5ea5e6fc19a88c1aab5c61f827f103ab798a5c84da8dc4d45cea86d5aacb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://234d8b914e6a88480021647ae65d74e9b21c55b6788312a094d00c8a42677d48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://234d8b914e6a88480021647ae65d74e9b21c55b6788312a094d00c8a42677d48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:08Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.474267 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:08Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.504170 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.504199 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.504209 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.504225 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.504234 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:08Z","lastTransitionTime":"2026-01-21T11:12:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.606530 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.606564 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.606573 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.606586 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.606595 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:08Z","lastTransitionTime":"2026-01-21T11:12:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.697791 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.697840 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.697852 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.697866 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.697876 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:08Z","lastTransitionTime":"2026-01-21T11:12:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:08 crc kubenswrapper[4824]: E0121 11:12:08.708724 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:12:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:12:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:12:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:12:08Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:12:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:12:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:12:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:12:08Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"695c676c-720c-4921-9e71-56b96f2fb46b\\\",\\\"systemUUID\\\":\\\"d00624f5-786f-41dc-bdc1-a1875f167d65\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:08Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.711741 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.711789 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.711803 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.711821 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.711830 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:08Z","lastTransitionTime":"2026-01-21T11:12:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:08 crc kubenswrapper[4824]: E0121 11:12:08.721278 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:12:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:12:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:12:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:12:08Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:12:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:12:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:12:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:12:08Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"695c676c-720c-4921-9e71-56b96f2fb46b\\\",\\\"systemUUID\\\":\\\"d00624f5-786f-41dc-bdc1-a1875f167d65\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:08Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.724475 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.724507 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.724518 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.724532 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.724542 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:08Z","lastTransitionTime":"2026-01-21T11:12:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:08 crc kubenswrapper[4824]: E0121 11:12:08.733855 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:12:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:12:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:12:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:12:08Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:12:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:12:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:12:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:12:08Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"695c676c-720c-4921-9e71-56b96f2fb46b\\\",\\\"systemUUID\\\":\\\"d00624f5-786f-41dc-bdc1-a1875f167d65\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:08Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.737561 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.737600 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.737610 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.737626 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.737638 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:08Z","lastTransitionTime":"2026-01-21T11:12:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:08 crc kubenswrapper[4824]: E0121 11:12:08.747256 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:12:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:12:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:12:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:12:08Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:12:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:12:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:12:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:12:08Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"695c676c-720c-4921-9e71-56b96f2fb46b\\\",\\\"systemUUID\\\":\\\"d00624f5-786f-41dc-bdc1-a1875f167d65\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:08Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.749612 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.749636 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.749648 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.749659 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.749667 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:08Z","lastTransitionTime":"2026-01-21T11:12:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:08 crc kubenswrapper[4824]: E0121 11:12:08.760142 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:12:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:12:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:12:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:12:08Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:12:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:12:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:12:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:12:08Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"695c676c-720c-4921-9e71-56b96f2fb46b\\\",\\\"systemUUID\\\":\\\"d00624f5-786f-41dc-bdc1-a1875f167d65\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:08Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:08 crc kubenswrapper[4824]: E0121 11:12:08.760253 4824 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.761596 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.761634 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.761644 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.761675 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.761687 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:08Z","lastTransitionTime":"2026-01-21T11:12:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.863842 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.863891 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.863901 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.863917 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.863930 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:08Z","lastTransitionTime":"2026-01-21T11:12:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.966363 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.966396 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.966408 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.966437 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:08 crc kubenswrapper[4824]: I0121 11:12:08.966450 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:08Z","lastTransitionTime":"2026-01-21T11:12:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:09 crc kubenswrapper[4824]: I0121 11:12:09.022317 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-05 20:24:07.499506491 +0000 UTC Jan 21 11:12:09 crc kubenswrapper[4824]: I0121 11:12:09.048555 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 11:12:09 crc kubenswrapper[4824]: I0121 11:12:09.048692 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 11:12:09 crc kubenswrapper[4824]: E0121 11:12:09.048799 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 21 11:12:09 crc kubenswrapper[4824]: E0121 11:12:09.048928 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 21 11:12:09 crc kubenswrapper[4824]: I0121 11:12:09.067893 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:09 crc kubenswrapper[4824]: I0121 11:12:09.067922 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:09 crc kubenswrapper[4824]: I0121 11:12:09.067933 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:09 crc kubenswrapper[4824]: I0121 11:12:09.067945 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:09 crc kubenswrapper[4824]: I0121 11:12:09.067977 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:09Z","lastTransitionTime":"2026-01-21T11:12:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:09 crc kubenswrapper[4824]: I0121 11:12:09.170730 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:09 crc kubenswrapper[4824]: I0121 11:12:09.170766 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:09 crc kubenswrapper[4824]: I0121 11:12:09.170777 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:09 crc kubenswrapper[4824]: I0121 11:12:09.170789 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:09 crc kubenswrapper[4824]: I0121 11:12:09.170798 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:09Z","lastTransitionTime":"2026-01-21T11:12:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:09 crc kubenswrapper[4824]: I0121 11:12:09.272921 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:09 crc kubenswrapper[4824]: I0121 11:12:09.272950 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:09 crc kubenswrapper[4824]: I0121 11:12:09.272976 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:09 crc kubenswrapper[4824]: I0121 11:12:09.272992 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:09 crc kubenswrapper[4824]: I0121 11:12:09.273003 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:09Z","lastTransitionTime":"2026-01-21T11:12:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:09 crc kubenswrapper[4824]: I0121 11:12:09.324633 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-htkvk_26b1108a-4b98-4e39-a3cd-e0d055089fd8/ovnkube-controller/3.log" Jan 21 11:12:09 crc kubenswrapper[4824]: I0121 11:12:09.328307 4824 scope.go:117] "RemoveContainer" containerID="86a6881f109dab1f910fe505a21c7c800e587de4dded2635147997c51c05c483" Jan 21 11:12:09 crc kubenswrapper[4824]: E0121 11:12:09.328462 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-htkvk_openshift-ovn-kubernetes(26b1108a-4b98-4e39-a3cd-e0d055089fd8)\"" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" podUID="26b1108a-4b98-4e39-a3cd-e0d055089fd8" Jan 21 11:12:09 crc kubenswrapper[4824]: I0121 11:12:09.344835 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06f56d86-ec6f-4b57-b314-f508c0dd4d9f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d25c94f1ae0bbabb733f0e1ea1d1ba39562a92af73047a39aacc03013e239681\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca06234486b59cf990c69d9501e23f935328e2ba21013cd71ed64831ea749bcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b63e45d66bcd11ff3479cc3ec6b450017da8d21d36e967b8d496ae0f24b0ee9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52758bcd31f2ba64edf03adfd4d1be6428015abff5b64c7c029fa787085f9c29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2736e73342d2adb175b0721dd4f40cc42d26fca06d5eb3334927518531c3b107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b35ad7dd0570c191eb0080739304f3cd43f17b3d1dddf895549441ebe50deaa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b35ad7dd0570c191eb0080739304f3cd43f17b3d1dddf895549441ebe50deaa6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://80cd5ea5e6fc19a88c1aab5c61f827f103ab798a5c84da8dc4d45cea86d5aacb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80cd5ea5e6fc19a88c1aab5c61f827f103ab798a5c84da8dc4d45cea86d5aacb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://234d8b914e6a88480021647ae65d74e9b21c55b6788312a094d00c8a42677d48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://234d8b914e6a88480021647ae65d74e9b21c55b6788312a094d00c8a42677d48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:09Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:09 crc kubenswrapper[4824]: I0121 11:12:09.354983 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:09Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:09 crc kubenswrapper[4824]: I0121 11:12:09.364347 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86d3005cd61c0973659e1c1c0d3ef1ddd6fcf95bede5465f4f2099e64a74bb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:09Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:09 crc kubenswrapper[4824]: I0121 11:12:09.373596 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:09Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:09 crc kubenswrapper[4824]: I0121 11:12:09.375912 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:09 crc kubenswrapper[4824]: I0121 11:12:09.376001 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:09 crc kubenswrapper[4824]: I0121 11:12:09.376017 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:09 crc kubenswrapper[4824]: I0121 11:12:09.376038 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:09 crc kubenswrapper[4824]: I0121 11:12:09.376053 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:09Z","lastTransitionTime":"2026-01-21T11:12:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:09 crc kubenswrapper[4824]: I0121 11:12:09.385805 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x94nb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:12:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:12:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49ce055ec8090d6957cb7e23fe75edb82c0c2d7f87e7203d280606368dd3179f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75d52224135b75133fc000611cd179240b9174f6492d3263341716c7e42ccb15\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-21T11:12:02Z\\\",\\\"message\\\":\\\"2026-01-21T11:11:17+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_c9e2a0c0-e545-4ddc-b22f-97601d5900ca\\\\n2026-01-21T11:11:17+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_c9e2a0c0-e545-4ddc-b22f-97601d5900ca to /host/opt/cni/bin/\\\\n2026-01-21T11:11:17Z [verbose] multus-daemon started\\\\n2026-01-21T11:11:17Z [verbose] Readiness Indicator file check\\\\n2026-01-21T11:12:02Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:12:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgr8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x94nb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:09Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:09 crc kubenswrapper[4824]: I0121 11:12:09.396500 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33f3d922-4ffe-409b-a49a-d88c85898260\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a9e57b39eed8300107e944b9fcf6a2ff1e75991f8f37387a3de615b14f9bc78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpwst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b1b37c6d2994021cfc30f85e92c40ae35677113b17aa066d52bc7a76ae5c153\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpwst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zqjg9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:09Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:09 crc kubenswrapper[4824]: I0121 11:12:09.407529 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gjtnh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a8d2cac-2d0f-4894-9bc0-2abd81699c47\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67c122a26811deaa9b70d19a1a65daee318ca3af2c626712482179d72fe67875\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2239186f1de46481cebfcafb1cab85cd320a205b5a0dfc6dd1b321d9baedeba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b2239186f1de46481cebfcafb1cab85cd320a205b5a0dfc6dd1b321d9baedeba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0db5ecd7759d974ae1b0d8065b3b8bd4d7957fae16eb5bab2d71c2fe69243816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0db5ecd7759d974ae1b0d8065b3b8bd4d7957fae16eb5bab2d71c2fe69243816\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba9b3278c208fc7aac630ca7a0daa3de0823f85d0799d6f1f278da80ae14c862\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba9b3278c208fc7aac630ca7a0daa3de0823f85d0799d6f1f278da80ae14c862\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79b8f6c016b43c3ac3ba7b22cca9a6b72c262cb17ddd6236c5d4970ecc09680c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79b8f6c016b43c3ac3ba7b22cca9a6b72c262cb17ddd6236c5d4970ecc09680c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38c4f218eb2ca7ea0c92bd0ae2ae9efeaa40746751608fbcf7f6c3f8d9e3c758\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38c4f218eb2ca7ea0c92bd0ae2ae9efeaa40746751608fbcf7f6c3f8d9e3c758\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d66ad9831d6029c8c350054525a00069e869d39c1b5638e06e9b772666bd8630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d66ad9831d6029c8c350054525a00069e869d39c1b5638e06e9b772666bd8630\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gjtnh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:09Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:09 crc kubenswrapper[4824]: I0121 11:12:09.415605 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-9rmnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f9bd12b-22ee-44ed-a8b9-c100d043f691\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2fgn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2fgn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:28Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-9rmnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:09Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:09 crc kubenswrapper[4824]: I0121 11:12:09.423533 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d87422473838f5218c9fb646ab2515f2f42592754ffdf07b19f7a8e093fa48a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:09Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:09 crc kubenswrapper[4824]: I0121 11:12:09.432798 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://516c009d4720ab5b393c7482872f12fed22d4f4414bd1f3400411c0e8da88ce4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee18d293036971832c8d9c3f7b9eca84ab779ff6768a2db3c6f7bdcf3433cdef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:09Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:09 crc kubenswrapper[4824]: I0121 11:12:09.446048 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"26b1108a-4b98-4e39-a3cd-e0d055089fd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc319bbc06568cc811ded2b418b04b986a31f321fed492e8be50104af79bb30b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cee9cc9eec3fc815f1621a22a37805691288dbb60dd3ff64194a84be62abcddc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc4c96e33e069f7d08d33caf92cd0987668506f931fd134b246d5ce5e934f95c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ddbb0e5ae02e9d5d82a80ead7efc91f1b8e557a9fb6f39940e5530fae8de2d38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://64a9065254f889b1c43d63759283335bf78fc4fa1531a4baae472c06c075dae3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e94ef8a8f47dc75ecfbab2e3fb8e7097c1b73eed32d7ca947f11ea198cea880\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86a6881f109dab1f910fe505a21c7c800e587de4dded2635147997c51c05c483\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86a6881f109dab1f910fe505a21c7c800e587de4dded2635147997c51c05c483\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-21T11:12:07Z\\\",\\\"message\\\":\\\"UUIDName:}]\\\\nI0121 11:12:07.749346 6924 services_controller.go:453] Built service openshift-console/console template LB for network=default: []services.LB{}\\\\nI0121 11:12:07.749353 6924 services_controller.go:454] Service openshift-console/console for network=default has 1 cluster-wide, 0 per-node configs, 0 template configs, making 1 (cluster) 0 (per node) and 0 (template) load balancers\\\\nI0121 11:12:07.747279 6924 services_controller.go:434] Service openshift-kube-storage-version-migrator-operator/metrics retrieved from lister for network=default: \\\\u0026Service{ObjectMeta:{metrics openshift-kube-storage-version-migrator-operator e1639a86-fb7f-46de-9d5e-4aee16dccea1 4372 0 2025-02-23 05:12:25 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[app:kube-storage-version-migrator-operator] map[include.release.openshift.io/hypershift:true include.release.openshift.io/ibm-cloud-managed:true include.release.openshift.io/self-managed-high-availability:true service.alpha.openshift.io/serving-cert-secret-name:serving-cert service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc007778417 \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:htt\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T11:12:07Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-htkvk_openshift-ovn-kubernetes(26b1108a-4b98-4e39-a3cd-e0d055089fd8)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948e0497eaffa7b0d84b4085c9ed8d987eb2c6527c9a82c10e576ed9c9c1e21b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-htkvk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:09Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:09 crc kubenswrapper[4824]: I0121 11:12:09.453481 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-np2v8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4698abe4-0f33-4349-a2a4-614de80ae21b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://191328471bb6f2b34cde2eb927da697619826bf0b6174ecb583c8c25d27a6654\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n9qbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://050bd5bdded2162d5e9dac81713f4ed63cb575d1776e828b2f98e50d89f11b04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n9qbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-np2v8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:09Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:09 crc kubenswrapper[4824]: I0121 11:12:09.462990 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d31ab61-f94c-4f49-85d0-c9e91b83069b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a66b5c33884af00408dc35cf02f08ebd8570bd695f9f1fe558bf54fd06930431\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7514786602746dabfa690da48d48e233fe68d6dd6674aa5ecd489c0b3185f08\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61b14b6846f94cdeba8f1f61b3754e7663041f92c70356c4ee6cbb988a2e680f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a7023cfe4c098bcdd9c0477a8187f4ed7a0d2abdbe1e143436863948f441b4f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af11ad95bcf0c6bbd5fa4ab99bb592ce1c87f40d29c5ade7832414315820a569\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW0121 11:11:14.604290 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0121 11:11:14.604572 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0121 11:11:14.605875 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3284452394/tls.crt::/tmp/serving-cert-3284452394/tls.key\\\\\\\"\\\\nI0121 11:11:14.923427 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0121 11:11:14.928141 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0121 11:11:14.928162 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0121 11:11:14.928179 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0121 11:11:14.928184 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0121 11:11:14.933154 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0121 11:11:14.933230 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 11:11:14.933259 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 11:11:14.933282 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0121 11:11:14.933299 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0121 11:11:14.933315 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0121 11:11:14.933331 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0121 11:11:14.933177 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0121 11:11:14.934666 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48070c09fad511108ff22e44211cf26911509b8406d66948985cbfc3009d8233\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:09Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:09 crc kubenswrapper[4824]: I0121 11:12:09.472117 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9bb99c0a-683b-44da-9c0e-e8d284e87d53\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05e076febd697ed802ce5c7ee92a8c2c636a1a2c6609ad7eb01f17dace247c4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43529d1b9e1e5e22b0d56b023946119da8aacbae54b68c26ff0a1a82dc890c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ea232ad87393685aa323ce6393ff3869f8f8d423c2406a501c69c355535adfa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc6c7eff9e17b8a8ecc88bbfe703b5a01dbf1e02d342cede4f6a263a6526af43\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:09Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:09 crc kubenswrapper[4824]: I0121 11:12:09.478353 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:09 crc kubenswrapper[4824]: I0121 11:12:09.478388 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:09 crc kubenswrapper[4824]: I0121 11:12:09.478399 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:09 crc kubenswrapper[4824]: I0121 11:12:09.478429 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:09 crc kubenswrapper[4824]: I0121 11:12:09.478441 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:09Z","lastTransitionTime":"2026-01-21T11:12:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:09 crc kubenswrapper[4824]: I0121 11:12:09.480016 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e570fff1-6227-44d6-b71f-ee377b1099ae\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4d41492c3802a8168fffd249c15aeb9c902a5c7e01550e1aa6f9b769adfdc4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2690a1ecbc8fe1290e68bba2695a3edd833b5da933fb71e8ddc926a1340a9a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b85db961e1e078e1f54323c79a99c9413365256ac2fce411db7932bd0580ab6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://595783a1454735bd33f4862180b60d99a4e1b6fae1e735f2b7328391951bf422\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://595783a1454735bd33f4862180b60d99a4e1b6fae1e735f2b7328391951bf422\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:09Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:09 crc kubenswrapper[4824]: I0121 11:12:09.488573 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:09Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:09 crc kubenswrapper[4824]: I0121 11:12:09.496329 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lv4rn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1267539-9adf-466f-8027-8ede009d19a0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eed1b884aa53e1ee318169a0cc2c9ad10130aa363ae4ac55e4bf065237756f53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbmfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lv4rn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:09Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:09 crc kubenswrapper[4824]: I0121 11:12:09.503053 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jr64s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9768c60d-5bee-4286-bced-076e584b04e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91158f14d2e98ac9f875a5e7101df3107ef28aa5fc5d62218f460af98ff920fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flrnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:17Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jr64s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:09Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:09 crc kubenswrapper[4824]: I0121 11:12:09.581796 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:09 crc kubenswrapper[4824]: I0121 11:12:09.581844 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:09 crc kubenswrapper[4824]: I0121 11:12:09.581864 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:09 crc kubenswrapper[4824]: I0121 11:12:09.581888 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:09 crc kubenswrapper[4824]: I0121 11:12:09.581911 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:09Z","lastTransitionTime":"2026-01-21T11:12:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:09 crc kubenswrapper[4824]: I0121 11:12:09.684329 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:09 crc kubenswrapper[4824]: I0121 11:12:09.684358 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:09 crc kubenswrapper[4824]: I0121 11:12:09.684369 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:09 crc kubenswrapper[4824]: I0121 11:12:09.684384 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:09 crc kubenswrapper[4824]: I0121 11:12:09.684394 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:09Z","lastTransitionTime":"2026-01-21T11:12:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:09 crc kubenswrapper[4824]: I0121 11:12:09.786210 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:09 crc kubenswrapper[4824]: I0121 11:12:09.786262 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:09 crc kubenswrapper[4824]: I0121 11:12:09.786276 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:09 crc kubenswrapper[4824]: I0121 11:12:09.786296 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:09 crc kubenswrapper[4824]: I0121 11:12:09.786310 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:09Z","lastTransitionTime":"2026-01-21T11:12:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:09 crc kubenswrapper[4824]: I0121 11:12:09.888702 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:09 crc kubenswrapper[4824]: I0121 11:12:09.888748 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:09 crc kubenswrapper[4824]: I0121 11:12:09.888759 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:09 crc kubenswrapper[4824]: I0121 11:12:09.888777 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:09 crc kubenswrapper[4824]: I0121 11:12:09.888786 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:09Z","lastTransitionTime":"2026-01-21T11:12:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:09 crc kubenswrapper[4824]: I0121 11:12:09.990712 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:09 crc kubenswrapper[4824]: I0121 11:12:09.990744 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:09 crc kubenswrapper[4824]: I0121 11:12:09.990752 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:09 crc kubenswrapper[4824]: I0121 11:12:09.990766 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:09 crc kubenswrapper[4824]: I0121 11:12:09.990775 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:09Z","lastTransitionTime":"2026-01-21T11:12:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:10 crc kubenswrapper[4824]: I0121 11:12:10.022594 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-22 02:45:45.467635292 +0000 UTC Jan 21 11:12:10 crc kubenswrapper[4824]: I0121 11:12:10.048324 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9rmnw" Jan 21 11:12:10 crc kubenswrapper[4824]: E0121 11:12:10.048438 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9rmnw" podUID="8f9bd12b-22ee-44ed-a8b9-c100d043f691" Jan 21 11:12:10 crc kubenswrapper[4824]: I0121 11:12:10.048468 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 11:12:10 crc kubenswrapper[4824]: E0121 11:12:10.048565 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 21 11:12:10 crc kubenswrapper[4824]: I0121 11:12:10.093040 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:10 crc kubenswrapper[4824]: I0121 11:12:10.093149 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:10 crc kubenswrapper[4824]: I0121 11:12:10.093221 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:10 crc kubenswrapper[4824]: I0121 11:12:10.093275 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:10 crc kubenswrapper[4824]: I0121 11:12:10.093326 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:10Z","lastTransitionTime":"2026-01-21T11:12:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:10 crc kubenswrapper[4824]: I0121 11:12:10.195312 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:10 crc kubenswrapper[4824]: I0121 11:12:10.195522 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:10 crc kubenswrapper[4824]: I0121 11:12:10.195586 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:10 crc kubenswrapper[4824]: I0121 11:12:10.195642 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:10 crc kubenswrapper[4824]: I0121 11:12:10.195692 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:10Z","lastTransitionTime":"2026-01-21T11:12:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:10 crc kubenswrapper[4824]: I0121 11:12:10.297481 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:10 crc kubenswrapper[4824]: I0121 11:12:10.297551 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:10 crc kubenswrapper[4824]: I0121 11:12:10.297566 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:10 crc kubenswrapper[4824]: I0121 11:12:10.297591 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:10 crc kubenswrapper[4824]: I0121 11:12:10.297603 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:10Z","lastTransitionTime":"2026-01-21T11:12:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:10 crc kubenswrapper[4824]: I0121 11:12:10.400241 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:10 crc kubenswrapper[4824]: I0121 11:12:10.400440 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:10 crc kubenswrapper[4824]: I0121 11:12:10.400624 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:10 crc kubenswrapper[4824]: I0121 11:12:10.400780 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:10 crc kubenswrapper[4824]: I0121 11:12:10.400912 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:10Z","lastTransitionTime":"2026-01-21T11:12:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:10 crc kubenswrapper[4824]: I0121 11:12:10.503276 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:10 crc kubenswrapper[4824]: I0121 11:12:10.503322 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:10 crc kubenswrapper[4824]: I0121 11:12:10.503334 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:10 crc kubenswrapper[4824]: I0121 11:12:10.503351 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:10 crc kubenswrapper[4824]: I0121 11:12:10.503363 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:10Z","lastTransitionTime":"2026-01-21T11:12:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:10 crc kubenswrapper[4824]: I0121 11:12:10.606238 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:10 crc kubenswrapper[4824]: I0121 11:12:10.606370 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:10 crc kubenswrapper[4824]: I0121 11:12:10.606680 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:10 crc kubenswrapper[4824]: I0121 11:12:10.606756 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:10 crc kubenswrapper[4824]: I0121 11:12:10.606824 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:10Z","lastTransitionTime":"2026-01-21T11:12:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:10 crc kubenswrapper[4824]: I0121 11:12:10.708152 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:10 crc kubenswrapper[4824]: I0121 11:12:10.708178 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:10 crc kubenswrapper[4824]: I0121 11:12:10.708187 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:10 crc kubenswrapper[4824]: I0121 11:12:10.708200 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:10 crc kubenswrapper[4824]: I0121 11:12:10.708209 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:10Z","lastTransitionTime":"2026-01-21T11:12:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:10 crc kubenswrapper[4824]: I0121 11:12:10.809549 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:10 crc kubenswrapper[4824]: I0121 11:12:10.809581 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:10 crc kubenswrapper[4824]: I0121 11:12:10.809590 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:10 crc kubenswrapper[4824]: I0121 11:12:10.809605 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:10 crc kubenswrapper[4824]: I0121 11:12:10.809614 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:10Z","lastTransitionTime":"2026-01-21T11:12:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:10 crc kubenswrapper[4824]: I0121 11:12:10.910701 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:10 crc kubenswrapper[4824]: I0121 11:12:10.910734 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:10 crc kubenswrapper[4824]: I0121 11:12:10.910743 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:10 crc kubenswrapper[4824]: I0121 11:12:10.910756 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:10 crc kubenswrapper[4824]: I0121 11:12:10.910764 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:10Z","lastTransitionTime":"2026-01-21T11:12:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:11 crc kubenswrapper[4824]: I0121 11:12:11.012371 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:11 crc kubenswrapper[4824]: I0121 11:12:11.012427 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:11 crc kubenswrapper[4824]: I0121 11:12:11.012436 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:11 crc kubenswrapper[4824]: I0121 11:12:11.012449 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:11 crc kubenswrapper[4824]: I0121 11:12:11.012457 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:11Z","lastTransitionTime":"2026-01-21T11:12:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:11 crc kubenswrapper[4824]: I0121 11:12:11.023708 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-29 22:47:03.820783195 +0000 UTC Jan 21 11:12:11 crc kubenswrapper[4824]: I0121 11:12:11.049129 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 11:12:11 crc kubenswrapper[4824]: I0121 11:12:11.049286 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 11:12:11 crc kubenswrapper[4824]: E0121 11:12:11.049333 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 21 11:12:11 crc kubenswrapper[4824]: E0121 11:12:11.049554 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 21 11:12:11 crc kubenswrapper[4824]: I0121 11:12:11.114823 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:11 crc kubenswrapper[4824]: I0121 11:12:11.114867 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:11 crc kubenswrapper[4824]: I0121 11:12:11.114877 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:11 crc kubenswrapper[4824]: I0121 11:12:11.114888 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:11 crc kubenswrapper[4824]: I0121 11:12:11.114898 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:11Z","lastTransitionTime":"2026-01-21T11:12:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:11 crc kubenswrapper[4824]: I0121 11:12:11.217498 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:11 crc kubenswrapper[4824]: I0121 11:12:11.217543 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:11 crc kubenswrapper[4824]: I0121 11:12:11.217553 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:11 crc kubenswrapper[4824]: I0121 11:12:11.217567 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:11 crc kubenswrapper[4824]: I0121 11:12:11.217578 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:11Z","lastTransitionTime":"2026-01-21T11:12:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:11 crc kubenswrapper[4824]: I0121 11:12:11.319363 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:11 crc kubenswrapper[4824]: I0121 11:12:11.319423 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:11 crc kubenswrapper[4824]: I0121 11:12:11.319433 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:11 crc kubenswrapper[4824]: I0121 11:12:11.319449 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:11 crc kubenswrapper[4824]: I0121 11:12:11.319461 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:11Z","lastTransitionTime":"2026-01-21T11:12:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:11 crc kubenswrapper[4824]: I0121 11:12:11.421493 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:11 crc kubenswrapper[4824]: I0121 11:12:11.421544 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:11 crc kubenswrapper[4824]: I0121 11:12:11.421553 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:11 crc kubenswrapper[4824]: I0121 11:12:11.421566 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:11 crc kubenswrapper[4824]: I0121 11:12:11.421576 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:11Z","lastTransitionTime":"2026-01-21T11:12:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:11 crc kubenswrapper[4824]: I0121 11:12:11.523674 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:11 crc kubenswrapper[4824]: I0121 11:12:11.523717 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:11 crc kubenswrapper[4824]: I0121 11:12:11.523726 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:11 crc kubenswrapper[4824]: I0121 11:12:11.523741 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:11 crc kubenswrapper[4824]: I0121 11:12:11.523751 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:11Z","lastTransitionTime":"2026-01-21T11:12:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:11 crc kubenswrapper[4824]: I0121 11:12:11.625189 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:11 crc kubenswrapper[4824]: I0121 11:12:11.625219 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:11 crc kubenswrapper[4824]: I0121 11:12:11.625246 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:11 crc kubenswrapper[4824]: I0121 11:12:11.625260 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:11 crc kubenswrapper[4824]: I0121 11:12:11.625266 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:11Z","lastTransitionTime":"2026-01-21T11:12:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:11 crc kubenswrapper[4824]: I0121 11:12:11.727760 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:11 crc kubenswrapper[4824]: I0121 11:12:11.727788 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:11 crc kubenswrapper[4824]: I0121 11:12:11.727797 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:11 crc kubenswrapper[4824]: I0121 11:12:11.727808 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:11 crc kubenswrapper[4824]: I0121 11:12:11.727815 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:11Z","lastTransitionTime":"2026-01-21T11:12:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:11 crc kubenswrapper[4824]: I0121 11:12:11.829585 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:11 crc kubenswrapper[4824]: I0121 11:12:11.829625 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:11 crc kubenswrapper[4824]: I0121 11:12:11.829636 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:11 crc kubenswrapper[4824]: I0121 11:12:11.829653 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:11 crc kubenswrapper[4824]: I0121 11:12:11.829664 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:11Z","lastTransitionTime":"2026-01-21T11:12:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:11 crc kubenswrapper[4824]: I0121 11:12:11.931453 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:11 crc kubenswrapper[4824]: I0121 11:12:11.931494 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:11 crc kubenswrapper[4824]: I0121 11:12:11.931503 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:11 crc kubenswrapper[4824]: I0121 11:12:11.931520 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:11 crc kubenswrapper[4824]: I0121 11:12:11.931528 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:11Z","lastTransitionTime":"2026-01-21T11:12:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:12 crc kubenswrapper[4824]: I0121 11:12:12.024762 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-30 19:09:19.15886122 +0000 UTC Jan 21 11:12:12 crc kubenswrapper[4824]: I0121 11:12:12.033365 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:12 crc kubenswrapper[4824]: I0121 11:12:12.033389 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:12 crc kubenswrapper[4824]: I0121 11:12:12.033399 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:12 crc kubenswrapper[4824]: I0121 11:12:12.033419 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:12 crc kubenswrapper[4824]: I0121 11:12:12.033427 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:12Z","lastTransitionTime":"2026-01-21T11:12:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:12 crc kubenswrapper[4824]: I0121 11:12:12.048973 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9rmnw" Jan 21 11:12:12 crc kubenswrapper[4824]: I0121 11:12:12.048991 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 11:12:12 crc kubenswrapper[4824]: E0121 11:12:12.049082 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9rmnw" podUID="8f9bd12b-22ee-44ed-a8b9-c100d043f691" Jan 21 11:12:12 crc kubenswrapper[4824]: E0121 11:12:12.049215 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 21 11:12:12 crc kubenswrapper[4824]: I0121 11:12:12.135398 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:12 crc kubenswrapper[4824]: I0121 11:12:12.135464 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:12 crc kubenswrapper[4824]: I0121 11:12:12.135487 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:12 crc kubenswrapper[4824]: I0121 11:12:12.135500 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:12 crc kubenswrapper[4824]: I0121 11:12:12.135511 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:12Z","lastTransitionTime":"2026-01-21T11:12:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:12 crc kubenswrapper[4824]: I0121 11:12:12.237012 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:12 crc kubenswrapper[4824]: I0121 11:12:12.237041 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:12 crc kubenswrapper[4824]: I0121 11:12:12.237050 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:12 crc kubenswrapper[4824]: I0121 11:12:12.237064 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:12 crc kubenswrapper[4824]: I0121 11:12:12.237075 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:12Z","lastTransitionTime":"2026-01-21T11:12:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:12 crc kubenswrapper[4824]: I0121 11:12:12.338679 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:12 crc kubenswrapper[4824]: I0121 11:12:12.338729 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:12 crc kubenswrapper[4824]: I0121 11:12:12.338741 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:12 crc kubenswrapper[4824]: I0121 11:12:12.338759 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:12 crc kubenswrapper[4824]: I0121 11:12:12.338768 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:12Z","lastTransitionTime":"2026-01-21T11:12:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:12 crc kubenswrapper[4824]: I0121 11:12:12.440774 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:12 crc kubenswrapper[4824]: I0121 11:12:12.440805 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:12 crc kubenswrapper[4824]: I0121 11:12:12.440814 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:12 crc kubenswrapper[4824]: I0121 11:12:12.440825 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:12 crc kubenswrapper[4824]: I0121 11:12:12.440834 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:12Z","lastTransitionTime":"2026-01-21T11:12:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:12 crc kubenswrapper[4824]: I0121 11:12:12.542804 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:12 crc kubenswrapper[4824]: I0121 11:12:12.542839 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:12 crc kubenswrapper[4824]: I0121 11:12:12.542847 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:12 crc kubenswrapper[4824]: I0121 11:12:12.542860 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:12 crc kubenswrapper[4824]: I0121 11:12:12.542868 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:12Z","lastTransitionTime":"2026-01-21T11:12:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:12 crc kubenswrapper[4824]: I0121 11:12:12.644675 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:12 crc kubenswrapper[4824]: I0121 11:12:12.644730 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:12 crc kubenswrapper[4824]: I0121 11:12:12.644740 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:12 crc kubenswrapper[4824]: I0121 11:12:12.644757 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:12 crc kubenswrapper[4824]: I0121 11:12:12.644766 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:12Z","lastTransitionTime":"2026-01-21T11:12:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:12 crc kubenswrapper[4824]: I0121 11:12:12.746332 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:12 crc kubenswrapper[4824]: I0121 11:12:12.746360 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:12 crc kubenswrapper[4824]: I0121 11:12:12.746369 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:12 crc kubenswrapper[4824]: I0121 11:12:12.746380 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:12 crc kubenswrapper[4824]: I0121 11:12:12.746390 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:12Z","lastTransitionTime":"2026-01-21T11:12:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:12 crc kubenswrapper[4824]: I0121 11:12:12.848101 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:12 crc kubenswrapper[4824]: I0121 11:12:12.848139 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:12 crc kubenswrapper[4824]: I0121 11:12:12.848150 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:12 crc kubenswrapper[4824]: I0121 11:12:12.848165 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:12 crc kubenswrapper[4824]: I0121 11:12:12.848174 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:12Z","lastTransitionTime":"2026-01-21T11:12:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:12 crc kubenswrapper[4824]: I0121 11:12:12.949865 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:12 crc kubenswrapper[4824]: I0121 11:12:12.949899 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:12 crc kubenswrapper[4824]: I0121 11:12:12.949907 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:12 crc kubenswrapper[4824]: I0121 11:12:12.949920 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:12 crc kubenswrapper[4824]: I0121 11:12:12.949928 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:12Z","lastTransitionTime":"2026-01-21T11:12:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:13 crc kubenswrapper[4824]: I0121 11:12:13.025560 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-07 22:38:28.950155576 +0000 UTC Jan 21 11:12:13 crc kubenswrapper[4824]: I0121 11:12:13.049191 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 11:12:13 crc kubenswrapper[4824]: I0121 11:12:13.049211 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 11:12:13 crc kubenswrapper[4824]: E0121 11:12:13.049281 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 21 11:12:13 crc kubenswrapper[4824]: E0121 11:12:13.049358 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 21 11:12:13 crc kubenswrapper[4824]: I0121 11:12:13.051278 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:13 crc kubenswrapper[4824]: I0121 11:12:13.051300 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:13 crc kubenswrapper[4824]: I0121 11:12:13.051311 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:13 crc kubenswrapper[4824]: I0121 11:12:13.051322 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:13 crc kubenswrapper[4824]: I0121 11:12:13.051331 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:13Z","lastTransitionTime":"2026-01-21T11:12:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:13 crc kubenswrapper[4824]: I0121 11:12:13.152486 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:13 crc kubenswrapper[4824]: I0121 11:12:13.152512 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:13 crc kubenswrapper[4824]: I0121 11:12:13.152520 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:13 crc kubenswrapper[4824]: I0121 11:12:13.152531 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:13 crc kubenswrapper[4824]: I0121 11:12:13.152538 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:13Z","lastTransitionTime":"2026-01-21T11:12:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:13 crc kubenswrapper[4824]: I0121 11:12:13.256163 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:13 crc kubenswrapper[4824]: I0121 11:12:13.256194 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:13 crc kubenswrapper[4824]: I0121 11:12:13.256202 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:13 crc kubenswrapper[4824]: I0121 11:12:13.256214 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:13 crc kubenswrapper[4824]: I0121 11:12:13.256222 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:13Z","lastTransitionTime":"2026-01-21T11:12:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:13 crc kubenswrapper[4824]: I0121 11:12:13.358037 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:13 crc kubenswrapper[4824]: I0121 11:12:13.358071 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:13 crc kubenswrapper[4824]: I0121 11:12:13.358081 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:13 crc kubenswrapper[4824]: I0121 11:12:13.358094 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:13 crc kubenswrapper[4824]: I0121 11:12:13.358103 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:13Z","lastTransitionTime":"2026-01-21T11:12:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:13 crc kubenswrapper[4824]: I0121 11:12:13.460093 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:13 crc kubenswrapper[4824]: I0121 11:12:13.460121 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:13 crc kubenswrapper[4824]: I0121 11:12:13.460128 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:13 crc kubenswrapper[4824]: I0121 11:12:13.460137 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:13 crc kubenswrapper[4824]: I0121 11:12:13.460144 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:13Z","lastTransitionTime":"2026-01-21T11:12:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:13 crc kubenswrapper[4824]: I0121 11:12:13.562250 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:13 crc kubenswrapper[4824]: I0121 11:12:13.562281 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:13 crc kubenswrapper[4824]: I0121 11:12:13.562289 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:13 crc kubenswrapper[4824]: I0121 11:12:13.562299 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:13 crc kubenswrapper[4824]: I0121 11:12:13.562306 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:13Z","lastTransitionTime":"2026-01-21T11:12:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:13 crc kubenswrapper[4824]: I0121 11:12:13.663846 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:13 crc kubenswrapper[4824]: I0121 11:12:13.663876 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:13 crc kubenswrapper[4824]: I0121 11:12:13.663883 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:13 crc kubenswrapper[4824]: I0121 11:12:13.663909 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:13 crc kubenswrapper[4824]: I0121 11:12:13.663918 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:13Z","lastTransitionTime":"2026-01-21T11:12:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:13 crc kubenswrapper[4824]: I0121 11:12:13.765580 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:13 crc kubenswrapper[4824]: I0121 11:12:13.765608 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:13 crc kubenswrapper[4824]: I0121 11:12:13.765616 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:13 crc kubenswrapper[4824]: I0121 11:12:13.765628 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:13 crc kubenswrapper[4824]: I0121 11:12:13.765635 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:13Z","lastTransitionTime":"2026-01-21T11:12:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:13 crc kubenswrapper[4824]: I0121 11:12:13.867831 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:13 crc kubenswrapper[4824]: I0121 11:12:13.867860 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:13 crc kubenswrapper[4824]: I0121 11:12:13.867868 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:13 crc kubenswrapper[4824]: I0121 11:12:13.867878 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:13 crc kubenswrapper[4824]: I0121 11:12:13.867886 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:13Z","lastTransitionTime":"2026-01-21T11:12:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:13 crc kubenswrapper[4824]: I0121 11:12:13.971278 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:13 crc kubenswrapper[4824]: I0121 11:12:13.971306 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:13 crc kubenswrapper[4824]: I0121 11:12:13.971314 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:13 crc kubenswrapper[4824]: I0121 11:12:13.971327 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:13 crc kubenswrapper[4824]: I0121 11:12:13.971337 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:13Z","lastTransitionTime":"2026-01-21T11:12:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:14 crc kubenswrapper[4824]: I0121 11:12:14.025977 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-12 00:29:47.057543229 +0000 UTC Jan 21 11:12:14 crc kubenswrapper[4824]: I0121 11:12:14.048505 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9rmnw" Jan 21 11:12:14 crc kubenswrapper[4824]: I0121 11:12:14.048584 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 11:12:14 crc kubenswrapper[4824]: E0121 11:12:14.048665 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 21 11:12:14 crc kubenswrapper[4824]: E0121 11:12:14.048728 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9rmnw" podUID="8f9bd12b-22ee-44ed-a8b9-c100d043f691" Jan 21 11:12:14 crc kubenswrapper[4824]: I0121 11:12:14.056510 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Jan 21 11:12:14 crc kubenswrapper[4824]: I0121 11:12:14.073472 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:14 crc kubenswrapper[4824]: I0121 11:12:14.073499 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:14 crc kubenswrapper[4824]: I0121 11:12:14.073507 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:14 crc kubenswrapper[4824]: I0121 11:12:14.073517 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:14 crc kubenswrapper[4824]: I0121 11:12:14.073525 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:14Z","lastTransitionTime":"2026-01-21T11:12:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:14 crc kubenswrapper[4824]: I0121 11:12:14.175619 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:14 crc kubenswrapper[4824]: I0121 11:12:14.175650 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:14 crc kubenswrapper[4824]: I0121 11:12:14.175658 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:14 crc kubenswrapper[4824]: I0121 11:12:14.175669 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:14 crc kubenswrapper[4824]: I0121 11:12:14.175677 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:14Z","lastTransitionTime":"2026-01-21T11:12:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:14 crc kubenswrapper[4824]: I0121 11:12:14.277860 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:14 crc kubenswrapper[4824]: I0121 11:12:14.277895 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:14 crc kubenswrapper[4824]: I0121 11:12:14.277903 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:14 crc kubenswrapper[4824]: I0121 11:12:14.277916 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:14 crc kubenswrapper[4824]: I0121 11:12:14.277924 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:14Z","lastTransitionTime":"2026-01-21T11:12:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:14 crc kubenswrapper[4824]: I0121 11:12:14.379783 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:14 crc kubenswrapper[4824]: I0121 11:12:14.379827 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:14 crc kubenswrapper[4824]: I0121 11:12:14.379835 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:14 crc kubenswrapper[4824]: I0121 11:12:14.379846 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:14 crc kubenswrapper[4824]: I0121 11:12:14.379853 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:14Z","lastTransitionTime":"2026-01-21T11:12:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:14 crc kubenswrapper[4824]: I0121 11:12:14.481669 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:14 crc kubenswrapper[4824]: I0121 11:12:14.481698 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:14 crc kubenswrapper[4824]: I0121 11:12:14.481706 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:14 crc kubenswrapper[4824]: I0121 11:12:14.481718 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:14 crc kubenswrapper[4824]: I0121 11:12:14.481730 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:14Z","lastTransitionTime":"2026-01-21T11:12:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:14 crc kubenswrapper[4824]: I0121 11:12:14.583596 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:14 crc kubenswrapper[4824]: I0121 11:12:14.583636 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:14 crc kubenswrapper[4824]: I0121 11:12:14.583645 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:14 crc kubenswrapper[4824]: I0121 11:12:14.583673 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:14 crc kubenswrapper[4824]: I0121 11:12:14.583680 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:14Z","lastTransitionTime":"2026-01-21T11:12:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:14 crc kubenswrapper[4824]: I0121 11:12:14.685546 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:14 crc kubenswrapper[4824]: I0121 11:12:14.685583 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:14 crc kubenswrapper[4824]: I0121 11:12:14.685592 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:14 crc kubenswrapper[4824]: I0121 11:12:14.685605 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:14 crc kubenswrapper[4824]: I0121 11:12:14.685613 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:14Z","lastTransitionTime":"2026-01-21T11:12:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:14 crc kubenswrapper[4824]: I0121 11:12:14.787228 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:14 crc kubenswrapper[4824]: I0121 11:12:14.787273 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:14 crc kubenswrapper[4824]: I0121 11:12:14.787283 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:14 crc kubenswrapper[4824]: I0121 11:12:14.787296 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:14 crc kubenswrapper[4824]: I0121 11:12:14.787304 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:14Z","lastTransitionTime":"2026-01-21T11:12:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:14 crc kubenswrapper[4824]: I0121 11:12:14.889510 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:14 crc kubenswrapper[4824]: I0121 11:12:14.889544 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:14 crc kubenswrapper[4824]: I0121 11:12:14.889569 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:14 crc kubenswrapper[4824]: I0121 11:12:14.889583 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:14 crc kubenswrapper[4824]: I0121 11:12:14.889592 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:14Z","lastTransitionTime":"2026-01-21T11:12:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:14 crc kubenswrapper[4824]: I0121 11:12:14.991124 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:14 crc kubenswrapper[4824]: I0121 11:12:14.991162 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:14 crc kubenswrapper[4824]: I0121 11:12:14.991170 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:14 crc kubenswrapper[4824]: I0121 11:12:14.991184 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:14 crc kubenswrapper[4824]: I0121 11:12:14.991193 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:14Z","lastTransitionTime":"2026-01-21T11:12:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:15 crc kubenswrapper[4824]: I0121 11:12:15.026683 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-26 00:19:52.153953239 +0000 UTC Jan 21 11:12:15 crc kubenswrapper[4824]: I0121 11:12:15.049221 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 11:12:15 crc kubenswrapper[4824]: I0121 11:12:15.049245 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 11:12:15 crc kubenswrapper[4824]: E0121 11:12:15.049326 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 21 11:12:15 crc kubenswrapper[4824]: E0121 11:12:15.049391 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 21 11:12:15 crc kubenswrapper[4824]: I0121 11:12:15.093027 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:15 crc kubenswrapper[4824]: I0121 11:12:15.093058 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:15 crc kubenswrapper[4824]: I0121 11:12:15.093068 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:15 crc kubenswrapper[4824]: I0121 11:12:15.093082 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:15 crc kubenswrapper[4824]: I0121 11:12:15.093090 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:15Z","lastTransitionTime":"2026-01-21T11:12:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:15 crc kubenswrapper[4824]: I0121 11:12:15.195114 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:15 crc kubenswrapper[4824]: I0121 11:12:15.195150 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:15 crc kubenswrapper[4824]: I0121 11:12:15.195158 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:15 crc kubenswrapper[4824]: I0121 11:12:15.195174 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:15 crc kubenswrapper[4824]: I0121 11:12:15.195182 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:15Z","lastTransitionTime":"2026-01-21T11:12:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:15 crc kubenswrapper[4824]: I0121 11:12:15.296646 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:15 crc kubenswrapper[4824]: I0121 11:12:15.296699 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:15 crc kubenswrapper[4824]: I0121 11:12:15.296708 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:15 crc kubenswrapper[4824]: I0121 11:12:15.296722 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:15 crc kubenswrapper[4824]: I0121 11:12:15.296732 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:15Z","lastTransitionTime":"2026-01-21T11:12:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:15 crc kubenswrapper[4824]: I0121 11:12:15.398165 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:15 crc kubenswrapper[4824]: I0121 11:12:15.398199 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:15 crc kubenswrapper[4824]: I0121 11:12:15.398207 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:15 crc kubenswrapper[4824]: I0121 11:12:15.398218 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:15 crc kubenswrapper[4824]: I0121 11:12:15.398227 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:15Z","lastTransitionTime":"2026-01-21T11:12:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:15 crc kubenswrapper[4824]: I0121 11:12:15.499414 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:15 crc kubenswrapper[4824]: I0121 11:12:15.499446 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:15 crc kubenswrapper[4824]: I0121 11:12:15.499455 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:15 crc kubenswrapper[4824]: I0121 11:12:15.499468 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:15 crc kubenswrapper[4824]: I0121 11:12:15.499476 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:15Z","lastTransitionTime":"2026-01-21T11:12:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:15 crc kubenswrapper[4824]: I0121 11:12:15.600727 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:15 crc kubenswrapper[4824]: I0121 11:12:15.600755 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:15 crc kubenswrapper[4824]: I0121 11:12:15.600765 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:15 crc kubenswrapper[4824]: I0121 11:12:15.600777 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:15 crc kubenswrapper[4824]: I0121 11:12:15.600785 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:15Z","lastTransitionTime":"2026-01-21T11:12:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:15 crc kubenswrapper[4824]: I0121 11:12:15.702237 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:15 crc kubenswrapper[4824]: I0121 11:12:15.702268 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:15 crc kubenswrapper[4824]: I0121 11:12:15.702277 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:15 crc kubenswrapper[4824]: I0121 11:12:15.702306 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:15 crc kubenswrapper[4824]: I0121 11:12:15.702316 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:15Z","lastTransitionTime":"2026-01-21T11:12:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:15 crc kubenswrapper[4824]: I0121 11:12:15.806731 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:15 crc kubenswrapper[4824]: I0121 11:12:15.806767 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:15 crc kubenswrapper[4824]: I0121 11:12:15.806777 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:15 crc kubenswrapper[4824]: I0121 11:12:15.806789 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:15 crc kubenswrapper[4824]: I0121 11:12:15.806797 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:15Z","lastTransitionTime":"2026-01-21T11:12:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:15 crc kubenswrapper[4824]: I0121 11:12:15.908015 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:15 crc kubenswrapper[4824]: I0121 11:12:15.908050 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:15 crc kubenswrapper[4824]: I0121 11:12:15.908059 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:15 crc kubenswrapper[4824]: I0121 11:12:15.908073 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:15 crc kubenswrapper[4824]: I0121 11:12:15.908081 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:15Z","lastTransitionTime":"2026-01-21T11:12:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:16 crc kubenswrapper[4824]: I0121 11:12:16.010068 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:16 crc kubenswrapper[4824]: I0121 11:12:16.010108 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:16 crc kubenswrapper[4824]: I0121 11:12:16.010119 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:16 crc kubenswrapper[4824]: I0121 11:12:16.010133 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:16 crc kubenswrapper[4824]: I0121 11:12:16.010141 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:16Z","lastTransitionTime":"2026-01-21T11:12:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:16 crc kubenswrapper[4824]: I0121 11:12:16.027420 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-02 09:48:22.72384722 +0000 UTC Jan 21 11:12:16 crc kubenswrapper[4824]: I0121 11:12:16.048778 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 11:12:16 crc kubenswrapper[4824]: E0121 11:12:16.048883 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 21 11:12:16 crc kubenswrapper[4824]: I0121 11:12:16.048914 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9rmnw" Jan 21 11:12:16 crc kubenswrapper[4824]: E0121 11:12:16.049041 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9rmnw" podUID="8f9bd12b-22ee-44ed-a8b9-c100d043f691" Jan 21 11:12:16 crc kubenswrapper[4824]: I0121 11:12:16.112028 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:16 crc kubenswrapper[4824]: I0121 11:12:16.112054 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:16 crc kubenswrapper[4824]: I0121 11:12:16.112063 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:16 crc kubenswrapper[4824]: I0121 11:12:16.112082 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:16 crc kubenswrapper[4824]: I0121 11:12:16.112091 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:16Z","lastTransitionTime":"2026-01-21T11:12:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:16 crc kubenswrapper[4824]: I0121 11:12:16.214153 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:16 crc kubenswrapper[4824]: I0121 11:12:16.214191 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:16 crc kubenswrapper[4824]: I0121 11:12:16.214202 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:16 crc kubenswrapper[4824]: I0121 11:12:16.214217 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:16 crc kubenswrapper[4824]: I0121 11:12:16.214225 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:16Z","lastTransitionTime":"2026-01-21T11:12:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:16 crc kubenswrapper[4824]: I0121 11:12:16.316442 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:16 crc kubenswrapper[4824]: I0121 11:12:16.316470 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:16 crc kubenswrapper[4824]: I0121 11:12:16.316478 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:16 crc kubenswrapper[4824]: I0121 11:12:16.316491 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:16 crc kubenswrapper[4824]: I0121 11:12:16.316501 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:16Z","lastTransitionTime":"2026-01-21T11:12:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:16 crc kubenswrapper[4824]: I0121 11:12:16.418382 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:16 crc kubenswrapper[4824]: I0121 11:12:16.418431 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:16 crc kubenswrapper[4824]: I0121 11:12:16.418439 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:16 crc kubenswrapper[4824]: I0121 11:12:16.418452 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:16 crc kubenswrapper[4824]: I0121 11:12:16.418461 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:16Z","lastTransitionTime":"2026-01-21T11:12:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:16 crc kubenswrapper[4824]: I0121 11:12:16.520043 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:16 crc kubenswrapper[4824]: I0121 11:12:16.520074 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:16 crc kubenswrapper[4824]: I0121 11:12:16.520082 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:16 crc kubenswrapper[4824]: I0121 11:12:16.520094 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:16 crc kubenswrapper[4824]: I0121 11:12:16.520103 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:16Z","lastTransitionTime":"2026-01-21T11:12:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:16 crc kubenswrapper[4824]: I0121 11:12:16.622003 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:16 crc kubenswrapper[4824]: I0121 11:12:16.622048 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:16 crc kubenswrapper[4824]: I0121 11:12:16.622056 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:16 crc kubenswrapper[4824]: I0121 11:12:16.622068 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:16 crc kubenswrapper[4824]: I0121 11:12:16.622077 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:16Z","lastTransitionTime":"2026-01-21T11:12:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:16 crc kubenswrapper[4824]: I0121 11:12:16.723710 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:16 crc kubenswrapper[4824]: I0121 11:12:16.723740 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:16 crc kubenswrapper[4824]: I0121 11:12:16.723750 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:16 crc kubenswrapper[4824]: I0121 11:12:16.723763 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:16 crc kubenswrapper[4824]: I0121 11:12:16.723772 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:16Z","lastTransitionTime":"2026-01-21T11:12:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:16 crc kubenswrapper[4824]: I0121 11:12:16.825449 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:16 crc kubenswrapper[4824]: I0121 11:12:16.825480 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:16 crc kubenswrapper[4824]: I0121 11:12:16.825489 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:16 crc kubenswrapper[4824]: I0121 11:12:16.825501 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:16 crc kubenswrapper[4824]: I0121 11:12:16.825510 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:16Z","lastTransitionTime":"2026-01-21T11:12:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:16 crc kubenswrapper[4824]: I0121 11:12:16.927004 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:16 crc kubenswrapper[4824]: I0121 11:12:16.927034 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:16 crc kubenswrapper[4824]: I0121 11:12:16.927042 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:16 crc kubenswrapper[4824]: I0121 11:12:16.927054 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:16 crc kubenswrapper[4824]: I0121 11:12:16.927061 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:16Z","lastTransitionTime":"2026-01-21T11:12:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:17 crc kubenswrapper[4824]: I0121 11:12:17.027997 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-15 08:58:56.390507961 +0000 UTC Jan 21 11:12:17 crc kubenswrapper[4824]: I0121 11:12:17.029301 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:17 crc kubenswrapper[4824]: I0121 11:12:17.029339 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:17 crc kubenswrapper[4824]: I0121 11:12:17.029347 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:17 crc kubenswrapper[4824]: I0121 11:12:17.029359 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:17 crc kubenswrapper[4824]: I0121 11:12:17.029369 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:17Z","lastTransitionTime":"2026-01-21T11:12:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:17 crc kubenswrapper[4824]: I0121 11:12:17.048579 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 11:12:17 crc kubenswrapper[4824]: I0121 11:12:17.048601 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 11:12:17 crc kubenswrapper[4824]: E0121 11:12:17.048684 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 21 11:12:17 crc kubenswrapper[4824]: E0121 11:12:17.048760 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 21 11:12:17 crc kubenswrapper[4824]: I0121 11:12:17.131334 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:17 crc kubenswrapper[4824]: I0121 11:12:17.131367 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:17 crc kubenswrapper[4824]: I0121 11:12:17.131375 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:17 crc kubenswrapper[4824]: I0121 11:12:17.131402 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:17 crc kubenswrapper[4824]: I0121 11:12:17.131412 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:17Z","lastTransitionTime":"2026-01-21T11:12:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:17 crc kubenswrapper[4824]: I0121 11:12:17.233226 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:17 crc kubenswrapper[4824]: I0121 11:12:17.233259 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:17 crc kubenswrapper[4824]: I0121 11:12:17.233268 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:17 crc kubenswrapper[4824]: I0121 11:12:17.233283 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:17 crc kubenswrapper[4824]: I0121 11:12:17.233291 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:17Z","lastTransitionTime":"2026-01-21T11:12:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:17 crc kubenswrapper[4824]: I0121 11:12:17.335173 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:17 crc kubenswrapper[4824]: I0121 11:12:17.335207 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:17 crc kubenswrapper[4824]: I0121 11:12:17.335217 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:17 crc kubenswrapper[4824]: I0121 11:12:17.335231 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:17 crc kubenswrapper[4824]: I0121 11:12:17.335239 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:17Z","lastTransitionTime":"2026-01-21T11:12:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:17 crc kubenswrapper[4824]: I0121 11:12:17.436997 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:17 crc kubenswrapper[4824]: I0121 11:12:17.437035 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:17 crc kubenswrapper[4824]: I0121 11:12:17.437046 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:17 crc kubenswrapper[4824]: I0121 11:12:17.437058 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:17 crc kubenswrapper[4824]: I0121 11:12:17.437067 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:17Z","lastTransitionTime":"2026-01-21T11:12:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:17 crc kubenswrapper[4824]: I0121 11:12:17.539156 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:17 crc kubenswrapper[4824]: I0121 11:12:17.539187 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:17 crc kubenswrapper[4824]: I0121 11:12:17.539196 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:17 crc kubenswrapper[4824]: I0121 11:12:17.539209 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:17 crc kubenswrapper[4824]: I0121 11:12:17.539218 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:17Z","lastTransitionTime":"2026-01-21T11:12:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:17 crc kubenswrapper[4824]: I0121 11:12:17.641472 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:17 crc kubenswrapper[4824]: I0121 11:12:17.641524 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:17 crc kubenswrapper[4824]: I0121 11:12:17.641535 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:17 crc kubenswrapper[4824]: I0121 11:12:17.641548 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:17 crc kubenswrapper[4824]: I0121 11:12:17.641558 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:17Z","lastTransitionTime":"2026-01-21T11:12:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:17 crc kubenswrapper[4824]: I0121 11:12:17.743800 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:17 crc kubenswrapper[4824]: I0121 11:12:17.743833 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:17 crc kubenswrapper[4824]: I0121 11:12:17.743841 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:17 crc kubenswrapper[4824]: I0121 11:12:17.743854 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:17 crc kubenswrapper[4824]: I0121 11:12:17.743862 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:17Z","lastTransitionTime":"2026-01-21T11:12:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:17 crc kubenswrapper[4824]: I0121 11:12:17.845556 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:17 crc kubenswrapper[4824]: I0121 11:12:17.845608 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:17 crc kubenswrapper[4824]: I0121 11:12:17.845619 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:17 crc kubenswrapper[4824]: I0121 11:12:17.845631 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:17 crc kubenswrapper[4824]: I0121 11:12:17.845640 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:17Z","lastTransitionTime":"2026-01-21T11:12:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:17 crc kubenswrapper[4824]: I0121 11:12:17.947479 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:17 crc kubenswrapper[4824]: I0121 11:12:17.947518 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:17 crc kubenswrapper[4824]: I0121 11:12:17.947527 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:17 crc kubenswrapper[4824]: I0121 11:12:17.947540 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:17 crc kubenswrapper[4824]: I0121 11:12:17.947550 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:17Z","lastTransitionTime":"2026-01-21T11:12:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:18 crc kubenswrapper[4824]: I0121 11:12:18.028278 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-09 19:47:14.845753186 +0000 UTC Jan 21 11:12:18 crc kubenswrapper[4824]: I0121 11:12:18.048511 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9rmnw" Jan 21 11:12:18 crc kubenswrapper[4824]: I0121 11:12:18.048548 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 11:12:18 crc kubenswrapper[4824]: E0121 11:12:18.048888 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9rmnw" podUID="8f9bd12b-22ee-44ed-a8b9-c100d043f691" Jan 21 11:12:18 crc kubenswrapper[4824]: E0121 11:12:18.049000 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 21 11:12:18 crc kubenswrapper[4824]: I0121 11:12:18.049774 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:18 crc kubenswrapper[4824]: I0121 11:12:18.050257 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:18 crc kubenswrapper[4824]: I0121 11:12:18.050287 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:18 crc kubenswrapper[4824]: I0121 11:12:18.050301 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:18 crc kubenswrapper[4824]: I0121 11:12:18.050309 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:18Z","lastTransitionTime":"2026-01-21T11:12:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:18 crc kubenswrapper[4824]: I0121 11:12:18.058769 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jr64s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9768c60d-5bee-4286-bced-076e584b04e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91158f14d2e98ac9f875a5e7101df3107ef28aa5fc5d62218f460af98ff920fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flrnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:17Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jr64s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:18Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:18 crc kubenswrapper[4824]: I0121 11:12:18.066229 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b44f8a34-1578-4634-b62e-8e222844b397\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fd58f2438b6f9cbb2c37687e39f02cdf96a029477b1ae33f6dc9f3b70ecc926\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef10a172a3279201375f7445673820033978ce1eb2bbef7a2c0ed297fd783875\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef10a172a3279201375f7445673820033978ce1eb2bbef7a2c0ed297fd783875\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:18Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:18 crc kubenswrapper[4824]: I0121 11:12:18.075735 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d31ab61-f94c-4f49-85d0-c9e91b83069b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a66b5c33884af00408dc35cf02f08ebd8570bd695f9f1fe558bf54fd06930431\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7514786602746dabfa690da48d48e233fe68d6dd6674aa5ecd489c0b3185f08\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61b14b6846f94cdeba8f1f61b3754e7663041f92c70356c4ee6cbb988a2e680f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a7023cfe4c098bcdd9c0477a8187f4ed7a0d2abdbe1e143436863948f441b4f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af11ad95bcf0c6bbd5fa4ab99bb592ce1c87f40d29c5ade7832414315820a569\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW0121 11:11:14.604290 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0121 11:11:14.604572 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0121 11:11:14.605875 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3284452394/tls.crt::/tmp/serving-cert-3284452394/tls.key\\\\\\\"\\\\nI0121 11:11:14.923427 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0121 11:11:14.928141 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0121 11:11:14.928162 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0121 11:11:14.928179 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0121 11:11:14.928184 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0121 11:11:14.933154 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0121 11:11:14.933230 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 11:11:14.933259 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 11:11:14.933282 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0121 11:11:14.933299 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0121 11:11:14.933315 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0121 11:11:14.933331 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0121 11:11:14.933177 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0121 11:11:14.934666 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48070c09fad511108ff22e44211cf26911509b8406d66948985cbfc3009d8233\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:18Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:18 crc kubenswrapper[4824]: I0121 11:12:18.083866 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9bb99c0a-683b-44da-9c0e-e8d284e87d53\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05e076febd697ed802ce5c7ee92a8c2c636a1a2c6609ad7eb01f17dace247c4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43529d1b9e1e5e22b0d56b023946119da8aacbae54b68c26ff0a1a82dc890c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ea232ad87393685aa323ce6393ff3869f8f8d423c2406a501c69c355535adfa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc6c7eff9e17b8a8ecc88bbfe703b5a01dbf1e02d342cede4f6a263a6526af43\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:18Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:18 crc kubenswrapper[4824]: I0121 11:12:18.092524 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e570fff1-6227-44d6-b71f-ee377b1099ae\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4d41492c3802a8168fffd249c15aeb9c902a5c7e01550e1aa6f9b769adfdc4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2690a1ecbc8fe1290e68bba2695a3edd833b5da933fb71e8ddc926a1340a9a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b85db961e1e078e1f54323c79a99c9413365256ac2fce411db7932bd0580ab6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://595783a1454735bd33f4862180b60d99a4e1b6fae1e735f2b7328391951bf422\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://595783a1454735bd33f4862180b60d99a4e1b6fae1e735f2b7328391951bf422\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:18Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:18 crc kubenswrapper[4824]: I0121 11:12:18.101943 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:18Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:18 crc kubenswrapper[4824]: I0121 11:12:18.108801 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lv4rn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1267539-9adf-466f-8027-8ede009d19a0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eed1b884aa53e1ee318169a0cc2c9ad10130aa363ae4ac55e4bf065237756f53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbmfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lv4rn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:18Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:18 crc kubenswrapper[4824]: I0121 11:12:18.121714 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06f56d86-ec6f-4b57-b314-f508c0dd4d9f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d25c94f1ae0bbabb733f0e1ea1d1ba39562a92af73047a39aacc03013e239681\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca06234486b59cf990c69d9501e23f935328e2ba21013cd71ed64831ea749bcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b63e45d66bcd11ff3479cc3ec6b450017da8d21d36e967b8d496ae0f24b0ee9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52758bcd31f2ba64edf03adfd4d1be6428015abff5b64c7c029fa787085f9c29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2736e73342d2adb175b0721dd4f40cc42d26fca06d5eb3334927518531c3b107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b35ad7dd0570c191eb0080739304f3cd43f17b3d1dddf895549441ebe50deaa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b35ad7dd0570c191eb0080739304f3cd43f17b3d1dddf895549441ebe50deaa6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://80cd5ea5e6fc19a88c1aab5c61f827f103ab798a5c84da8dc4d45cea86d5aacb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80cd5ea5e6fc19a88c1aab5c61f827f103ab798a5c84da8dc4d45cea86d5aacb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://234d8b914e6a88480021647ae65d74e9b21c55b6788312a094d00c8a42677d48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://234d8b914e6a88480021647ae65d74e9b21c55b6788312a094d00c8a42677d48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:18Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:18 crc kubenswrapper[4824]: I0121 11:12:18.129512 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:18Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:18 crc kubenswrapper[4824]: I0121 11:12:18.137797 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86d3005cd61c0973659e1c1c0d3ef1ddd6fcf95bede5465f4f2099e64a74bb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:18Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:18 crc kubenswrapper[4824]: I0121 11:12:18.145396 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:18Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:18 crc kubenswrapper[4824]: I0121 11:12:18.151934 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:18 crc kubenswrapper[4824]: I0121 11:12:18.151982 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:18 crc kubenswrapper[4824]: I0121 11:12:18.151992 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:18 crc kubenswrapper[4824]: I0121 11:12:18.152004 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:18 crc kubenswrapper[4824]: I0121 11:12:18.152013 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:18Z","lastTransitionTime":"2026-01-21T11:12:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:18 crc kubenswrapper[4824]: I0121 11:12:18.153479 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x94nb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:12:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:12:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49ce055ec8090d6957cb7e23fe75edb82c0c2d7f87e7203d280606368dd3179f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75d52224135b75133fc000611cd179240b9174f6492d3263341716c7e42ccb15\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-21T11:12:02Z\\\",\\\"message\\\":\\\"2026-01-21T11:11:17+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_c9e2a0c0-e545-4ddc-b22f-97601d5900ca\\\\n2026-01-21T11:11:17+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_c9e2a0c0-e545-4ddc-b22f-97601d5900ca to /host/opt/cni/bin/\\\\n2026-01-21T11:11:17Z [verbose] multus-daemon started\\\\n2026-01-21T11:11:17Z [verbose] Readiness Indicator file check\\\\n2026-01-21T11:12:02Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:12:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgr8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x94nb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:18Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:18 crc kubenswrapper[4824]: I0121 11:12:18.161067 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33f3d922-4ffe-409b-a49a-d88c85898260\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a9e57b39eed8300107e944b9fcf6a2ff1e75991f8f37387a3de615b14f9bc78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpwst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b1b37c6d2994021cfc30f85e92c40ae35677113b17aa066d52bc7a76ae5c153\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpwst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zqjg9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:18Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:18 crc kubenswrapper[4824]: I0121 11:12:18.170782 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gjtnh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a8d2cac-2d0f-4894-9bc0-2abd81699c47\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67c122a26811deaa9b70d19a1a65daee318ca3af2c626712482179d72fe67875\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2239186f1de46481cebfcafb1cab85cd320a205b5a0dfc6dd1b321d9baedeba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b2239186f1de46481cebfcafb1cab85cd320a205b5a0dfc6dd1b321d9baedeba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0db5ecd7759d974ae1b0d8065b3b8bd4d7957fae16eb5bab2d71c2fe69243816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0db5ecd7759d974ae1b0d8065b3b8bd4d7957fae16eb5bab2d71c2fe69243816\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba9b3278c208fc7aac630ca7a0daa3de0823f85d0799d6f1f278da80ae14c862\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba9b3278c208fc7aac630ca7a0daa3de0823f85d0799d6f1f278da80ae14c862\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79b8f6c016b43c3ac3ba7b22cca9a6b72c262cb17ddd6236c5d4970ecc09680c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79b8f6c016b43c3ac3ba7b22cca9a6b72c262cb17ddd6236c5d4970ecc09680c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38c4f218eb2ca7ea0c92bd0ae2ae9efeaa40746751608fbcf7f6c3f8d9e3c758\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38c4f218eb2ca7ea0c92bd0ae2ae9efeaa40746751608fbcf7f6c3f8d9e3c758\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d66ad9831d6029c8c350054525a00069e869d39c1b5638e06e9b772666bd8630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d66ad9831d6029c8c350054525a00069e869d39c1b5638e06e9b772666bd8630\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gjtnh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:18Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:18 crc kubenswrapper[4824]: I0121 11:12:18.177928 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-9rmnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f9bd12b-22ee-44ed-a8b9-c100d043f691\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2fgn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2fgn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:28Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-9rmnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:18Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:18 crc kubenswrapper[4824]: I0121 11:12:18.185300 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d87422473838f5218c9fb646ab2515f2f42592754ffdf07b19f7a8e093fa48a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:18Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:18 crc kubenswrapper[4824]: I0121 11:12:18.193555 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://516c009d4720ab5b393c7482872f12fed22d4f4414bd1f3400411c0e8da88ce4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee18d293036971832c8d9c3f7b9eca84ab779ff6768a2db3c6f7bdcf3433cdef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:18Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:18 crc kubenswrapper[4824]: I0121 11:12:18.205805 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"26b1108a-4b98-4e39-a3cd-e0d055089fd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc319bbc06568cc811ded2b418b04b986a31f321fed492e8be50104af79bb30b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cee9cc9eec3fc815f1621a22a37805691288dbb60dd3ff64194a84be62abcddc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc4c96e33e069f7d08d33caf92cd0987668506f931fd134b246d5ce5e934f95c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ddbb0e5ae02e9d5d82a80ead7efc91f1b8e557a9fb6f39940e5530fae8de2d38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://64a9065254f889b1c43d63759283335bf78fc4fa1531a4baae472c06c075dae3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e94ef8a8f47dc75ecfbab2e3fb8e7097c1b73eed32d7ca947f11ea198cea880\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86a6881f109dab1f910fe505a21c7c800e587de4dded2635147997c51c05c483\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86a6881f109dab1f910fe505a21c7c800e587de4dded2635147997c51c05c483\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-21T11:12:07Z\\\",\\\"message\\\":\\\"UUIDName:}]\\\\nI0121 11:12:07.749346 6924 services_controller.go:453] Built service openshift-console/console template LB for network=default: []services.LB{}\\\\nI0121 11:12:07.749353 6924 services_controller.go:454] Service openshift-console/console for network=default has 1 cluster-wide, 0 per-node configs, 0 template configs, making 1 (cluster) 0 (per node) and 0 (template) load balancers\\\\nI0121 11:12:07.747279 6924 services_controller.go:434] Service openshift-kube-storage-version-migrator-operator/metrics retrieved from lister for network=default: \\\\u0026Service{ObjectMeta:{metrics openshift-kube-storage-version-migrator-operator e1639a86-fb7f-46de-9d5e-4aee16dccea1 4372 0 2025-02-23 05:12:25 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[app:kube-storage-version-migrator-operator] map[include.release.openshift.io/hypershift:true include.release.openshift.io/ibm-cloud-managed:true include.release.openshift.io/self-managed-high-availability:true service.alpha.openshift.io/serving-cert-secret-name:serving-cert service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc007778417 \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:htt\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T11:12:07Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-htkvk_openshift-ovn-kubernetes(26b1108a-4b98-4e39-a3cd-e0d055089fd8)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948e0497eaffa7b0d84b4085c9ed8d987eb2c6527c9a82c10e576ed9c9c1e21b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-htkvk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:18Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:18 crc kubenswrapper[4824]: I0121 11:12:18.213535 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-np2v8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4698abe4-0f33-4349-a2a4-614de80ae21b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://191328471bb6f2b34cde2eb927da697619826bf0b6174ecb583c8c25d27a6654\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n9qbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://050bd5bdded2162d5e9dac81713f4ed63cb575d1776e828b2f98e50d89f11b04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n9qbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-np2v8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:18Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:18 crc kubenswrapper[4824]: I0121 11:12:18.253838 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:18 crc kubenswrapper[4824]: I0121 11:12:18.253870 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:18 crc kubenswrapper[4824]: I0121 11:12:18.253878 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:18 crc kubenswrapper[4824]: I0121 11:12:18.253893 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:18 crc kubenswrapper[4824]: I0121 11:12:18.253901 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:18Z","lastTransitionTime":"2026-01-21T11:12:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:18 crc kubenswrapper[4824]: I0121 11:12:18.355206 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:18 crc kubenswrapper[4824]: I0121 11:12:18.355241 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:18 crc kubenswrapper[4824]: I0121 11:12:18.355250 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:18 crc kubenswrapper[4824]: I0121 11:12:18.355261 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:18 crc kubenswrapper[4824]: I0121 11:12:18.355269 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:18Z","lastTransitionTime":"2026-01-21T11:12:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:18 crc kubenswrapper[4824]: I0121 11:12:18.457303 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:18 crc kubenswrapper[4824]: I0121 11:12:18.457352 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:18 crc kubenswrapper[4824]: I0121 11:12:18.457360 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:18 crc kubenswrapper[4824]: I0121 11:12:18.457376 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:18 crc kubenswrapper[4824]: I0121 11:12:18.457398 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:18Z","lastTransitionTime":"2026-01-21T11:12:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:18 crc kubenswrapper[4824]: I0121 11:12:18.559606 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:18 crc kubenswrapper[4824]: I0121 11:12:18.559638 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:18 crc kubenswrapper[4824]: I0121 11:12:18.559647 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:18 crc kubenswrapper[4824]: I0121 11:12:18.559677 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:18 crc kubenswrapper[4824]: I0121 11:12:18.559695 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:18Z","lastTransitionTime":"2026-01-21T11:12:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:18 crc kubenswrapper[4824]: I0121 11:12:18.662138 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:18 crc kubenswrapper[4824]: I0121 11:12:18.662190 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:18 crc kubenswrapper[4824]: I0121 11:12:18.662199 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:18 crc kubenswrapper[4824]: I0121 11:12:18.662214 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:18 crc kubenswrapper[4824]: I0121 11:12:18.662223 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:18Z","lastTransitionTime":"2026-01-21T11:12:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:18 crc kubenswrapper[4824]: I0121 11:12:18.764120 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:18 crc kubenswrapper[4824]: I0121 11:12:18.764152 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:18 crc kubenswrapper[4824]: I0121 11:12:18.764160 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:18 crc kubenswrapper[4824]: I0121 11:12:18.764172 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:18 crc kubenswrapper[4824]: I0121 11:12:18.764181 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:18Z","lastTransitionTime":"2026-01-21T11:12:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:18 crc kubenswrapper[4824]: I0121 11:12:18.866362 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:18 crc kubenswrapper[4824]: I0121 11:12:18.866422 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:18 crc kubenswrapper[4824]: I0121 11:12:18.866433 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:18 crc kubenswrapper[4824]: I0121 11:12:18.866445 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:18 crc kubenswrapper[4824]: I0121 11:12:18.866452 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:18Z","lastTransitionTime":"2026-01-21T11:12:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:18 crc kubenswrapper[4824]: I0121 11:12:18.875795 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 21 11:12:18 crc kubenswrapper[4824]: E0121 11:12:18.875920 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-21 11:13:22.875904663 +0000 UTC m=+145.168933955 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 11:12:18 crc kubenswrapper[4824]: I0121 11:12:18.875994 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 11:12:18 crc kubenswrapper[4824]: I0121 11:12:18.876016 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 11:12:18 crc kubenswrapper[4824]: E0121 11:12:18.876077 4824 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Jan 21 11:12:18 crc kubenswrapper[4824]: E0121 11:12:18.876119 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-21 11:13:22.876111963 +0000 UTC m=+145.169141255 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Jan 21 11:12:18 crc kubenswrapper[4824]: E0121 11:12:18.876276 4824 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 21 11:12:18 crc kubenswrapper[4824]: E0121 11:12:18.876422 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-21 11:13:22.876406517 +0000 UTC m=+145.169435810 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 21 11:12:18 crc kubenswrapper[4824]: I0121 11:12:18.968140 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:18 crc kubenswrapper[4824]: I0121 11:12:18.968398 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:18 crc kubenswrapper[4824]: I0121 11:12:18.968480 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:18 crc kubenswrapper[4824]: I0121 11:12:18.968550 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:18 crc kubenswrapper[4824]: I0121 11:12:18.968630 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:18Z","lastTransitionTime":"2026-01-21T11:12:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:18 crc kubenswrapper[4824]: I0121 11:12:18.976512 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 11:12:18 crc kubenswrapper[4824]: I0121 11:12:18.976556 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 11:12:18 crc kubenswrapper[4824]: E0121 11:12:18.976653 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 21 11:12:18 crc kubenswrapper[4824]: E0121 11:12:18.976674 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 21 11:12:18 crc kubenswrapper[4824]: E0121 11:12:18.976685 4824 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 21 11:12:18 crc kubenswrapper[4824]: E0121 11:12:18.976715 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-01-21 11:13:22.976704539 +0000 UTC m=+145.269733831 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 21 11:12:18 crc kubenswrapper[4824]: E0121 11:12:18.976912 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 21 11:12:18 crc kubenswrapper[4824]: E0121 11:12:18.977007 4824 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 21 11:12:18 crc kubenswrapper[4824]: E0121 11:12:18.977070 4824 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 21 11:12:18 crc kubenswrapper[4824]: E0121 11:12:18.977162 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-01-21 11:13:22.977149426 +0000 UTC m=+145.270178718 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 21 11:12:19 crc kubenswrapper[4824]: I0121 11:12:19.001601 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:19 crc kubenswrapper[4824]: I0121 11:12:19.001637 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:19 crc kubenswrapper[4824]: I0121 11:12:19.001645 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:19 crc kubenswrapper[4824]: I0121 11:12:19.001662 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:19 crc kubenswrapper[4824]: I0121 11:12:19.001671 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:19Z","lastTransitionTime":"2026-01-21T11:12:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:19 crc kubenswrapper[4824]: E0121 11:12:19.010511 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:12:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:12:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:12:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:12:19Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:12:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:12:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:12:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:12:19Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"695c676c-720c-4921-9e71-56b96f2fb46b\\\",\\\"systemUUID\\\":\\\"d00624f5-786f-41dc-bdc1-a1875f167d65\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:19Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:19 crc kubenswrapper[4824]: I0121 11:12:19.013160 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:19 crc kubenswrapper[4824]: I0121 11:12:19.013192 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:19 crc kubenswrapper[4824]: I0121 11:12:19.013200 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:19 crc kubenswrapper[4824]: I0121 11:12:19.013212 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:19 crc kubenswrapper[4824]: I0121 11:12:19.013220 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:19Z","lastTransitionTime":"2026-01-21T11:12:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:19 crc kubenswrapper[4824]: E0121 11:12:19.023117 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:12:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:12:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:12:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:12:19Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:12:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:12:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:12:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:12:19Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"695c676c-720c-4921-9e71-56b96f2fb46b\\\",\\\"systemUUID\\\":\\\"d00624f5-786f-41dc-bdc1-a1875f167d65\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:19Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:19 crc kubenswrapper[4824]: I0121 11:12:19.025917 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:19 crc kubenswrapper[4824]: I0121 11:12:19.025970 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:19 crc kubenswrapper[4824]: I0121 11:12:19.025980 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:19 crc kubenswrapper[4824]: I0121 11:12:19.025993 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:19 crc kubenswrapper[4824]: I0121 11:12:19.026001 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:19Z","lastTransitionTime":"2026-01-21T11:12:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:19 crc kubenswrapper[4824]: I0121 11:12:19.028788 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-28 15:37:34.072729343 +0000 UTC Jan 21 11:12:19 crc kubenswrapper[4824]: E0121 11:12:19.034032 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:12:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:12:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:12:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:12:19Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:12:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:12:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:12:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:12:19Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"695c676c-720c-4921-9e71-56b96f2fb46b\\\",\\\"systemUUID\\\":\\\"d00624f5-786f-41dc-bdc1-a1875f167d65\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:19Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:19 crc kubenswrapper[4824]: I0121 11:12:19.036376 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:19 crc kubenswrapper[4824]: I0121 11:12:19.036418 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:19 crc kubenswrapper[4824]: I0121 11:12:19.036427 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:19 crc kubenswrapper[4824]: I0121 11:12:19.036441 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:19 crc kubenswrapper[4824]: I0121 11:12:19.036449 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:19Z","lastTransitionTime":"2026-01-21T11:12:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:19 crc kubenswrapper[4824]: E0121 11:12:19.044767 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:12:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:12:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:12:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:12:19Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:12:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:12:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:12:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:12:19Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"695c676c-720c-4921-9e71-56b96f2fb46b\\\",\\\"systemUUID\\\":\\\"d00624f5-786f-41dc-bdc1-a1875f167d65\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:19Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:19 crc kubenswrapper[4824]: I0121 11:12:19.047074 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:19 crc kubenswrapper[4824]: I0121 11:12:19.047099 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:19 crc kubenswrapper[4824]: I0121 11:12:19.047108 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:19 crc kubenswrapper[4824]: I0121 11:12:19.047137 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:19 crc kubenswrapper[4824]: I0121 11:12:19.047146 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:19Z","lastTransitionTime":"2026-01-21T11:12:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:19 crc kubenswrapper[4824]: I0121 11:12:19.048225 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 11:12:19 crc kubenswrapper[4824]: I0121 11:12:19.048245 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 11:12:19 crc kubenswrapper[4824]: E0121 11:12:19.048326 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 21 11:12:19 crc kubenswrapper[4824]: E0121 11:12:19.048396 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 21 11:12:19 crc kubenswrapper[4824]: E0121 11:12:19.056004 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:12:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:12:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:12:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:12:19Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:12:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:12:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:12:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:12:19Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"695c676c-720c-4921-9e71-56b96f2fb46b\\\",\\\"systemUUID\\\":\\\"d00624f5-786f-41dc-bdc1-a1875f167d65\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:19Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:19 crc kubenswrapper[4824]: E0121 11:12:19.056117 4824 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Jan 21 11:12:19 crc kubenswrapper[4824]: I0121 11:12:19.070340 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:19 crc kubenswrapper[4824]: I0121 11:12:19.070372 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:19 crc kubenswrapper[4824]: I0121 11:12:19.070391 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:19 crc kubenswrapper[4824]: I0121 11:12:19.070405 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:19 crc kubenswrapper[4824]: I0121 11:12:19.070414 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:19Z","lastTransitionTime":"2026-01-21T11:12:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:19 crc kubenswrapper[4824]: I0121 11:12:19.172753 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:19 crc kubenswrapper[4824]: I0121 11:12:19.172996 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:19 crc kubenswrapper[4824]: I0121 11:12:19.173117 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:19 crc kubenswrapper[4824]: I0121 11:12:19.173205 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:19 crc kubenswrapper[4824]: I0121 11:12:19.173285 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:19Z","lastTransitionTime":"2026-01-21T11:12:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:19 crc kubenswrapper[4824]: I0121 11:12:19.275586 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:19 crc kubenswrapper[4824]: I0121 11:12:19.275855 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:19 crc kubenswrapper[4824]: I0121 11:12:19.275948 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:19 crc kubenswrapper[4824]: I0121 11:12:19.276043 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:19 crc kubenswrapper[4824]: I0121 11:12:19.276126 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:19Z","lastTransitionTime":"2026-01-21T11:12:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:19 crc kubenswrapper[4824]: I0121 11:12:19.377626 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:19 crc kubenswrapper[4824]: I0121 11:12:19.377844 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:19 crc kubenswrapper[4824]: I0121 11:12:19.377943 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:19 crc kubenswrapper[4824]: I0121 11:12:19.378057 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:19 crc kubenswrapper[4824]: I0121 11:12:19.378203 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:19Z","lastTransitionTime":"2026-01-21T11:12:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:19 crc kubenswrapper[4824]: I0121 11:12:19.479897 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:19 crc kubenswrapper[4824]: I0121 11:12:19.480123 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:19 crc kubenswrapper[4824]: I0121 11:12:19.480132 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:19 crc kubenswrapper[4824]: I0121 11:12:19.480144 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:19 crc kubenswrapper[4824]: I0121 11:12:19.480153 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:19Z","lastTransitionTime":"2026-01-21T11:12:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:19 crc kubenswrapper[4824]: I0121 11:12:19.581365 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:19 crc kubenswrapper[4824]: I0121 11:12:19.581393 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:19 crc kubenswrapper[4824]: I0121 11:12:19.581401 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:19 crc kubenswrapper[4824]: I0121 11:12:19.581410 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:19 crc kubenswrapper[4824]: I0121 11:12:19.581419 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:19Z","lastTransitionTime":"2026-01-21T11:12:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:19 crc kubenswrapper[4824]: I0121 11:12:19.682692 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:19 crc kubenswrapper[4824]: I0121 11:12:19.682803 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:19 crc kubenswrapper[4824]: I0121 11:12:19.682879 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:19 crc kubenswrapper[4824]: I0121 11:12:19.682971 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:19 crc kubenswrapper[4824]: I0121 11:12:19.683135 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:19Z","lastTransitionTime":"2026-01-21T11:12:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:19 crc kubenswrapper[4824]: I0121 11:12:19.785336 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:19 crc kubenswrapper[4824]: I0121 11:12:19.785368 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:19 crc kubenswrapper[4824]: I0121 11:12:19.785388 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:19 crc kubenswrapper[4824]: I0121 11:12:19.785400 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:19 crc kubenswrapper[4824]: I0121 11:12:19.785409 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:19Z","lastTransitionTime":"2026-01-21T11:12:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:19 crc kubenswrapper[4824]: I0121 11:12:19.887481 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:19 crc kubenswrapper[4824]: I0121 11:12:19.887512 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:19 crc kubenswrapper[4824]: I0121 11:12:19.887519 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:19 crc kubenswrapper[4824]: I0121 11:12:19.887533 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:19 crc kubenswrapper[4824]: I0121 11:12:19.887541 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:19Z","lastTransitionTime":"2026-01-21T11:12:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:19 crc kubenswrapper[4824]: I0121 11:12:19.989710 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:19 crc kubenswrapper[4824]: I0121 11:12:19.989742 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:19 crc kubenswrapper[4824]: I0121 11:12:19.989750 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:19 crc kubenswrapper[4824]: I0121 11:12:19.989764 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:19 crc kubenswrapper[4824]: I0121 11:12:19.989772 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:19Z","lastTransitionTime":"2026-01-21T11:12:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:20 crc kubenswrapper[4824]: I0121 11:12:20.029135 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-22 02:02:09.851170023 +0000 UTC Jan 21 11:12:20 crc kubenswrapper[4824]: I0121 11:12:20.048403 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9rmnw" Jan 21 11:12:20 crc kubenswrapper[4824]: I0121 11:12:20.048454 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 11:12:20 crc kubenswrapper[4824]: E0121 11:12:20.048501 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9rmnw" podUID="8f9bd12b-22ee-44ed-a8b9-c100d043f691" Jan 21 11:12:20 crc kubenswrapper[4824]: E0121 11:12:20.048545 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 21 11:12:20 crc kubenswrapper[4824]: I0121 11:12:20.049186 4824 scope.go:117] "RemoveContainer" containerID="86a6881f109dab1f910fe505a21c7c800e587de4dded2635147997c51c05c483" Jan 21 11:12:20 crc kubenswrapper[4824]: E0121 11:12:20.049316 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-htkvk_openshift-ovn-kubernetes(26b1108a-4b98-4e39-a3cd-e0d055089fd8)\"" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" podUID="26b1108a-4b98-4e39-a3cd-e0d055089fd8" Jan 21 11:12:20 crc kubenswrapper[4824]: I0121 11:12:20.090916 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:20 crc kubenswrapper[4824]: I0121 11:12:20.090943 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:20 crc kubenswrapper[4824]: I0121 11:12:20.090951 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:20 crc kubenswrapper[4824]: I0121 11:12:20.090981 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:20 crc kubenswrapper[4824]: I0121 11:12:20.090991 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:20Z","lastTransitionTime":"2026-01-21T11:12:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:20 crc kubenswrapper[4824]: I0121 11:12:20.192883 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:20 crc kubenswrapper[4824]: I0121 11:12:20.192913 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:20 crc kubenswrapper[4824]: I0121 11:12:20.192921 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:20 crc kubenswrapper[4824]: I0121 11:12:20.192932 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:20 crc kubenswrapper[4824]: I0121 11:12:20.192941 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:20Z","lastTransitionTime":"2026-01-21T11:12:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:20 crc kubenswrapper[4824]: I0121 11:12:20.294697 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:20 crc kubenswrapper[4824]: I0121 11:12:20.294724 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:20 crc kubenswrapper[4824]: I0121 11:12:20.294732 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:20 crc kubenswrapper[4824]: I0121 11:12:20.294742 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:20 crc kubenswrapper[4824]: I0121 11:12:20.294749 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:20Z","lastTransitionTime":"2026-01-21T11:12:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:20 crc kubenswrapper[4824]: I0121 11:12:20.396170 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:20 crc kubenswrapper[4824]: I0121 11:12:20.396198 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:20 crc kubenswrapper[4824]: I0121 11:12:20.396206 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:20 crc kubenswrapper[4824]: I0121 11:12:20.396216 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:20 crc kubenswrapper[4824]: I0121 11:12:20.396223 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:20Z","lastTransitionTime":"2026-01-21T11:12:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:20 crc kubenswrapper[4824]: I0121 11:12:20.498044 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:20 crc kubenswrapper[4824]: I0121 11:12:20.498098 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:20 crc kubenswrapper[4824]: I0121 11:12:20.498107 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:20 crc kubenswrapper[4824]: I0121 11:12:20.498118 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:20 crc kubenswrapper[4824]: I0121 11:12:20.498126 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:20Z","lastTransitionTime":"2026-01-21T11:12:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:20 crc kubenswrapper[4824]: I0121 11:12:20.599629 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:20 crc kubenswrapper[4824]: I0121 11:12:20.599656 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:20 crc kubenswrapper[4824]: I0121 11:12:20.599663 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:20 crc kubenswrapper[4824]: I0121 11:12:20.599674 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:20 crc kubenswrapper[4824]: I0121 11:12:20.599683 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:20Z","lastTransitionTime":"2026-01-21T11:12:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:20 crc kubenswrapper[4824]: I0121 11:12:20.701692 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:20 crc kubenswrapper[4824]: I0121 11:12:20.701723 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:20 crc kubenswrapper[4824]: I0121 11:12:20.701732 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:20 crc kubenswrapper[4824]: I0121 11:12:20.701744 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:20 crc kubenswrapper[4824]: I0121 11:12:20.701754 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:20Z","lastTransitionTime":"2026-01-21T11:12:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:20 crc kubenswrapper[4824]: I0121 11:12:20.802884 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:20 crc kubenswrapper[4824]: I0121 11:12:20.802925 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:20 crc kubenswrapper[4824]: I0121 11:12:20.802937 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:20 crc kubenswrapper[4824]: I0121 11:12:20.802982 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:20 crc kubenswrapper[4824]: I0121 11:12:20.802993 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:20Z","lastTransitionTime":"2026-01-21T11:12:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:20 crc kubenswrapper[4824]: I0121 11:12:20.905076 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:20 crc kubenswrapper[4824]: I0121 11:12:20.905105 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:20 crc kubenswrapper[4824]: I0121 11:12:20.905134 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:20 crc kubenswrapper[4824]: I0121 11:12:20.905146 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:20 crc kubenswrapper[4824]: I0121 11:12:20.905154 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:20Z","lastTransitionTime":"2026-01-21T11:12:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:21 crc kubenswrapper[4824]: I0121 11:12:21.006764 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:21 crc kubenswrapper[4824]: I0121 11:12:21.006796 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:21 crc kubenswrapper[4824]: I0121 11:12:21.006806 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:21 crc kubenswrapper[4824]: I0121 11:12:21.006818 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:21 crc kubenswrapper[4824]: I0121 11:12:21.006846 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:21Z","lastTransitionTime":"2026-01-21T11:12:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:21 crc kubenswrapper[4824]: I0121 11:12:21.030174 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-12 15:21:32.155937181 +0000 UTC Jan 21 11:12:21 crc kubenswrapper[4824]: I0121 11:12:21.048451 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 11:12:21 crc kubenswrapper[4824]: E0121 11:12:21.048528 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 21 11:12:21 crc kubenswrapper[4824]: I0121 11:12:21.048616 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 11:12:21 crc kubenswrapper[4824]: E0121 11:12:21.048746 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 21 11:12:21 crc kubenswrapper[4824]: I0121 11:12:21.108697 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:21 crc kubenswrapper[4824]: I0121 11:12:21.108734 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:21 crc kubenswrapper[4824]: I0121 11:12:21.108744 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:21 crc kubenswrapper[4824]: I0121 11:12:21.108757 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:21 crc kubenswrapper[4824]: I0121 11:12:21.108766 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:21Z","lastTransitionTime":"2026-01-21T11:12:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:21 crc kubenswrapper[4824]: I0121 11:12:21.209927 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:21 crc kubenswrapper[4824]: I0121 11:12:21.209969 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:21 crc kubenswrapper[4824]: I0121 11:12:21.209978 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:21 crc kubenswrapper[4824]: I0121 11:12:21.209988 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:21 crc kubenswrapper[4824]: I0121 11:12:21.209996 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:21Z","lastTransitionTime":"2026-01-21T11:12:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:21 crc kubenswrapper[4824]: I0121 11:12:21.312188 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:21 crc kubenswrapper[4824]: I0121 11:12:21.312841 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:21 crc kubenswrapper[4824]: I0121 11:12:21.312988 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:21 crc kubenswrapper[4824]: I0121 11:12:21.313122 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:21 crc kubenswrapper[4824]: I0121 11:12:21.313187 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:21Z","lastTransitionTime":"2026-01-21T11:12:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:21 crc kubenswrapper[4824]: I0121 11:12:21.415701 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:21 crc kubenswrapper[4824]: I0121 11:12:21.415913 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:21 crc kubenswrapper[4824]: I0121 11:12:21.415996 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:21 crc kubenswrapper[4824]: I0121 11:12:21.416062 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:21 crc kubenswrapper[4824]: I0121 11:12:21.416132 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:21Z","lastTransitionTime":"2026-01-21T11:12:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:21 crc kubenswrapper[4824]: I0121 11:12:21.518421 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:21 crc kubenswrapper[4824]: I0121 11:12:21.518664 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:21 crc kubenswrapper[4824]: I0121 11:12:21.518754 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:21 crc kubenswrapper[4824]: I0121 11:12:21.518820 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:21 crc kubenswrapper[4824]: I0121 11:12:21.518880 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:21Z","lastTransitionTime":"2026-01-21T11:12:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:21 crc kubenswrapper[4824]: I0121 11:12:21.621575 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:21 crc kubenswrapper[4824]: I0121 11:12:21.621609 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:21 crc kubenswrapper[4824]: I0121 11:12:21.621617 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:21 crc kubenswrapper[4824]: I0121 11:12:21.621632 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:21 crc kubenswrapper[4824]: I0121 11:12:21.621641 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:21Z","lastTransitionTime":"2026-01-21T11:12:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:21 crc kubenswrapper[4824]: I0121 11:12:21.723435 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:21 crc kubenswrapper[4824]: I0121 11:12:21.723471 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:21 crc kubenswrapper[4824]: I0121 11:12:21.723481 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:21 crc kubenswrapper[4824]: I0121 11:12:21.723501 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:21 crc kubenswrapper[4824]: I0121 11:12:21.723509 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:21Z","lastTransitionTime":"2026-01-21T11:12:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:21 crc kubenswrapper[4824]: I0121 11:12:21.825160 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:21 crc kubenswrapper[4824]: I0121 11:12:21.825193 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:21 crc kubenswrapper[4824]: I0121 11:12:21.825203 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:21 crc kubenswrapper[4824]: I0121 11:12:21.825217 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:21 crc kubenswrapper[4824]: I0121 11:12:21.825225 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:21Z","lastTransitionTime":"2026-01-21T11:12:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:21 crc kubenswrapper[4824]: I0121 11:12:21.927611 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:21 crc kubenswrapper[4824]: I0121 11:12:21.927869 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:21 crc kubenswrapper[4824]: I0121 11:12:21.927953 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:21 crc kubenswrapper[4824]: I0121 11:12:21.928050 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:21 crc kubenswrapper[4824]: I0121 11:12:21.928113 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:21Z","lastTransitionTime":"2026-01-21T11:12:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:22 crc kubenswrapper[4824]: I0121 11:12:22.029900 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:22 crc kubenswrapper[4824]: I0121 11:12:22.029933 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:22 crc kubenswrapper[4824]: I0121 11:12:22.029942 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:22 crc kubenswrapper[4824]: I0121 11:12:22.029977 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:22 crc kubenswrapper[4824]: I0121 11:12:22.029986 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:22Z","lastTransitionTime":"2026-01-21T11:12:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:22 crc kubenswrapper[4824]: I0121 11:12:22.030256 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-24 13:55:53.470736332 +0000 UTC Jan 21 11:12:22 crc kubenswrapper[4824]: I0121 11:12:22.048479 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9rmnw" Jan 21 11:12:22 crc kubenswrapper[4824]: I0121 11:12:22.048484 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 11:12:22 crc kubenswrapper[4824]: E0121 11:12:22.048773 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 21 11:12:22 crc kubenswrapper[4824]: E0121 11:12:22.048853 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9rmnw" podUID="8f9bd12b-22ee-44ed-a8b9-c100d043f691" Jan 21 11:12:22 crc kubenswrapper[4824]: I0121 11:12:22.132049 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:22 crc kubenswrapper[4824]: I0121 11:12:22.132081 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:22 crc kubenswrapper[4824]: I0121 11:12:22.132093 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:22 crc kubenswrapper[4824]: I0121 11:12:22.132104 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:22 crc kubenswrapper[4824]: I0121 11:12:22.132112 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:22Z","lastTransitionTime":"2026-01-21T11:12:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:22 crc kubenswrapper[4824]: I0121 11:12:22.233635 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:22 crc kubenswrapper[4824]: I0121 11:12:22.233666 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:22 crc kubenswrapper[4824]: I0121 11:12:22.233673 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:22 crc kubenswrapper[4824]: I0121 11:12:22.233684 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:22 crc kubenswrapper[4824]: I0121 11:12:22.233693 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:22Z","lastTransitionTime":"2026-01-21T11:12:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:22 crc kubenswrapper[4824]: I0121 11:12:22.335787 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:22 crc kubenswrapper[4824]: I0121 11:12:22.335813 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:22 crc kubenswrapper[4824]: I0121 11:12:22.335820 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:22 crc kubenswrapper[4824]: I0121 11:12:22.335832 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:22 crc kubenswrapper[4824]: I0121 11:12:22.335839 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:22Z","lastTransitionTime":"2026-01-21T11:12:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:22 crc kubenswrapper[4824]: I0121 11:12:22.437607 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:22 crc kubenswrapper[4824]: I0121 11:12:22.437658 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:22 crc kubenswrapper[4824]: I0121 11:12:22.437667 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:22 crc kubenswrapper[4824]: I0121 11:12:22.437682 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:22 crc kubenswrapper[4824]: I0121 11:12:22.437690 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:22Z","lastTransitionTime":"2026-01-21T11:12:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:22 crc kubenswrapper[4824]: I0121 11:12:22.539739 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:22 crc kubenswrapper[4824]: I0121 11:12:22.539781 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:22 crc kubenswrapper[4824]: I0121 11:12:22.539790 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:22 crc kubenswrapper[4824]: I0121 11:12:22.539804 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:22 crc kubenswrapper[4824]: I0121 11:12:22.539813 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:22Z","lastTransitionTime":"2026-01-21T11:12:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:22 crc kubenswrapper[4824]: I0121 11:12:22.641409 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:22 crc kubenswrapper[4824]: I0121 11:12:22.641447 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:22 crc kubenswrapper[4824]: I0121 11:12:22.641457 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:22 crc kubenswrapper[4824]: I0121 11:12:22.641471 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:22 crc kubenswrapper[4824]: I0121 11:12:22.641480 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:22Z","lastTransitionTime":"2026-01-21T11:12:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:22 crc kubenswrapper[4824]: I0121 11:12:22.743084 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:22 crc kubenswrapper[4824]: I0121 11:12:22.743122 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:22 crc kubenswrapper[4824]: I0121 11:12:22.743131 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:22 crc kubenswrapper[4824]: I0121 11:12:22.743145 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:22 crc kubenswrapper[4824]: I0121 11:12:22.743154 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:22Z","lastTransitionTime":"2026-01-21T11:12:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:22 crc kubenswrapper[4824]: I0121 11:12:22.844832 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:22 crc kubenswrapper[4824]: I0121 11:12:22.844875 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:22 crc kubenswrapper[4824]: I0121 11:12:22.844884 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:22 crc kubenswrapper[4824]: I0121 11:12:22.844895 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:22 crc kubenswrapper[4824]: I0121 11:12:22.844903 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:22Z","lastTransitionTime":"2026-01-21T11:12:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:22 crc kubenswrapper[4824]: I0121 11:12:22.946519 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:22 crc kubenswrapper[4824]: I0121 11:12:22.946547 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:22 crc kubenswrapper[4824]: I0121 11:12:22.946559 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:22 crc kubenswrapper[4824]: I0121 11:12:22.946570 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:22 crc kubenswrapper[4824]: I0121 11:12:22.946579 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:22Z","lastTransitionTime":"2026-01-21T11:12:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:23 crc kubenswrapper[4824]: I0121 11:12:23.031150 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-16 20:58:29.913446283 +0000 UTC Jan 21 11:12:23 crc kubenswrapper[4824]: I0121 11:12:23.048269 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 11:12:23 crc kubenswrapper[4824]: I0121 11:12:23.048316 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 11:12:23 crc kubenswrapper[4824]: E0121 11:12:23.048423 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 21 11:12:23 crc kubenswrapper[4824]: E0121 11:12:23.048495 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 21 11:12:23 crc kubenswrapper[4824]: I0121 11:12:23.048741 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:23 crc kubenswrapper[4824]: I0121 11:12:23.048769 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:23 crc kubenswrapper[4824]: I0121 11:12:23.048777 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:23 crc kubenswrapper[4824]: I0121 11:12:23.048788 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:23 crc kubenswrapper[4824]: I0121 11:12:23.048796 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:23Z","lastTransitionTime":"2026-01-21T11:12:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:23 crc kubenswrapper[4824]: I0121 11:12:23.150351 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:23 crc kubenswrapper[4824]: I0121 11:12:23.150398 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:23 crc kubenswrapper[4824]: I0121 11:12:23.150406 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:23 crc kubenswrapper[4824]: I0121 11:12:23.150418 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:23 crc kubenswrapper[4824]: I0121 11:12:23.150426 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:23Z","lastTransitionTime":"2026-01-21T11:12:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:23 crc kubenswrapper[4824]: I0121 11:12:23.252389 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:23 crc kubenswrapper[4824]: I0121 11:12:23.252421 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:23 crc kubenswrapper[4824]: I0121 11:12:23.252429 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:23 crc kubenswrapper[4824]: I0121 11:12:23.252441 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:23 crc kubenswrapper[4824]: I0121 11:12:23.252449 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:23Z","lastTransitionTime":"2026-01-21T11:12:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:23 crc kubenswrapper[4824]: I0121 11:12:23.354830 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:23 crc kubenswrapper[4824]: I0121 11:12:23.354901 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:23 crc kubenswrapper[4824]: I0121 11:12:23.354910 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:23 crc kubenswrapper[4824]: I0121 11:12:23.354921 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:23 crc kubenswrapper[4824]: I0121 11:12:23.354929 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:23Z","lastTransitionTime":"2026-01-21T11:12:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:23 crc kubenswrapper[4824]: I0121 11:12:23.457033 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:23 crc kubenswrapper[4824]: I0121 11:12:23.457062 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:23 crc kubenswrapper[4824]: I0121 11:12:23.457070 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:23 crc kubenswrapper[4824]: I0121 11:12:23.457080 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:23 crc kubenswrapper[4824]: I0121 11:12:23.457090 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:23Z","lastTransitionTime":"2026-01-21T11:12:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:23 crc kubenswrapper[4824]: I0121 11:12:23.558534 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:23 crc kubenswrapper[4824]: I0121 11:12:23.558564 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:23 crc kubenswrapper[4824]: I0121 11:12:23.558571 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:23 crc kubenswrapper[4824]: I0121 11:12:23.558582 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:23 crc kubenswrapper[4824]: I0121 11:12:23.558590 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:23Z","lastTransitionTime":"2026-01-21T11:12:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:23 crc kubenswrapper[4824]: I0121 11:12:23.660554 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:23 crc kubenswrapper[4824]: I0121 11:12:23.660592 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:23 crc kubenswrapper[4824]: I0121 11:12:23.660600 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:23 crc kubenswrapper[4824]: I0121 11:12:23.660668 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:23 crc kubenswrapper[4824]: I0121 11:12:23.660676 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:23Z","lastTransitionTime":"2026-01-21T11:12:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:23 crc kubenswrapper[4824]: I0121 11:12:23.762837 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:23 crc kubenswrapper[4824]: I0121 11:12:23.762887 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:23 crc kubenswrapper[4824]: I0121 11:12:23.762897 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:23 crc kubenswrapper[4824]: I0121 11:12:23.762909 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:23 crc kubenswrapper[4824]: I0121 11:12:23.762917 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:23Z","lastTransitionTime":"2026-01-21T11:12:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:23 crc kubenswrapper[4824]: I0121 11:12:23.865381 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:23 crc kubenswrapper[4824]: I0121 11:12:23.865431 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:23 crc kubenswrapper[4824]: I0121 11:12:23.865440 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:23 crc kubenswrapper[4824]: I0121 11:12:23.865453 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:23 crc kubenswrapper[4824]: I0121 11:12:23.865461 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:23Z","lastTransitionTime":"2026-01-21T11:12:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:23 crc kubenswrapper[4824]: I0121 11:12:23.967767 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:23 crc kubenswrapper[4824]: I0121 11:12:23.967798 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:23 crc kubenswrapper[4824]: I0121 11:12:23.967807 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:23 crc kubenswrapper[4824]: I0121 11:12:23.967819 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:23 crc kubenswrapper[4824]: I0121 11:12:23.967827 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:23Z","lastTransitionTime":"2026-01-21T11:12:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:24 crc kubenswrapper[4824]: I0121 11:12:24.031689 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-17 08:59:56.766731956 +0000 UTC Jan 21 11:12:24 crc kubenswrapper[4824]: I0121 11:12:24.048986 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9rmnw" Jan 21 11:12:24 crc kubenswrapper[4824]: I0121 11:12:24.049058 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 11:12:24 crc kubenswrapper[4824]: E0121 11:12:24.049109 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9rmnw" podUID="8f9bd12b-22ee-44ed-a8b9-c100d043f691" Jan 21 11:12:24 crc kubenswrapper[4824]: E0121 11:12:24.049240 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 21 11:12:24 crc kubenswrapper[4824]: I0121 11:12:24.070194 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:24 crc kubenswrapper[4824]: I0121 11:12:24.070232 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:24 crc kubenswrapper[4824]: I0121 11:12:24.070242 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:24 crc kubenswrapper[4824]: I0121 11:12:24.070256 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:24 crc kubenswrapper[4824]: I0121 11:12:24.070266 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:24Z","lastTransitionTime":"2026-01-21T11:12:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:24 crc kubenswrapper[4824]: I0121 11:12:24.171893 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:24 crc kubenswrapper[4824]: I0121 11:12:24.171919 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:24 crc kubenswrapper[4824]: I0121 11:12:24.171927 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:24 crc kubenswrapper[4824]: I0121 11:12:24.171937 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:24 crc kubenswrapper[4824]: I0121 11:12:24.171945 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:24Z","lastTransitionTime":"2026-01-21T11:12:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:24 crc kubenswrapper[4824]: I0121 11:12:24.273886 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:24 crc kubenswrapper[4824]: I0121 11:12:24.273926 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:24 crc kubenswrapper[4824]: I0121 11:12:24.273935 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:24 crc kubenswrapper[4824]: I0121 11:12:24.273949 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:24 crc kubenswrapper[4824]: I0121 11:12:24.273976 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:24Z","lastTransitionTime":"2026-01-21T11:12:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:24 crc kubenswrapper[4824]: I0121 11:12:24.376211 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:24 crc kubenswrapper[4824]: I0121 11:12:24.376264 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:24 crc kubenswrapper[4824]: I0121 11:12:24.376275 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:24 crc kubenswrapper[4824]: I0121 11:12:24.376291 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:24 crc kubenswrapper[4824]: I0121 11:12:24.376306 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:24Z","lastTransitionTime":"2026-01-21T11:12:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:24 crc kubenswrapper[4824]: I0121 11:12:24.477741 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:24 crc kubenswrapper[4824]: I0121 11:12:24.477793 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:24 crc kubenswrapper[4824]: I0121 11:12:24.477803 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:24 crc kubenswrapper[4824]: I0121 11:12:24.477819 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:24 crc kubenswrapper[4824]: I0121 11:12:24.477829 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:24Z","lastTransitionTime":"2026-01-21T11:12:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:24 crc kubenswrapper[4824]: I0121 11:12:24.579626 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:24 crc kubenswrapper[4824]: I0121 11:12:24.579654 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:24 crc kubenswrapper[4824]: I0121 11:12:24.579662 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:24 crc kubenswrapper[4824]: I0121 11:12:24.579673 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:24 crc kubenswrapper[4824]: I0121 11:12:24.579682 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:24Z","lastTransitionTime":"2026-01-21T11:12:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:24 crc kubenswrapper[4824]: I0121 11:12:24.681239 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:24 crc kubenswrapper[4824]: I0121 11:12:24.681283 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:24 crc kubenswrapper[4824]: I0121 11:12:24.681292 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:24 crc kubenswrapper[4824]: I0121 11:12:24.681307 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:24 crc kubenswrapper[4824]: I0121 11:12:24.681316 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:24Z","lastTransitionTime":"2026-01-21T11:12:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:24 crc kubenswrapper[4824]: I0121 11:12:24.782625 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:24 crc kubenswrapper[4824]: I0121 11:12:24.782660 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:24 crc kubenswrapper[4824]: I0121 11:12:24.782667 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:24 crc kubenswrapper[4824]: I0121 11:12:24.782679 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:24 crc kubenswrapper[4824]: I0121 11:12:24.782687 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:24Z","lastTransitionTime":"2026-01-21T11:12:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:24 crc kubenswrapper[4824]: I0121 11:12:24.884671 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:24 crc kubenswrapper[4824]: I0121 11:12:24.884708 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:24 crc kubenswrapper[4824]: I0121 11:12:24.884717 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:24 crc kubenswrapper[4824]: I0121 11:12:24.884732 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:24 crc kubenswrapper[4824]: I0121 11:12:24.884740 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:24Z","lastTransitionTime":"2026-01-21T11:12:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:24 crc kubenswrapper[4824]: I0121 11:12:24.986819 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:24 crc kubenswrapper[4824]: I0121 11:12:24.986853 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:24 crc kubenswrapper[4824]: I0121 11:12:24.986862 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:24 crc kubenswrapper[4824]: I0121 11:12:24.986876 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:24 crc kubenswrapper[4824]: I0121 11:12:24.986885 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:24Z","lastTransitionTime":"2026-01-21T11:12:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:25 crc kubenswrapper[4824]: I0121 11:12:25.032561 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-27 05:37:05.730617272 +0000 UTC Jan 21 11:12:25 crc kubenswrapper[4824]: I0121 11:12:25.048788 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 11:12:25 crc kubenswrapper[4824]: I0121 11:12:25.048788 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 11:12:25 crc kubenswrapper[4824]: E0121 11:12:25.048882 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 21 11:12:25 crc kubenswrapper[4824]: E0121 11:12:25.048931 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 21 11:12:25 crc kubenswrapper[4824]: I0121 11:12:25.088337 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:25 crc kubenswrapper[4824]: I0121 11:12:25.088373 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:25 crc kubenswrapper[4824]: I0121 11:12:25.088382 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:25 crc kubenswrapper[4824]: I0121 11:12:25.088391 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:25 crc kubenswrapper[4824]: I0121 11:12:25.088398 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:25Z","lastTransitionTime":"2026-01-21T11:12:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:25 crc kubenswrapper[4824]: I0121 11:12:25.189889 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:25 crc kubenswrapper[4824]: I0121 11:12:25.189919 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:25 crc kubenswrapper[4824]: I0121 11:12:25.189928 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:25 crc kubenswrapper[4824]: I0121 11:12:25.189936 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:25 crc kubenswrapper[4824]: I0121 11:12:25.189943 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:25Z","lastTransitionTime":"2026-01-21T11:12:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:25 crc kubenswrapper[4824]: I0121 11:12:25.291494 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:25 crc kubenswrapper[4824]: I0121 11:12:25.291537 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:25 crc kubenswrapper[4824]: I0121 11:12:25.291547 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:25 crc kubenswrapper[4824]: I0121 11:12:25.291560 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:25 crc kubenswrapper[4824]: I0121 11:12:25.291570 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:25Z","lastTransitionTime":"2026-01-21T11:12:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:25 crc kubenswrapper[4824]: I0121 11:12:25.393432 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:25 crc kubenswrapper[4824]: I0121 11:12:25.393467 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:25 crc kubenswrapper[4824]: I0121 11:12:25.393477 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:25 crc kubenswrapper[4824]: I0121 11:12:25.393488 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:25 crc kubenswrapper[4824]: I0121 11:12:25.393498 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:25Z","lastTransitionTime":"2026-01-21T11:12:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:25 crc kubenswrapper[4824]: I0121 11:12:25.494636 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:25 crc kubenswrapper[4824]: I0121 11:12:25.494680 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:25 crc kubenswrapper[4824]: I0121 11:12:25.494688 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:25 crc kubenswrapper[4824]: I0121 11:12:25.494702 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:25 crc kubenswrapper[4824]: I0121 11:12:25.494711 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:25Z","lastTransitionTime":"2026-01-21T11:12:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:25 crc kubenswrapper[4824]: I0121 11:12:25.596477 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:25 crc kubenswrapper[4824]: I0121 11:12:25.596502 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:25 crc kubenswrapper[4824]: I0121 11:12:25.596509 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:25 crc kubenswrapper[4824]: I0121 11:12:25.596519 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:25 crc kubenswrapper[4824]: I0121 11:12:25.596527 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:25Z","lastTransitionTime":"2026-01-21T11:12:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:25 crc kubenswrapper[4824]: I0121 11:12:25.698439 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:25 crc kubenswrapper[4824]: I0121 11:12:25.698469 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:25 crc kubenswrapper[4824]: I0121 11:12:25.698478 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:25 crc kubenswrapper[4824]: I0121 11:12:25.698489 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:25 crc kubenswrapper[4824]: I0121 11:12:25.698496 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:25Z","lastTransitionTime":"2026-01-21T11:12:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:25 crc kubenswrapper[4824]: I0121 11:12:25.800140 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:25 crc kubenswrapper[4824]: I0121 11:12:25.800171 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:25 crc kubenswrapper[4824]: I0121 11:12:25.800179 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:25 crc kubenswrapper[4824]: I0121 11:12:25.800188 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:25 crc kubenswrapper[4824]: I0121 11:12:25.800195 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:25Z","lastTransitionTime":"2026-01-21T11:12:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:25 crc kubenswrapper[4824]: I0121 11:12:25.902068 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:25 crc kubenswrapper[4824]: I0121 11:12:25.902096 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:25 crc kubenswrapper[4824]: I0121 11:12:25.902104 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:25 crc kubenswrapper[4824]: I0121 11:12:25.902114 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:25 crc kubenswrapper[4824]: I0121 11:12:25.902121 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:25Z","lastTransitionTime":"2026-01-21T11:12:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:26 crc kubenswrapper[4824]: I0121 11:12:26.003451 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:26 crc kubenswrapper[4824]: I0121 11:12:26.003481 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:26 crc kubenswrapper[4824]: I0121 11:12:26.003489 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:26 crc kubenswrapper[4824]: I0121 11:12:26.003500 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:26 crc kubenswrapper[4824]: I0121 11:12:26.003525 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:26Z","lastTransitionTime":"2026-01-21T11:12:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:26 crc kubenswrapper[4824]: I0121 11:12:26.032940 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-18 19:58:06.973520129 +0000 UTC Jan 21 11:12:26 crc kubenswrapper[4824]: I0121 11:12:26.049254 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9rmnw" Jan 21 11:12:26 crc kubenswrapper[4824]: E0121 11:12:26.049378 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9rmnw" podUID="8f9bd12b-22ee-44ed-a8b9-c100d043f691" Jan 21 11:12:26 crc kubenswrapper[4824]: I0121 11:12:26.049464 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 11:12:26 crc kubenswrapper[4824]: E0121 11:12:26.049617 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 21 11:12:26 crc kubenswrapper[4824]: I0121 11:12:26.105286 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:26 crc kubenswrapper[4824]: I0121 11:12:26.105320 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:26 crc kubenswrapper[4824]: I0121 11:12:26.105329 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:26 crc kubenswrapper[4824]: I0121 11:12:26.105342 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:26 crc kubenswrapper[4824]: I0121 11:12:26.105353 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:26Z","lastTransitionTime":"2026-01-21T11:12:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:26 crc kubenswrapper[4824]: I0121 11:12:26.207028 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:26 crc kubenswrapper[4824]: I0121 11:12:26.207065 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:26 crc kubenswrapper[4824]: I0121 11:12:26.207073 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:26 crc kubenswrapper[4824]: I0121 11:12:26.207084 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:26 crc kubenswrapper[4824]: I0121 11:12:26.207092 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:26Z","lastTransitionTime":"2026-01-21T11:12:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:26 crc kubenswrapper[4824]: I0121 11:12:26.309552 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:26 crc kubenswrapper[4824]: I0121 11:12:26.309597 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:26 crc kubenswrapper[4824]: I0121 11:12:26.309607 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:26 crc kubenswrapper[4824]: I0121 11:12:26.309621 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:26 crc kubenswrapper[4824]: I0121 11:12:26.309631 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:26Z","lastTransitionTime":"2026-01-21T11:12:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:26 crc kubenswrapper[4824]: I0121 11:12:26.411553 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:26 crc kubenswrapper[4824]: I0121 11:12:26.411586 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:26 crc kubenswrapper[4824]: I0121 11:12:26.411594 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:26 crc kubenswrapper[4824]: I0121 11:12:26.411608 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:26 crc kubenswrapper[4824]: I0121 11:12:26.411618 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:26Z","lastTransitionTime":"2026-01-21T11:12:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:26 crc kubenswrapper[4824]: I0121 11:12:26.513002 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:26 crc kubenswrapper[4824]: I0121 11:12:26.513038 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:26 crc kubenswrapper[4824]: I0121 11:12:26.513048 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:26 crc kubenswrapper[4824]: I0121 11:12:26.513062 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:26 crc kubenswrapper[4824]: I0121 11:12:26.513070 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:26Z","lastTransitionTime":"2026-01-21T11:12:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:26 crc kubenswrapper[4824]: I0121 11:12:26.614674 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:26 crc kubenswrapper[4824]: I0121 11:12:26.614731 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:26 crc kubenswrapper[4824]: I0121 11:12:26.614740 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:26 crc kubenswrapper[4824]: I0121 11:12:26.614755 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:26 crc kubenswrapper[4824]: I0121 11:12:26.614764 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:26Z","lastTransitionTime":"2026-01-21T11:12:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:26 crc kubenswrapper[4824]: I0121 11:12:26.716467 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:26 crc kubenswrapper[4824]: I0121 11:12:26.716495 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:26 crc kubenswrapper[4824]: I0121 11:12:26.716503 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:26 crc kubenswrapper[4824]: I0121 11:12:26.716513 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:26 crc kubenswrapper[4824]: I0121 11:12:26.716521 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:26Z","lastTransitionTime":"2026-01-21T11:12:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:26 crc kubenswrapper[4824]: I0121 11:12:26.818033 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:26 crc kubenswrapper[4824]: I0121 11:12:26.818066 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:26 crc kubenswrapper[4824]: I0121 11:12:26.818075 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:26 crc kubenswrapper[4824]: I0121 11:12:26.818087 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:26 crc kubenswrapper[4824]: I0121 11:12:26.818095 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:26Z","lastTransitionTime":"2026-01-21T11:12:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:26 crc kubenswrapper[4824]: I0121 11:12:26.920216 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:26 crc kubenswrapper[4824]: I0121 11:12:26.920253 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:26 crc kubenswrapper[4824]: I0121 11:12:26.920262 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:26 crc kubenswrapper[4824]: I0121 11:12:26.920275 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:26 crc kubenswrapper[4824]: I0121 11:12:26.920283 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:26Z","lastTransitionTime":"2026-01-21T11:12:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:27 crc kubenswrapper[4824]: I0121 11:12:27.022137 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:27 crc kubenswrapper[4824]: I0121 11:12:27.022172 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:27 crc kubenswrapper[4824]: I0121 11:12:27.022180 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:27 crc kubenswrapper[4824]: I0121 11:12:27.022194 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:27 crc kubenswrapper[4824]: I0121 11:12:27.022202 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:27Z","lastTransitionTime":"2026-01-21T11:12:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:27 crc kubenswrapper[4824]: I0121 11:12:27.033397 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-31 13:34:16.597437611 +0000 UTC Jan 21 11:12:27 crc kubenswrapper[4824]: I0121 11:12:27.048874 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 11:12:27 crc kubenswrapper[4824]: I0121 11:12:27.048893 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 11:12:27 crc kubenswrapper[4824]: E0121 11:12:27.049156 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 21 11:12:27 crc kubenswrapper[4824]: E0121 11:12:27.049069 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 21 11:12:27 crc kubenswrapper[4824]: I0121 11:12:27.124015 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:27 crc kubenswrapper[4824]: I0121 11:12:27.124049 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:27 crc kubenswrapper[4824]: I0121 11:12:27.124057 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:27 crc kubenswrapper[4824]: I0121 11:12:27.124068 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:27 crc kubenswrapper[4824]: I0121 11:12:27.124076 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:27Z","lastTransitionTime":"2026-01-21T11:12:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:27 crc kubenswrapper[4824]: I0121 11:12:27.226369 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:27 crc kubenswrapper[4824]: I0121 11:12:27.226404 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:27 crc kubenswrapper[4824]: I0121 11:12:27.226412 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:27 crc kubenswrapper[4824]: I0121 11:12:27.226424 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:27 crc kubenswrapper[4824]: I0121 11:12:27.226432 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:27Z","lastTransitionTime":"2026-01-21T11:12:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:27 crc kubenswrapper[4824]: I0121 11:12:27.328192 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:27 crc kubenswrapper[4824]: I0121 11:12:27.328228 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:27 crc kubenswrapper[4824]: I0121 11:12:27.328238 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:27 crc kubenswrapper[4824]: I0121 11:12:27.328251 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:27 crc kubenswrapper[4824]: I0121 11:12:27.328258 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:27Z","lastTransitionTime":"2026-01-21T11:12:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:27 crc kubenswrapper[4824]: I0121 11:12:27.429700 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:27 crc kubenswrapper[4824]: I0121 11:12:27.429735 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:27 crc kubenswrapper[4824]: I0121 11:12:27.429745 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:27 crc kubenswrapper[4824]: I0121 11:12:27.429758 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:27 crc kubenswrapper[4824]: I0121 11:12:27.429767 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:27Z","lastTransitionTime":"2026-01-21T11:12:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:27 crc kubenswrapper[4824]: I0121 11:12:27.531797 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:27 crc kubenswrapper[4824]: I0121 11:12:27.531830 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:27 crc kubenswrapper[4824]: I0121 11:12:27.531840 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:27 crc kubenswrapper[4824]: I0121 11:12:27.531851 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:27 crc kubenswrapper[4824]: I0121 11:12:27.531859 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:27Z","lastTransitionTime":"2026-01-21T11:12:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:27 crc kubenswrapper[4824]: I0121 11:12:27.633557 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:27 crc kubenswrapper[4824]: I0121 11:12:27.633589 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:27 crc kubenswrapper[4824]: I0121 11:12:27.633600 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:27 crc kubenswrapper[4824]: I0121 11:12:27.633613 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:27 crc kubenswrapper[4824]: I0121 11:12:27.633622 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:27Z","lastTransitionTime":"2026-01-21T11:12:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:27 crc kubenswrapper[4824]: I0121 11:12:27.735686 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:27 crc kubenswrapper[4824]: I0121 11:12:27.735726 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:27 crc kubenswrapper[4824]: I0121 11:12:27.735735 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:27 crc kubenswrapper[4824]: I0121 11:12:27.735751 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:27 crc kubenswrapper[4824]: I0121 11:12:27.735759 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:27Z","lastTransitionTime":"2026-01-21T11:12:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:27 crc kubenswrapper[4824]: I0121 11:12:27.837379 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:27 crc kubenswrapper[4824]: I0121 11:12:27.837411 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:27 crc kubenswrapper[4824]: I0121 11:12:27.837420 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:27 crc kubenswrapper[4824]: I0121 11:12:27.837432 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:27 crc kubenswrapper[4824]: I0121 11:12:27.837439 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:27Z","lastTransitionTime":"2026-01-21T11:12:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:27 crc kubenswrapper[4824]: I0121 11:12:27.938832 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:27 crc kubenswrapper[4824]: I0121 11:12:27.938854 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:27 crc kubenswrapper[4824]: I0121 11:12:27.938862 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:27 crc kubenswrapper[4824]: I0121 11:12:27.938871 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:27 crc kubenswrapper[4824]: I0121 11:12:27.938879 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:27Z","lastTransitionTime":"2026-01-21T11:12:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:28 crc kubenswrapper[4824]: I0121 11:12:28.033758 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-17 06:59:54.159032978 +0000 UTC Jan 21 11:12:28 crc kubenswrapper[4824]: I0121 11:12:28.040944 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:28 crc kubenswrapper[4824]: I0121 11:12:28.040983 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:28 crc kubenswrapper[4824]: I0121 11:12:28.040992 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:28 crc kubenswrapper[4824]: I0121 11:12:28.041001 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:28 crc kubenswrapper[4824]: I0121 11:12:28.041009 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:28Z","lastTransitionTime":"2026-01-21T11:12:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:28 crc kubenswrapper[4824]: I0121 11:12:28.048321 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9rmnw" Jan 21 11:12:28 crc kubenswrapper[4824]: E0121 11:12:28.048423 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9rmnw" podUID="8f9bd12b-22ee-44ed-a8b9-c100d043f691" Jan 21 11:12:28 crc kubenswrapper[4824]: I0121 11:12:28.048441 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 11:12:28 crc kubenswrapper[4824]: E0121 11:12:28.048499 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 21 11:12:28 crc kubenswrapper[4824]: I0121 11:12:28.062533 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06f56d86-ec6f-4b57-b314-f508c0dd4d9f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d25c94f1ae0bbabb733f0e1ea1d1ba39562a92af73047a39aacc03013e239681\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca06234486b59cf990c69d9501e23f935328e2ba21013cd71ed64831ea749bcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b63e45d66bcd11ff3479cc3ec6b450017da8d21d36e967b8d496ae0f24b0ee9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52758bcd31f2ba64edf03adfd4d1be6428015abff5b64c7c029fa787085f9c29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2736e73342d2adb175b0721dd4f40cc42d26fca06d5eb3334927518531c3b107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b35ad7dd0570c191eb0080739304f3cd43f17b3d1dddf895549441ebe50deaa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b35ad7dd0570c191eb0080739304f3cd43f17b3d1dddf895549441ebe50deaa6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://80cd5ea5e6fc19a88c1aab5c61f827f103ab798a5c84da8dc4d45cea86d5aacb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80cd5ea5e6fc19a88c1aab5c61f827f103ab798a5c84da8dc4d45cea86d5aacb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://234d8b914e6a88480021647ae65d74e9b21c55b6788312a094d00c8a42677d48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://234d8b914e6a88480021647ae65d74e9b21c55b6788312a094d00c8a42677d48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:28Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:28 crc kubenswrapper[4824]: I0121 11:12:28.071015 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:28Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:28 crc kubenswrapper[4824]: I0121 11:12:28.077971 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-9rmnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f9bd12b-22ee-44ed-a8b9-c100d043f691\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2fgn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2fgn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:28Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-9rmnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:28Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:28 crc kubenswrapper[4824]: I0121 11:12:28.085684 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86d3005cd61c0973659e1c1c0d3ef1ddd6fcf95bede5465f4f2099e64a74bb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:28Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:28 crc kubenswrapper[4824]: I0121 11:12:28.093032 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:28Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:28 crc kubenswrapper[4824]: I0121 11:12:28.103398 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x94nb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:12:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:12:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49ce055ec8090d6957cb7e23fe75edb82c0c2d7f87e7203d280606368dd3179f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75d52224135b75133fc000611cd179240b9174f6492d3263341716c7e42ccb15\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-21T11:12:02Z\\\",\\\"message\\\":\\\"2026-01-21T11:11:17+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_c9e2a0c0-e545-4ddc-b22f-97601d5900ca\\\\n2026-01-21T11:11:17+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_c9e2a0c0-e545-4ddc-b22f-97601d5900ca to /host/opt/cni/bin/\\\\n2026-01-21T11:11:17Z [verbose] multus-daemon started\\\\n2026-01-21T11:11:17Z [verbose] Readiness Indicator file check\\\\n2026-01-21T11:12:02Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:12:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgr8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x94nb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:28Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:28 crc kubenswrapper[4824]: I0121 11:12:28.110653 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33f3d922-4ffe-409b-a49a-d88c85898260\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a9e57b39eed8300107e944b9fcf6a2ff1e75991f8f37387a3de615b14f9bc78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpwst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b1b37c6d2994021cfc30f85e92c40ae35677113b17aa066d52bc7a76ae5c153\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpwst\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zqjg9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:28Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:28 crc kubenswrapper[4824]: I0121 11:12:28.119371 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gjtnh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a8d2cac-2d0f-4894-9bc0-2abd81699c47\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67c122a26811deaa9b70d19a1a65daee318ca3af2c626712482179d72fe67875\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2239186f1de46481cebfcafb1cab85cd320a205b5a0dfc6dd1b321d9baedeba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b2239186f1de46481cebfcafb1cab85cd320a205b5a0dfc6dd1b321d9baedeba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0db5ecd7759d974ae1b0d8065b3b8bd4d7957fae16eb5bab2d71c2fe69243816\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0db5ecd7759d974ae1b0d8065b3b8bd4d7957fae16eb5bab2d71c2fe69243816\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba9b3278c208fc7aac630ca7a0daa3de0823f85d0799d6f1f278da80ae14c862\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ba9b3278c208fc7aac630ca7a0daa3de0823f85d0799d6f1f278da80ae14c862\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79b8f6c016b43c3ac3ba7b22cca9a6b72c262cb17ddd6236c5d4970ecc09680c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79b8f6c016b43c3ac3ba7b22cca9a6b72c262cb17ddd6236c5d4970ecc09680c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38c4f218eb2ca7ea0c92bd0ae2ae9efeaa40746751608fbcf7f6c3f8d9e3c758\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38c4f218eb2ca7ea0c92bd0ae2ae9efeaa40746751608fbcf7f6c3f8d9e3c758\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d66ad9831d6029c8c350054525a00069e869d39c1b5638e06e9b772666bd8630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d66ad9831d6029c8c350054525a00069e869d39c1b5638e06e9b772666bd8630\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q78mk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gjtnh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:28Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:28 crc kubenswrapper[4824]: I0121 11:12:28.126084 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d87422473838f5218c9fb646ab2515f2f42592754ffdf07b19f7a8e093fa48a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:28Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:28 crc kubenswrapper[4824]: I0121 11:12:28.133804 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://516c009d4720ab5b393c7482872f12fed22d4f4414bd1f3400411c0e8da88ce4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee18d293036971832c8d9c3f7b9eca84ab779ff6768a2db3c6f7bdcf3433cdef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:28Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:28 crc kubenswrapper[4824]: I0121 11:12:28.142459 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:28 crc kubenswrapper[4824]: I0121 11:12:28.142498 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:28 crc kubenswrapper[4824]: I0121 11:12:28.142509 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:28 crc kubenswrapper[4824]: I0121 11:12:28.142523 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:28 crc kubenswrapper[4824]: I0121 11:12:28.142533 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:28Z","lastTransitionTime":"2026-01-21T11:12:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:28 crc kubenswrapper[4824]: I0121 11:12:28.145458 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"26b1108a-4b98-4e39-a3cd-e0d055089fd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc319bbc06568cc811ded2b418b04b986a31f321fed492e8be50104af79bb30b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cee9cc9eec3fc815f1621a22a37805691288dbb60dd3ff64194a84be62abcddc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc4c96e33e069f7d08d33caf92cd0987668506f931fd134b246d5ce5e934f95c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ddbb0e5ae02e9d5d82a80ead7efc91f1b8e557a9fb6f39940e5530fae8de2d38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://64a9065254f889b1c43d63759283335bf78fc4fa1531a4baae472c06c075dae3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e94ef8a8f47dc75ecfbab2e3fb8e7097c1b73eed32d7ca947f11ea198cea880\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86a6881f109dab1f910fe505a21c7c800e587de4dded2635147997c51c05c483\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://86a6881f109dab1f910fe505a21c7c800e587de4dded2635147997c51c05c483\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-21T11:12:07Z\\\",\\\"message\\\":\\\"UUIDName:}]\\\\nI0121 11:12:07.749346 6924 services_controller.go:453] Built service openshift-console/console template LB for network=default: []services.LB{}\\\\nI0121 11:12:07.749353 6924 services_controller.go:454] Service openshift-console/console for network=default has 1 cluster-wide, 0 per-node configs, 0 template configs, making 1 (cluster) 0 (per node) and 0 (template) load balancers\\\\nI0121 11:12:07.747279 6924 services_controller.go:434] Service openshift-kube-storage-version-migrator-operator/metrics retrieved from lister for network=default: \\\\u0026Service{ObjectMeta:{metrics openshift-kube-storage-version-migrator-operator e1639a86-fb7f-46de-9d5e-4aee16dccea1 4372 0 2025-02-23 05:12:25 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[app:kube-storage-version-migrator-operator] map[include.release.openshift.io/hypershift:true include.release.openshift.io/ibm-cloud-managed:true include.release.openshift.io/self-managed-high-availability:true service.alpha.openshift.io/serving-cert-secret-name:serving-cert service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc007778417 \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:htt\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T11:12:07Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-htkvk_openshift-ovn-kubernetes(26b1108a-4b98-4e39-a3cd-e0d055089fd8)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948e0497eaffa7b0d84b4085c9ed8d987eb2c6527c9a82c10e576ed9c9c1e21b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zcf8q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-htkvk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:28Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:28 crc kubenswrapper[4824]: I0121 11:12:28.152280 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-np2v8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4698abe4-0f33-4349-a2a4-614de80ae21b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://191328471bb6f2b34cde2eb927da697619826bf0b6174ecb583c8c25d27a6654\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n9qbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://050bd5bdded2162d5e9dac81713f4ed63cb575d1776e828b2f98e50d89f11b04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n9qbj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-np2v8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:28Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:28 crc kubenswrapper[4824]: I0121 11:12:28.158060 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lv4rn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1267539-9adf-466f-8027-8ede009d19a0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eed1b884aa53e1ee318169a0cc2c9ad10130aa363ae4ac55e4bf065237756f53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbmfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lv4rn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:28Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:28 crc kubenswrapper[4824]: I0121 11:12:28.163747 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jr64s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9768c60d-5bee-4286-bced-076e584b04e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91158f14d2e98ac9f875a5e7101df3107ef28aa5fc5d62218f460af98ff920fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flrnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:17Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jr64s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:28Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:28 crc kubenswrapper[4824]: I0121 11:12:28.169577 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b44f8a34-1578-4634-b62e-8e222844b397\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fd58f2438b6f9cbb2c37687e39f02cdf96a029477b1ae33f6dc9f3b70ecc926\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef10a172a3279201375f7445673820033978ce1eb2bbef7a2c0ed297fd783875\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef10a172a3279201375f7445673820033978ce1eb2bbef7a2c0ed297fd783875\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:28Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:28 crc kubenswrapper[4824]: I0121 11:12:28.178553 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d31ab61-f94c-4f49-85d0-c9e91b83069b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a66b5c33884af00408dc35cf02f08ebd8570bd695f9f1fe558bf54fd06930431\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7514786602746dabfa690da48d48e233fe68d6dd6674aa5ecd489c0b3185f08\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61b14b6846f94cdeba8f1f61b3754e7663041f92c70356c4ee6cbb988a2e680f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a7023cfe4c098bcdd9c0477a8187f4ed7a0d2abdbe1e143436863948f441b4f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af11ad95bcf0c6bbd5fa4ab99bb592ce1c87f40d29c5ade7832414315820a569\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW0121 11:11:14.604290 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0121 11:11:14.604572 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0121 11:11:14.605875 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3284452394/tls.crt::/tmp/serving-cert-3284452394/tls.key\\\\\\\"\\\\nI0121 11:11:14.923427 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0121 11:11:14.928141 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0121 11:11:14.928162 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0121 11:11:14.928179 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0121 11:11:14.928184 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0121 11:11:14.933154 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0121 11:11:14.933230 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 11:11:14.933259 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 11:11:14.933282 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0121 11:11:14.933299 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0121 11:11:14.933315 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0121 11:11:14.933331 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0121 11:11:14.933177 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0121 11:11:14.934666 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48070c09fad511108ff22e44211cf26911509b8406d66948985cbfc3009d8233\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:28Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:28 crc kubenswrapper[4824]: I0121 11:12:28.186212 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9bb99c0a-683b-44da-9c0e-e8d284e87d53\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05e076febd697ed802ce5c7ee92a8c2c636a1a2c6609ad7eb01f17dace247c4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43529d1b9e1e5e22b0d56b023946119da8aacbae54b68c26ff0a1a82dc890c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ea232ad87393685aa323ce6393ff3869f8f8d423c2406a501c69c355535adfa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc6c7eff9e17b8a8ecc88bbfe703b5a01dbf1e02d342cede4f6a263a6526af43\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:28Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:28 crc kubenswrapper[4824]: I0121 11:12:28.193224 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e570fff1-6227-44d6-b71f-ee377b1099ae\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4d41492c3802a8168fffd249c15aeb9c902a5c7e01550e1aa6f9b769adfdc4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2690a1ecbc8fe1290e68bba2695a3edd833b5da933fb71e8ddc926a1340a9a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b85db961e1e078e1f54323c79a99c9413365256ac2fce411db7932bd0580ab6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://595783a1454735bd33f4862180b60d99a4e1b6fae1e735f2b7328391951bf422\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://595783a1454735bd33f4862180b60d99a4e1b6fae1e735f2b7328391951bf422\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:28Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:28 crc kubenswrapper[4824]: I0121 11:12:28.200574 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:28Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:28 crc kubenswrapper[4824]: I0121 11:12:28.244812 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:28 crc kubenswrapper[4824]: I0121 11:12:28.244841 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:28 crc kubenswrapper[4824]: I0121 11:12:28.244850 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:28 crc kubenswrapper[4824]: I0121 11:12:28.244862 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:28 crc kubenswrapper[4824]: I0121 11:12:28.244870 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:28Z","lastTransitionTime":"2026-01-21T11:12:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:28 crc kubenswrapper[4824]: I0121 11:12:28.346013 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:28 crc kubenswrapper[4824]: I0121 11:12:28.346041 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:28 crc kubenswrapper[4824]: I0121 11:12:28.346051 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:28 crc kubenswrapper[4824]: I0121 11:12:28.346062 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:28 crc kubenswrapper[4824]: I0121 11:12:28.346072 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:28Z","lastTransitionTime":"2026-01-21T11:12:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:28 crc kubenswrapper[4824]: I0121 11:12:28.447612 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:28 crc kubenswrapper[4824]: I0121 11:12:28.447643 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:28 crc kubenswrapper[4824]: I0121 11:12:28.447652 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:28 crc kubenswrapper[4824]: I0121 11:12:28.447664 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:28 crc kubenswrapper[4824]: I0121 11:12:28.447674 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:28Z","lastTransitionTime":"2026-01-21T11:12:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:28 crc kubenswrapper[4824]: I0121 11:12:28.549748 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:28 crc kubenswrapper[4824]: I0121 11:12:28.549777 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:28 crc kubenswrapper[4824]: I0121 11:12:28.549785 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:28 crc kubenswrapper[4824]: I0121 11:12:28.549796 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:28 crc kubenswrapper[4824]: I0121 11:12:28.549805 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:28Z","lastTransitionTime":"2026-01-21T11:12:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:28 crc kubenswrapper[4824]: I0121 11:12:28.651450 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:28 crc kubenswrapper[4824]: I0121 11:12:28.651479 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:28 crc kubenswrapper[4824]: I0121 11:12:28.651487 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:28 crc kubenswrapper[4824]: I0121 11:12:28.651498 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:28 crc kubenswrapper[4824]: I0121 11:12:28.651507 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:28Z","lastTransitionTime":"2026-01-21T11:12:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:28 crc kubenswrapper[4824]: I0121 11:12:28.753212 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:28 crc kubenswrapper[4824]: I0121 11:12:28.753236 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:28 crc kubenswrapper[4824]: I0121 11:12:28.753244 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:28 crc kubenswrapper[4824]: I0121 11:12:28.753317 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:28 crc kubenswrapper[4824]: I0121 11:12:28.753329 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:28Z","lastTransitionTime":"2026-01-21T11:12:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:28 crc kubenswrapper[4824]: I0121 11:12:28.854949 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:28 crc kubenswrapper[4824]: I0121 11:12:28.854997 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:28 crc kubenswrapper[4824]: I0121 11:12:28.855005 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:28 crc kubenswrapper[4824]: I0121 11:12:28.855015 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:28 crc kubenswrapper[4824]: I0121 11:12:28.855023 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:28Z","lastTransitionTime":"2026-01-21T11:12:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:28 crc kubenswrapper[4824]: I0121 11:12:28.956544 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:28 crc kubenswrapper[4824]: I0121 11:12:28.956573 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:28 crc kubenswrapper[4824]: I0121 11:12:28.956584 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:28 crc kubenswrapper[4824]: I0121 11:12:28.956593 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:28 crc kubenswrapper[4824]: I0121 11:12:28.956601 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:28Z","lastTransitionTime":"2026-01-21T11:12:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:29 crc kubenswrapper[4824]: I0121 11:12:29.034786 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-16 20:56:25.353670252 +0000 UTC Jan 21 11:12:29 crc kubenswrapper[4824]: I0121 11:12:29.049185 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 11:12:29 crc kubenswrapper[4824]: I0121 11:12:29.049214 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 11:12:29 crc kubenswrapper[4824]: E0121 11:12:29.049274 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 21 11:12:29 crc kubenswrapper[4824]: E0121 11:12:29.049330 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 21 11:12:29 crc kubenswrapper[4824]: I0121 11:12:29.058075 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:29 crc kubenswrapper[4824]: I0121 11:12:29.058095 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:29 crc kubenswrapper[4824]: I0121 11:12:29.058103 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:29 crc kubenswrapper[4824]: I0121 11:12:29.058112 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:29 crc kubenswrapper[4824]: I0121 11:12:29.058119 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:29Z","lastTransitionTime":"2026-01-21T11:12:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:29 crc kubenswrapper[4824]: I0121 11:12:29.159858 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:29 crc kubenswrapper[4824]: I0121 11:12:29.159883 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:29 crc kubenswrapper[4824]: I0121 11:12:29.159891 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:29 crc kubenswrapper[4824]: I0121 11:12:29.159900 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:29 crc kubenswrapper[4824]: I0121 11:12:29.159908 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:29Z","lastTransitionTime":"2026-01-21T11:12:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:29 crc kubenswrapper[4824]: I0121 11:12:29.226107 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:29 crc kubenswrapper[4824]: I0121 11:12:29.226136 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:29 crc kubenswrapper[4824]: I0121 11:12:29.226146 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:29 crc kubenswrapper[4824]: I0121 11:12:29.226157 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:29 crc kubenswrapper[4824]: I0121 11:12:29.226166 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:29Z","lastTransitionTime":"2026-01-21T11:12:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:29 crc kubenswrapper[4824]: E0121 11:12:29.235163 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:12:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:12:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:12:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:12:29Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:12:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:12:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:12:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:12:29Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"695c676c-720c-4921-9e71-56b96f2fb46b\\\",\\\"systemUUID\\\":\\\"d00624f5-786f-41dc-bdc1-a1875f167d65\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:29Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:29 crc kubenswrapper[4824]: I0121 11:12:29.237843 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:29 crc kubenswrapper[4824]: I0121 11:12:29.237865 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:29 crc kubenswrapper[4824]: I0121 11:12:29.237873 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:29 crc kubenswrapper[4824]: I0121 11:12:29.237884 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:29 crc kubenswrapper[4824]: I0121 11:12:29.237891 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:29Z","lastTransitionTime":"2026-01-21T11:12:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:29 crc kubenswrapper[4824]: E0121 11:12:29.245375 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:12:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:12:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:12:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:12:29Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:12:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:12:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:12:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:12:29Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"695c676c-720c-4921-9e71-56b96f2fb46b\\\",\\\"systemUUID\\\":\\\"d00624f5-786f-41dc-bdc1-a1875f167d65\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:29Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:29 crc kubenswrapper[4824]: I0121 11:12:29.247426 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:29 crc kubenswrapper[4824]: I0121 11:12:29.247447 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:29 crc kubenswrapper[4824]: I0121 11:12:29.247456 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:29 crc kubenswrapper[4824]: I0121 11:12:29.247464 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:29 crc kubenswrapper[4824]: I0121 11:12:29.247472 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:29Z","lastTransitionTime":"2026-01-21T11:12:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:29 crc kubenswrapper[4824]: E0121 11:12:29.254715 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:12:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:12:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:12:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:12:29Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:12:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:12:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:12:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:12:29Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"695c676c-720c-4921-9e71-56b96f2fb46b\\\",\\\"systemUUID\\\":\\\"d00624f5-786f-41dc-bdc1-a1875f167d65\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:29Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:29 crc kubenswrapper[4824]: I0121 11:12:29.257011 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:29 crc kubenswrapper[4824]: I0121 11:12:29.257035 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:29 crc kubenswrapper[4824]: I0121 11:12:29.257043 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:29 crc kubenswrapper[4824]: I0121 11:12:29.257051 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:29 crc kubenswrapper[4824]: I0121 11:12:29.257058 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:29Z","lastTransitionTime":"2026-01-21T11:12:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:29 crc kubenswrapper[4824]: E0121 11:12:29.264390 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:12:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:12:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:12:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:12:29Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:12:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:12:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:12:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:12:29Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"695c676c-720c-4921-9e71-56b96f2fb46b\\\",\\\"systemUUID\\\":\\\"d00624f5-786f-41dc-bdc1-a1875f167d65\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:29Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:29 crc kubenswrapper[4824]: I0121 11:12:29.266303 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:29 crc kubenswrapper[4824]: I0121 11:12:29.266325 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:29 crc kubenswrapper[4824]: I0121 11:12:29.266333 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:29 crc kubenswrapper[4824]: I0121 11:12:29.266341 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:29 crc kubenswrapper[4824]: I0121 11:12:29.266359 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:29Z","lastTransitionTime":"2026-01-21T11:12:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:29 crc kubenswrapper[4824]: E0121 11:12:29.273473 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:12:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:12:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:12:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:12:29Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:12:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:12:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:12:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-21T11:12:29Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"695c676c-720c-4921-9e71-56b96f2fb46b\\\",\\\"systemUUID\\\":\\\"d00624f5-786f-41dc-bdc1-a1875f167d65\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:29Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:29 crc kubenswrapper[4824]: E0121 11:12:29.273592 4824 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Jan 21 11:12:29 crc kubenswrapper[4824]: I0121 11:12:29.274459 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:29 crc kubenswrapper[4824]: I0121 11:12:29.274484 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:29 crc kubenswrapper[4824]: I0121 11:12:29.274493 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:29 crc kubenswrapper[4824]: I0121 11:12:29.274503 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:29 crc kubenswrapper[4824]: I0121 11:12:29.274511 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:29Z","lastTransitionTime":"2026-01-21T11:12:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:29 crc kubenswrapper[4824]: I0121 11:12:29.376411 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:29 crc kubenswrapper[4824]: I0121 11:12:29.376433 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:29 crc kubenswrapper[4824]: I0121 11:12:29.376440 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:29 crc kubenswrapper[4824]: I0121 11:12:29.376450 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:29 crc kubenswrapper[4824]: I0121 11:12:29.376457 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:29Z","lastTransitionTime":"2026-01-21T11:12:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:29 crc kubenswrapper[4824]: I0121 11:12:29.478225 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:29 crc kubenswrapper[4824]: I0121 11:12:29.478248 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:29 crc kubenswrapper[4824]: I0121 11:12:29.478256 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:29 crc kubenswrapper[4824]: I0121 11:12:29.478266 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:29 crc kubenswrapper[4824]: I0121 11:12:29.478274 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:29Z","lastTransitionTime":"2026-01-21T11:12:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:29 crc kubenswrapper[4824]: I0121 11:12:29.580378 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:29 crc kubenswrapper[4824]: I0121 11:12:29.580405 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:29 crc kubenswrapper[4824]: I0121 11:12:29.580413 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:29 crc kubenswrapper[4824]: I0121 11:12:29.580424 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:29 crc kubenswrapper[4824]: I0121 11:12:29.580431 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:29Z","lastTransitionTime":"2026-01-21T11:12:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:29 crc kubenswrapper[4824]: I0121 11:12:29.682070 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:29 crc kubenswrapper[4824]: I0121 11:12:29.682098 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:29 crc kubenswrapper[4824]: I0121 11:12:29.682108 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:29 crc kubenswrapper[4824]: I0121 11:12:29.682116 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:29 crc kubenswrapper[4824]: I0121 11:12:29.682124 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:29Z","lastTransitionTime":"2026-01-21T11:12:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:29 crc kubenswrapper[4824]: I0121 11:12:29.783745 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:29 crc kubenswrapper[4824]: I0121 11:12:29.783765 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:29 crc kubenswrapper[4824]: I0121 11:12:29.783774 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:29 crc kubenswrapper[4824]: I0121 11:12:29.783782 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:29 crc kubenswrapper[4824]: I0121 11:12:29.783790 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:29Z","lastTransitionTime":"2026-01-21T11:12:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:29 crc kubenswrapper[4824]: I0121 11:12:29.884809 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:29 crc kubenswrapper[4824]: I0121 11:12:29.884834 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:29 crc kubenswrapper[4824]: I0121 11:12:29.884841 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:29 crc kubenswrapper[4824]: I0121 11:12:29.884851 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:29 crc kubenswrapper[4824]: I0121 11:12:29.884858 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:29Z","lastTransitionTime":"2026-01-21T11:12:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:29 crc kubenswrapper[4824]: I0121 11:12:29.986902 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:29 crc kubenswrapper[4824]: I0121 11:12:29.986927 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:29 crc kubenswrapper[4824]: I0121 11:12:29.986937 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:29 crc kubenswrapper[4824]: I0121 11:12:29.986948 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:29 crc kubenswrapper[4824]: I0121 11:12:29.986972 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:29Z","lastTransitionTime":"2026-01-21T11:12:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:30 crc kubenswrapper[4824]: I0121 11:12:30.035409 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-23 20:29:34.881837332 +0000 UTC Jan 21 11:12:30 crc kubenswrapper[4824]: I0121 11:12:30.049089 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 11:12:30 crc kubenswrapper[4824]: I0121 11:12:30.049139 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9rmnw" Jan 21 11:12:30 crc kubenswrapper[4824]: E0121 11:12:30.049187 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 21 11:12:30 crc kubenswrapper[4824]: E0121 11:12:30.049248 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9rmnw" podUID="8f9bd12b-22ee-44ed-a8b9-c100d043f691" Jan 21 11:12:30 crc kubenswrapper[4824]: I0121 11:12:30.088159 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:30 crc kubenswrapper[4824]: I0121 11:12:30.088179 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:30 crc kubenswrapper[4824]: I0121 11:12:30.088187 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:30 crc kubenswrapper[4824]: I0121 11:12:30.088196 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:30 crc kubenswrapper[4824]: I0121 11:12:30.088203 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:30Z","lastTransitionTime":"2026-01-21T11:12:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:30 crc kubenswrapper[4824]: I0121 11:12:30.190534 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:30 crc kubenswrapper[4824]: I0121 11:12:30.190565 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:30 crc kubenswrapper[4824]: I0121 11:12:30.190572 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:30 crc kubenswrapper[4824]: I0121 11:12:30.190580 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:30 crc kubenswrapper[4824]: I0121 11:12:30.190587 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:30Z","lastTransitionTime":"2026-01-21T11:12:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:30 crc kubenswrapper[4824]: I0121 11:12:30.292606 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:30 crc kubenswrapper[4824]: I0121 11:12:30.292625 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:30 crc kubenswrapper[4824]: I0121 11:12:30.292634 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:30 crc kubenswrapper[4824]: I0121 11:12:30.292643 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:30 crc kubenswrapper[4824]: I0121 11:12:30.292650 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:30Z","lastTransitionTime":"2026-01-21T11:12:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:30 crc kubenswrapper[4824]: I0121 11:12:30.393841 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:30 crc kubenswrapper[4824]: I0121 11:12:30.393906 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:30 crc kubenswrapper[4824]: I0121 11:12:30.393916 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:30 crc kubenswrapper[4824]: I0121 11:12:30.393929 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:30 crc kubenswrapper[4824]: I0121 11:12:30.393937 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:30Z","lastTransitionTime":"2026-01-21T11:12:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:30 crc kubenswrapper[4824]: I0121 11:12:30.496085 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:30 crc kubenswrapper[4824]: I0121 11:12:30.496120 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:30 crc kubenswrapper[4824]: I0121 11:12:30.496130 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:30 crc kubenswrapper[4824]: I0121 11:12:30.496142 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:30 crc kubenswrapper[4824]: I0121 11:12:30.496151 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:30Z","lastTransitionTime":"2026-01-21T11:12:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:30 crc kubenswrapper[4824]: I0121 11:12:30.597578 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:30 crc kubenswrapper[4824]: I0121 11:12:30.597604 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:30 crc kubenswrapper[4824]: I0121 11:12:30.597612 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:30 crc kubenswrapper[4824]: I0121 11:12:30.597622 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:30 crc kubenswrapper[4824]: I0121 11:12:30.597629 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:30Z","lastTransitionTime":"2026-01-21T11:12:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:30 crc kubenswrapper[4824]: I0121 11:12:30.699176 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:30 crc kubenswrapper[4824]: I0121 11:12:30.699202 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:30 crc kubenswrapper[4824]: I0121 11:12:30.699211 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:30 crc kubenswrapper[4824]: I0121 11:12:30.699220 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:30 crc kubenswrapper[4824]: I0121 11:12:30.699228 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:30Z","lastTransitionTime":"2026-01-21T11:12:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:30 crc kubenswrapper[4824]: I0121 11:12:30.800501 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:30 crc kubenswrapper[4824]: I0121 11:12:30.800534 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:30 crc kubenswrapper[4824]: I0121 11:12:30.800542 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:30 crc kubenswrapper[4824]: I0121 11:12:30.800554 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:30 crc kubenswrapper[4824]: I0121 11:12:30.800563 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:30Z","lastTransitionTime":"2026-01-21T11:12:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:30 crc kubenswrapper[4824]: I0121 11:12:30.903916 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:30 crc kubenswrapper[4824]: I0121 11:12:30.904082 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:30 crc kubenswrapper[4824]: I0121 11:12:30.904147 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:30 crc kubenswrapper[4824]: I0121 11:12:30.904175 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:30 crc kubenswrapper[4824]: I0121 11:12:30.904184 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:30Z","lastTransitionTime":"2026-01-21T11:12:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:31 crc kubenswrapper[4824]: I0121 11:12:31.006231 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:31 crc kubenswrapper[4824]: I0121 11:12:31.006267 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:31 crc kubenswrapper[4824]: I0121 11:12:31.006278 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:31 crc kubenswrapper[4824]: I0121 11:12:31.006291 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:31 crc kubenswrapper[4824]: I0121 11:12:31.006299 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:31Z","lastTransitionTime":"2026-01-21T11:12:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:31 crc kubenswrapper[4824]: I0121 11:12:31.035513 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-22 03:46:03.149802342 +0000 UTC Jan 21 11:12:31 crc kubenswrapper[4824]: I0121 11:12:31.048839 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 11:12:31 crc kubenswrapper[4824]: E0121 11:12:31.048925 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 21 11:12:31 crc kubenswrapper[4824]: I0121 11:12:31.048844 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 11:12:31 crc kubenswrapper[4824]: E0121 11:12:31.049036 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 21 11:12:31 crc kubenswrapper[4824]: I0121 11:12:31.108311 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:31 crc kubenswrapper[4824]: I0121 11:12:31.108357 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:31 crc kubenswrapper[4824]: I0121 11:12:31.108365 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:31 crc kubenswrapper[4824]: I0121 11:12:31.108377 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:31 crc kubenswrapper[4824]: I0121 11:12:31.108386 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:31Z","lastTransitionTime":"2026-01-21T11:12:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:31 crc kubenswrapper[4824]: I0121 11:12:31.210253 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:31 crc kubenswrapper[4824]: I0121 11:12:31.210284 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:31 crc kubenswrapper[4824]: I0121 11:12:31.210292 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:31 crc kubenswrapper[4824]: I0121 11:12:31.210305 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:31 crc kubenswrapper[4824]: I0121 11:12:31.210314 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:31Z","lastTransitionTime":"2026-01-21T11:12:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:31 crc kubenswrapper[4824]: I0121 11:12:31.312442 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:31 crc kubenswrapper[4824]: I0121 11:12:31.312474 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:31 crc kubenswrapper[4824]: I0121 11:12:31.312482 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:31 crc kubenswrapper[4824]: I0121 11:12:31.312493 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:31 crc kubenswrapper[4824]: I0121 11:12:31.312501 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:31Z","lastTransitionTime":"2026-01-21T11:12:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:31 crc kubenswrapper[4824]: I0121 11:12:31.414533 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:31 crc kubenswrapper[4824]: I0121 11:12:31.414564 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:31 crc kubenswrapper[4824]: I0121 11:12:31.414583 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:31 crc kubenswrapper[4824]: I0121 11:12:31.414595 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:31 crc kubenswrapper[4824]: I0121 11:12:31.414604 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:31Z","lastTransitionTime":"2026-01-21T11:12:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:31 crc kubenswrapper[4824]: I0121 11:12:31.516311 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:31 crc kubenswrapper[4824]: I0121 11:12:31.516410 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:31 crc kubenswrapper[4824]: I0121 11:12:31.516419 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:31 crc kubenswrapper[4824]: I0121 11:12:31.516428 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:31 crc kubenswrapper[4824]: I0121 11:12:31.516434 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:31Z","lastTransitionTime":"2026-01-21T11:12:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:31 crc kubenswrapper[4824]: I0121 11:12:31.617788 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:31 crc kubenswrapper[4824]: I0121 11:12:31.617813 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:31 crc kubenswrapper[4824]: I0121 11:12:31.617821 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:31 crc kubenswrapper[4824]: I0121 11:12:31.617829 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:31 crc kubenswrapper[4824]: I0121 11:12:31.617836 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:31Z","lastTransitionTime":"2026-01-21T11:12:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:31 crc kubenswrapper[4824]: I0121 11:12:31.719505 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:31 crc kubenswrapper[4824]: I0121 11:12:31.719532 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:31 crc kubenswrapper[4824]: I0121 11:12:31.719540 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:31 crc kubenswrapper[4824]: I0121 11:12:31.719549 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:31 crc kubenswrapper[4824]: I0121 11:12:31.719556 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:31Z","lastTransitionTime":"2026-01-21T11:12:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:31 crc kubenswrapper[4824]: I0121 11:12:31.821650 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:31 crc kubenswrapper[4824]: I0121 11:12:31.821669 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:31 crc kubenswrapper[4824]: I0121 11:12:31.821677 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:31 crc kubenswrapper[4824]: I0121 11:12:31.821686 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:31 crc kubenswrapper[4824]: I0121 11:12:31.821693 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:31Z","lastTransitionTime":"2026-01-21T11:12:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:31 crc kubenswrapper[4824]: I0121 11:12:31.923202 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:31 crc kubenswrapper[4824]: I0121 11:12:31.923222 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:31 crc kubenswrapper[4824]: I0121 11:12:31.923230 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:31 crc kubenswrapper[4824]: I0121 11:12:31.923238 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:31 crc kubenswrapper[4824]: I0121 11:12:31.923245 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:31Z","lastTransitionTime":"2026-01-21T11:12:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:32 crc kubenswrapper[4824]: I0121 11:12:32.024919 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:32 crc kubenswrapper[4824]: I0121 11:12:32.024935 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:32 crc kubenswrapper[4824]: I0121 11:12:32.024943 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:32 crc kubenswrapper[4824]: I0121 11:12:32.024974 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:32 crc kubenswrapper[4824]: I0121 11:12:32.024983 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:32Z","lastTransitionTime":"2026-01-21T11:12:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:32 crc kubenswrapper[4824]: I0121 11:12:32.036392 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-24 08:22:51.64614733 +0000 UTC Jan 21 11:12:32 crc kubenswrapper[4824]: I0121 11:12:32.048805 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9rmnw" Jan 21 11:12:32 crc kubenswrapper[4824]: I0121 11:12:32.049006 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 11:12:32 crc kubenswrapper[4824]: E0121 11:12:32.049131 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9rmnw" podUID="8f9bd12b-22ee-44ed-a8b9-c100d043f691" Jan 21 11:12:32 crc kubenswrapper[4824]: I0121 11:12:32.049230 4824 scope.go:117] "RemoveContainer" containerID="86a6881f109dab1f910fe505a21c7c800e587de4dded2635147997c51c05c483" Jan 21 11:12:32 crc kubenswrapper[4824]: E0121 11:12:32.049252 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 21 11:12:32 crc kubenswrapper[4824]: E0121 11:12:32.049385 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-htkvk_openshift-ovn-kubernetes(26b1108a-4b98-4e39-a3cd-e0d055089fd8)\"" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" podUID="26b1108a-4b98-4e39-a3cd-e0d055089fd8" Jan 21 11:12:32 crc kubenswrapper[4824]: I0121 11:12:32.126818 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:32 crc kubenswrapper[4824]: I0121 11:12:32.126842 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:32 crc kubenswrapper[4824]: I0121 11:12:32.126849 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:32 crc kubenswrapper[4824]: I0121 11:12:32.126860 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:32 crc kubenswrapper[4824]: I0121 11:12:32.126866 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:32Z","lastTransitionTime":"2026-01-21T11:12:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:32 crc kubenswrapper[4824]: I0121 11:12:32.179586 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8f9bd12b-22ee-44ed-a8b9-c100d043f691-metrics-certs\") pod \"network-metrics-daemon-9rmnw\" (UID: \"8f9bd12b-22ee-44ed-a8b9-c100d043f691\") " pod="openshift-multus/network-metrics-daemon-9rmnw" Jan 21 11:12:32 crc kubenswrapper[4824]: E0121 11:12:32.179665 4824 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Jan 21 11:12:32 crc kubenswrapper[4824]: E0121 11:12:32.179727 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8f9bd12b-22ee-44ed-a8b9-c100d043f691-metrics-certs podName:8f9bd12b-22ee-44ed-a8b9-c100d043f691 nodeName:}" failed. No retries permitted until 2026-01-21 11:13:36.179713996 +0000 UTC m=+158.472743288 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/8f9bd12b-22ee-44ed-a8b9-c100d043f691-metrics-certs") pod "network-metrics-daemon-9rmnw" (UID: "8f9bd12b-22ee-44ed-a8b9-c100d043f691") : object "openshift-multus"/"metrics-daemon-secret" not registered Jan 21 11:12:32 crc kubenswrapper[4824]: I0121 11:12:32.229469 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:32 crc kubenswrapper[4824]: I0121 11:12:32.229507 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:32 crc kubenswrapper[4824]: I0121 11:12:32.229515 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:32 crc kubenswrapper[4824]: I0121 11:12:32.229697 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:32 crc kubenswrapper[4824]: I0121 11:12:32.229727 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:32Z","lastTransitionTime":"2026-01-21T11:12:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:32 crc kubenswrapper[4824]: I0121 11:12:32.332992 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:32 crc kubenswrapper[4824]: I0121 11:12:32.333035 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:32 crc kubenswrapper[4824]: I0121 11:12:32.333044 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:32 crc kubenswrapper[4824]: I0121 11:12:32.333060 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:32 crc kubenswrapper[4824]: I0121 11:12:32.333070 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:32Z","lastTransitionTime":"2026-01-21T11:12:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:32 crc kubenswrapper[4824]: I0121 11:12:32.435855 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:32 crc kubenswrapper[4824]: I0121 11:12:32.435893 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:32 crc kubenswrapper[4824]: I0121 11:12:32.435922 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:32 crc kubenswrapper[4824]: I0121 11:12:32.435943 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:32 crc kubenswrapper[4824]: I0121 11:12:32.435971 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:32Z","lastTransitionTime":"2026-01-21T11:12:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:32 crc kubenswrapper[4824]: I0121 11:12:32.538417 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:32 crc kubenswrapper[4824]: I0121 11:12:32.538479 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:32 crc kubenswrapper[4824]: I0121 11:12:32.538489 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:32 crc kubenswrapper[4824]: I0121 11:12:32.538509 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:32 crc kubenswrapper[4824]: I0121 11:12:32.538519 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:32Z","lastTransitionTime":"2026-01-21T11:12:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:32 crc kubenswrapper[4824]: I0121 11:12:32.641548 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:32 crc kubenswrapper[4824]: I0121 11:12:32.641594 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:32 crc kubenswrapper[4824]: I0121 11:12:32.641604 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:32 crc kubenswrapper[4824]: I0121 11:12:32.641620 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:32 crc kubenswrapper[4824]: I0121 11:12:32.641627 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:32Z","lastTransitionTime":"2026-01-21T11:12:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:32 crc kubenswrapper[4824]: I0121 11:12:32.744218 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:32 crc kubenswrapper[4824]: I0121 11:12:32.744255 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:32 crc kubenswrapper[4824]: I0121 11:12:32.744264 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:32 crc kubenswrapper[4824]: I0121 11:12:32.744279 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:32 crc kubenswrapper[4824]: I0121 11:12:32.744287 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:32Z","lastTransitionTime":"2026-01-21T11:12:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:32 crc kubenswrapper[4824]: I0121 11:12:32.846415 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:32 crc kubenswrapper[4824]: I0121 11:12:32.846483 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:32 crc kubenswrapper[4824]: I0121 11:12:32.846498 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:32 crc kubenswrapper[4824]: I0121 11:12:32.846520 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:32 crc kubenswrapper[4824]: I0121 11:12:32.846531 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:32Z","lastTransitionTime":"2026-01-21T11:12:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:32 crc kubenswrapper[4824]: I0121 11:12:32.948997 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:32 crc kubenswrapper[4824]: I0121 11:12:32.949038 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:32 crc kubenswrapper[4824]: I0121 11:12:32.949047 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:32 crc kubenswrapper[4824]: I0121 11:12:32.949061 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:32 crc kubenswrapper[4824]: I0121 11:12:32.949070 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:32Z","lastTransitionTime":"2026-01-21T11:12:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:33 crc kubenswrapper[4824]: I0121 11:12:33.037174 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-12 12:45:29.05598313 +0000 UTC Jan 21 11:12:33 crc kubenswrapper[4824]: I0121 11:12:33.048518 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 11:12:33 crc kubenswrapper[4824]: I0121 11:12:33.048549 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 11:12:33 crc kubenswrapper[4824]: E0121 11:12:33.048677 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 21 11:12:33 crc kubenswrapper[4824]: E0121 11:12:33.048760 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 21 11:12:33 crc kubenswrapper[4824]: I0121 11:12:33.050692 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:33 crc kubenswrapper[4824]: I0121 11:12:33.050733 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:33 crc kubenswrapper[4824]: I0121 11:12:33.050741 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:33 crc kubenswrapper[4824]: I0121 11:12:33.050759 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:33 crc kubenswrapper[4824]: I0121 11:12:33.050768 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:33Z","lastTransitionTime":"2026-01-21T11:12:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:33 crc kubenswrapper[4824]: I0121 11:12:33.152648 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:33 crc kubenswrapper[4824]: I0121 11:12:33.152682 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:33 crc kubenswrapper[4824]: I0121 11:12:33.152690 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:33 crc kubenswrapper[4824]: I0121 11:12:33.152707 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:33 crc kubenswrapper[4824]: I0121 11:12:33.152715 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:33Z","lastTransitionTime":"2026-01-21T11:12:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:33 crc kubenswrapper[4824]: I0121 11:12:33.254796 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:33 crc kubenswrapper[4824]: I0121 11:12:33.254832 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:33 crc kubenswrapper[4824]: I0121 11:12:33.254841 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:33 crc kubenswrapper[4824]: I0121 11:12:33.254854 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:33 crc kubenswrapper[4824]: I0121 11:12:33.254863 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:33Z","lastTransitionTime":"2026-01-21T11:12:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:33 crc kubenswrapper[4824]: I0121 11:12:33.356821 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:33 crc kubenswrapper[4824]: I0121 11:12:33.356860 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:33 crc kubenswrapper[4824]: I0121 11:12:33.356869 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:33 crc kubenswrapper[4824]: I0121 11:12:33.356884 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:33 crc kubenswrapper[4824]: I0121 11:12:33.356894 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:33Z","lastTransitionTime":"2026-01-21T11:12:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:33 crc kubenswrapper[4824]: I0121 11:12:33.458384 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:33 crc kubenswrapper[4824]: I0121 11:12:33.458420 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:33 crc kubenswrapper[4824]: I0121 11:12:33.458429 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:33 crc kubenswrapper[4824]: I0121 11:12:33.458445 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:33 crc kubenswrapper[4824]: I0121 11:12:33.458454 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:33Z","lastTransitionTime":"2026-01-21T11:12:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:33 crc kubenswrapper[4824]: I0121 11:12:33.560308 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:33 crc kubenswrapper[4824]: I0121 11:12:33.560362 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:33 crc kubenswrapper[4824]: I0121 11:12:33.560373 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:33 crc kubenswrapper[4824]: I0121 11:12:33.560394 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:33 crc kubenswrapper[4824]: I0121 11:12:33.560405 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:33Z","lastTransitionTime":"2026-01-21T11:12:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:33 crc kubenswrapper[4824]: I0121 11:12:33.662420 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:33 crc kubenswrapper[4824]: I0121 11:12:33.662479 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:33 crc kubenswrapper[4824]: I0121 11:12:33.662487 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:33 crc kubenswrapper[4824]: I0121 11:12:33.662522 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:33 crc kubenswrapper[4824]: I0121 11:12:33.662531 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:33Z","lastTransitionTime":"2026-01-21T11:12:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:33 crc kubenswrapper[4824]: I0121 11:12:33.763870 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:33 crc kubenswrapper[4824]: I0121 11:12:33.763904 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:33 crc kubenswrapper[4824]: I0121 11:12:33.763912 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:33 crc kubenswrapper[4824]: I0121 11:12:33.763926 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:33 crc kubenswrapper[4824]: I0121 11:12:33.763935 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:33Z","lastTransitionTime":"2026-01-21T11:12:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:33 crc kubenswrapper[4824]: I0121 11:12:33.865926 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:33 crc kubenswrapper[4824]: I0121 11:12:33.866116 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:33 crc kubenswrapper[4824]: I0121 11:12:33.866127 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:33 crc kubenswrapper[4824]: I0121 11:12:33.866143 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:33 crc kubenswrapper[4824]: I0121 11:12:33.866153 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:33Z","lastTransitionTime":"2026-01-21T11:12:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:33 crc kubenswrapper[4824]: I0121 11:12:33.968650 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:33 crc kubenswrapper[4824]: I0121 11:12:33.968703 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:33 crc kubenswrapper[4824]: I0121 11:12:33.968713 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:33 crc kubenswrapper[4824]: I0121 11:12:33.968730 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:33 crc kubenswrapper[4824]: I0121 11:12:33.968738 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:33Z","lastTransitionTime":"2026-01-21T11:12:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:34 crc kubenswrapper[4824]: I0121 11:12:34.037584 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-18 05:58:41.730303231 +0000 UTC Jan 21 11:12:34 crc kubenswrapper[4824]: I0121 11:12:34.048953 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9rmnw" Jan 21 11:12:34 crc kubenswrapper[4824]: I0121 11:12:34.049053 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 11:12:34 crc kubenswrapper[4824]: E0121 11:12:34.049098 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9rmnw" podUID="8f9bd12b-22ee-44ed-a8b9-c100d043f691" Jan 21 11:12:34 crc kubenswrapper[4824]: E0121 11:12:34.049193 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 21 11:12:34 crc kubenswrapper[4824]: I0121 11:12:34.070991 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:34 crc kubenswrapper[4824]: I0121 11:12:34.071017 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:34 crc kubenswrapper[4824]: I0121 11:12:34.071026 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:34 crc kubenswrapper[4824]: I0121 11:12:34.071036 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:34 crc kubenswrapper[4824]: I0121 11:12:34.071044 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:34Z","lastTransitionTime":"2026-01-21T11:12:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:34 crc kubenswrapper[4824]: I0121 11:12:34.173305 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:34 crc kubenswrapper[4824]: I0121 11:12:34.173347 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:34 crc kubenswrapper[4824]: I0121 11:12:34.173356 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:34 crc kubenswrapper[4824]: I0121 11:12:34.173368 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:34 crc kubenswrapper[4824]: I0121 11:12:34.173376 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:34Z","lastTransitionTime":"2026-01-21T11:12:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:34 crc kubenswrapper[4824]: I0121 11:12:34.275681 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:34 crc kubenswrapper[4824]: I0121 11:12:34.275715 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:34 crc kubenswrapper[4824]: I0121 11:12:34.275723 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:34 crc kubenswrapper[4824]: I0121 11:12:34.275736 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:34 crc kubenswrapper[4824]: I0121 11:12:34.275744 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:34Z","lastTransitionTime":"2026-01-21T11:12:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:34 crc kubenswrapper[4824]: I0121 11:12:34.377644 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:34 crc kubenswrapper[4824]: I0121 11:12:34.377680 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:34 crc kubenswrapper[4824]: I0121 11:12:34.377689 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:34 crc kubenswrapper[4824]: I0121 11:12:34.377705 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:34 crc kubenswrapper[4824]: I0121 11:12:34.377716 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:34Z","lastTransitionTime":"2026-01-21T11:12:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:34 crc kubenswrapper[4824]: I0121 11:12:34.479320 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:34 crc kubenswrapper[4824]: I0121 11:12:34.479374 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:34 crc kubenswrapper[4824]: I0121 11:12:34.479383 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:34 crc kubenswrapper[4824]: I0121 11:12:34.479399 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:34 crc kubenswrapper[4824]: I0121 11:12:34.479408 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:34Z","lastTransitionTime":"2026-01-21T11:12:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:34 crc kubenswrapper[4824]: I0121 11:12:34.581620 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:34 crc kubenswrapper[4824]: I0121 11:12:34.581674 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:34 crc kubenswrapper[4824]: I0121 11:12:34.581683 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:34 crc kubenswrapper[4824]: I0121 11:12:34.581694 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:34 crc kubenswrapper[4824]: I0121 11:12:34.581702 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:34Z","lastTransitionTime":"2026-01-21T11:12:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:34 crc kubenswrapper[4824]: I0121 11:12:34.683540 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:34 crc kubenswrapper[4824]: I0121 11:12:34.683585 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:34 crc kubenswrapper[4824]: I0121 11:12:34.683594 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:34 crc kubenswrapper[4824]: I0121 11:12:34.683609 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:34 crc kubenswrapper[4824]: I0121 11:12:34.683618 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:34Z","lastTransitionTime":"2026-01-21T11:12:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:34 crc kubenswrapper[4824]: I0121 11:12:34.785646 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:34 crc kubenswrapper[4824]: I0121 11:12:34.785685 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:34 crc kubenswrapper[4824]: I0121 11:12:34.785694 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:34 crc kubenswrapper[4824]: I0121 11:12:34.785709 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:34 crc kubenswrapper[4824]: I0121 11:12:34.785719 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:34Z","lastTransitionTime":"2026-01-21T11:12:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:34 crc kubenswrapper[4824]: I0121 11:12:34.888047 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:34 crc kubenswrapper[4824]: I0121 11:12:34.888082 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:34 crc kubenswrapper[4824]: I0121 11:12:34.888090 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:34 crc kubenswrapper[4824]: I0121 11:12:34.888105 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:34 crc kubenswrapper[4824]: I0121 11:12:34.888113 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:34Z","lastTransitionTime":"2026-01-21T11:12:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:34 crc kubenswrapper[4824]: I0121 11:12:34.989671 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:34 crc kubenswrapper[4824]: I0121 11:12:34.989708 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:34 crc kubenswrapper[4824]: I0121 11:12:34.989719 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:34 crc kubenswrapper[4824]: I0121 11:12:34.989733 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:34 crc kubenswrapper[4824]: I0121 11:12:34.989742 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:34Z","lastTransitionTime":"2026-01-21T11:12:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:35 crc kubenswrapper[4824]: I0121 11:12:35.038400 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-16 09:30:56.989879119 +0000 UTC Jan 21 11:12:35 crc kubenswrapper[4824]: I0121 11:12:35.048670 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 11:12:35 crc kubenswrapper[4824]: I0121 11:12:35.048729 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 11:12:35 crc kubenswrapper[4824]: E0121 11:12:35.048790 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 21 11:12:35 crc kubenswrapper[4824]: E0121 11:12:35.048918 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 21 11:12:35 crc kubenswrapper[4824]: I0121 11:12:35.091693 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:35 crc kubenswrapper[4824]: I0121 11:12:35.091725 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:35 crc kubenswrapper[4824]: I0121 11:12:35.091732 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:35 crc kubenswrapper[4824]: I0121 11:12:35.091745 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:35 crc kubenswrapper[4824]: I0121 11:12:35.091755 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:35Z","lastTransitionTime":"2026-01-21T11:12:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:35 crc kubenswrapper[4824]: I0121 11:12:35.193902 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:35 crc kubenswrapper[4824]: I0121 11:12:35.193933 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:35 crc kubenswrapper[4824]: I0121 11:12:35.193941 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:35 crc kubenswrapper[4824]: I0121 11:12:35.193977 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:35 crc kubenswrapper[4824]: I0121 11:12:35.193986 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:35Z","lastTransitionTime":"2026-01-21T11:12:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:35 crc kubenswrapper[4824]: I0121 11:12:35.295858 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:35 crc kubenswrapper[4824]: I0121 11:12:35.295896 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:35 crc kubenswrapper[4824]: I0121 11:12:35.295906 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:35 crc kubenswrapper[4824]: I0121 11:12:35.295919 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:35 crc kubenswrapper[4824]: I0121 11:12:35.295927 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:35Z","lastTransitionTime":"2026-01-21T11:12:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:35 crc kubenswrapper[4824]: I0121 11:12:35.397262 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:35 crc kubenswrapper[4824]: I0121 11:12:35.397295 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:35 crc kubenswrapper[4824]: I0121 11:12:35.397304 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:35 crc kubenswrapper[4824]: I0121 11:12:35.397317 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:35 crc kubenswrapper[4824]: I0121 11:12:35.397339 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:35Z","lastTransitionTime":"2026-01-21T11:12:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:35 crc kubenswrapper[4824]: I0121 11:12:35.499648 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:35 crc kubenswrapper[4824]: I0121 11:12:35.499686 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:35 crc kubenswrapper[4824]: I0121 11:12:35.499709 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:35 crc kubenswrapper[4824]: I0121 11:12:35.499723 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:35 crc kubenswrapper[4824]: I0121 11:12:35.499731 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:35Z","lastTransitionTime":"2026-01-21T11:12:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:35 crc kubenswrapper[4824]: I0121 11:12:35.601901 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:35 crc kubenswrapper[4824]: I0121 11:12:35.601939 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:35 crc kubenswrapper[4824]: I0121 11:12:35.601948 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:35 crc kubenswrapper[4824]: I0121 11:12:35.601978 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:35 crc kubenswrapper[4824]: I0121 11:12:35.601987 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:35Z","lastTransitionTime":"2026-01-21T11:12:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:35 crc kubenswrapper[4824]: I0121 11:12:35.704642 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:35 crc kubenswrapper[4824]: I0121 11:12:35.704678 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:35 crc kubenswrapper[4824]: I0121 11:12:35.704687 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:35 crc kubenswrapper[4824]: I0121 11:12:35.704701 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:35 crc kubenswrapper[4824]: I0121 11:12:35.704709 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:35Z","lastTransitionTime":"2026-01-21T11:12:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:35 crc kubenswrapper[4824]: I0121 11:12:35.806992 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:35 crc kubenswrapper[4824]: I0121 11:12:35.807027 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:35 crc kubenswrapper[4824]: I0121 11:12:35.807036 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:35 crc kubenswrapper[4824]: I0121 11:12:35.807048 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:35 crc kubenswrapper[4824]: I0121 11:12:35.807056 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:35Z","lastTransitionTime":"2026-01-21T11:12:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:35 crc kubenswrapper[4824]: I0121 11:12:35.908755 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:35 crc kubenswrapper[4824]: I0121 11:12:35.908789 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:35 crc kubenswrapper[4824]: I0121 11:12:35.908799 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:35 crc kubenswrapper[4824]: I0121 11:12:35.908813 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:35 crc kubenswrapper[4824]: I0121 11:12:35.908822 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:35Z","lastTransitionTime":"2026-01-21T11:12:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:36 crc kubenswrapper[4824]: I0121 11:12:36.013940 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:36 crc kubenswrapper[4824]: I0121 11:12:36.013996 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:36 crc kubenswrapper[4824]: I0121 11:12:36.014005 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:36 crc kubenswrapper[4824]: I0121 11:12:36.014021 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:36 crc kubenswrapper[4824]: I0121 11:12:36.014030 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:36Z","lastTransitionTime":"2026-01-21T11:12:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:36 crc kubenswrapper[4824]: I0121 11:12:36.050280 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-25 11:01:25.456081842 +0000 UTC Jan 21 11:12:36 crc kubenswrapper[4824]: I0121 11:12:36.050335 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9rmnw" Jan 21 11:12:36 crc kubenswrapper[4824]: E0121 11:12:36.050443 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9rmnw" podUID="8f9bd12b-22ee-44ed-a8b9-c100d043f691" Jan 21 11:12:36 crc kubenswrapper[4824]: I0121 11:12:36.050468 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 11:12:36 crc kubenswrapper[4824]: E0121 11:12:36.050549 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 21 11:12:36 crc kubenswrapper[4824]: I0121 11:12:36.115855 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:36 crc kubenswrapper[4824]: I0121 11:12:36.115892 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:36 crc kubenswrapper[4824]: I0121 11:12:36.115899 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:36 crc kubenswrapper[4824]: I0121 11:12:36.115913 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:36 crc kubenswrapper[4824]: I0121 11:12:36.115922 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:36Z","lastTransitionTime":"2026-01-21T11:12:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:36 crc kubenswrapper[4824]: I0121 11:12:36.218096 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:36 crc kubenswrapper[4824]: I0121 11:12:36.218133 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:36 crc kubenswrapper[4824]: I0121 11:12:36.218141 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:36 crc kubenswrapper[4824]: I0121 11:12:36.218154 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:36 crc kubenswrapper[4824]: I0121 11:12:36.218163 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:36Z","lastTransitionTime":"2026-01-21T11:12:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:36 crc kubenswrapper[4824]: I0121 11:12:36.319801 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:36 crc kubenswrapper[4824]: I0121 11:12:36.319837 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:36 crc kubenswrapper[4824]: I0121 11:12:36.319847 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:36 crc kubenswrapper[4824]: I0121 11:12:36.319861 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:36 crc kubenswrapper[4824]: I0121 11:12:36.319870 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:36Z","lastTransitionTime":"2026-01-21T11:12:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:36 crc kubenswrapper[4824]: I0121 11:12:36.421413 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:36 crc kubenswrapper[4824]: I0121 11:12:36.421452 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:36 crc kubenswrapper[4824]: I0121 11:12:36.421460 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:36 crc kubenswrapper[4824]: I0121 11:12:36.421474 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:36 crc kubenswrapper[4824]: I0121 11:12:36.421481 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:36Z","lastTransitionTime":"2026-01-21T11:12:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:36 crc kubenswrapper[4824]: I0121 11:12:36.523705 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:36 crc kubenswrapper[4824]: I0121 11:12:36.523745 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:36 crc kubenswrapper[4824]: I0121 11:12:36.523754 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:36 crc kubenswrapper[4824]: I0121 11:12:36.523770 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:36 crc kubenswrapper[4824]: I0121 11:12:36.523779 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:36Z","lastTransitionTime":"2026-01-21T11:12:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:36 crc kubenswrapper[4824]: I0121 11:12:36.625670 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:36 crc kubenswrapper[4824]: I0121 11:12:36.625720 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:36 crc kubenswrapper[4824]: I0121 11:12:36.625730 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:36 crc kubenswrapper[4824]: I0121 11:12:36.625742 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:36 crc kubenswrapper[4824]: I0121 11:12:36.625750 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:36Z","lastTransitionTime":"2026-01-21T11:12:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:36 crc kubenswrapper[4824]: I0121 11:12:36.726844 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:36 crc kubenswrapper[4824]: I0121 11:12:36.726877 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:36 crc kubenswrapper[4824]: I0121 11:12:36.726886 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:36 crc kubenswrapper[4824]: I0121 11:12:36.726900 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:36 crc kubenswrapper[4824]: I0121 11:12:36.726909 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:36Z","lastTransitionTime":"2026-01-21T11:12:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:36 crc kubenswrapper[4824]: I0121 11:12:36.828653 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:36 crc kubenswrapper[4824]: I0121 11:12:36.828691 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:36 crc kubenswrapper[4824]: I0121 11:12:36.828699 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:36 crc kubenswrapper[4824]: I0121 11:12:36.828713 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:36 crc kubenswrapper[4824]: I0121 11:12:36.828723 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:36Z","lastTransitionTime":"2026-01-21T11:12:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:36 crc kubenswrapper[4824]: I0121 11:12:36.930415 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:36 crc kubenswrapper[4824]: I0121 11:12:36.930451 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:36 crc kubenswrapper[4824]: I0121 11:12:36.930461 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:36 crc kubenswrapper[4824]: I0121 11:12:36.930473 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:36 crc kubenswrapper[4824]: I0121 11:12:36.930482 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:36Z","lastTransitionTime":"2026-01-21T11:12:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:37 crc kubenswrapper[4824]: I0121 11:12:37.032521 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:37 crc kubenswrapper[4824]: I0121 11:12:37.032560 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:37 crc kubenswrapper[4824]: I0121 11:12:37.032569 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:37 crc kubenswrapper[4824]: I0121 11:12:37.032584 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:37 crc kubenswrapper[4824]: I0121 11:12:37.032592 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:37Z","lastTransitionTime":"2026-01-21T11:12:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:37 crc kubenswrapper[4824]: I0121 11:12:37.048864 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 11:12:37 crc kubenswrapper[4824]: I0121 11:12:37.048923 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 11:12:37 crc kubenswrapper[4824]: E0121 11:12:37.049019 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 21 11:12:37 crc kubenswrapper[4824]: E0121 11:12:37.049089 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 21 11:12:37 crc kubenswrapper[4824]: I0121 11:12:37.051129 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-07 19:12:19.471814047 +0000 UTC Jan 21 11:12:37 crc kubenswrapper[4824]: I0121 11:12:37.134646 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:37 crc kubenswrapper[4824]: I0121 11:12:37.134680 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:37 crc kubenswrapper[4824]: I0121 11:12:37.134688 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:37 crc kubenswrapper[4824]: I0121 11:12:37.134704 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:37 crc kubenswrapper[4824]: I0121 11:12:37.134713 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:37Z","lastTransitionTime":"2026-01-21T11:12:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:37 crc kubenswrapper[4824]: I0121 11:12:37.236715 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:37 crc kubenswrapper[4824]: I0121 11:12:37.236759 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:37 crc kubenswrapper[4824]: I0121 11:12:37.236788 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:37 crc kubenswrapper[4824]: I0121 11:12:37.236805 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:37 crc kubenswrapper[4824]: I0121 11:12:37.236815 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:37Z","lastTransitionTime":"2026-01-21T11:12:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:37 crc kubenswrapper[4824]: I0121 11:12:37.338429 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:37 crc kubenswrapper[4824]: I0121 11:12:37.338466 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:37 crc kubenswrapper[4824]: I0121 11:12:37.338476 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:37 crc kubenswrapper[4824]: I0121 11:12:37.338490 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:37 crc kubenswrapper[4824]: I0121 11:12:37.338498 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:37Z","lastTransitionTime":"2026-01-21T11:12:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:37 crc kubenswrapper[4824]: I0121 11:12:37.440005 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:37 crc kubenswrapper[4824]: I0121 11:12:37.440040 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:37 crc kubenswrapper[4824]: I0121 11:12:37.440048 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:37 crc kubenswrapper[4824]: I0121 11:12:37.440062 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:37 crc kubenswrapper[4824]: I0121 11:12:37.440070 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:37Z","lastTransitionTime":"2026-01-21T11:12:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:37 crc kubenswrapper[4824]: I0121 11:12:37.542150 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:37 crc kubenswrapper[4824]: I0121 11:12:37.542188 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:37 crc kubenswrapper[4824]: I0121 11:12:37.542197 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:37 crc kubenswrapper[4824]: I0121 11:12:37.542211 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:37 crc kubenswrapper[4824]: I0121 11:12:37.542219 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:37Z","lastTransitionTime":"2026-01-21T11:12:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:37 crc kubenswrapper[4824]: I0121 11:12:37.644381 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:37 crc kubenswrapper[4824]: I0121 11:12:37.644414 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:37 crc kubenswrapper[4824]: I0121 11:12:37.644423 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:37 crc kubenswrapper[4824]: I0121 11:12:37.644437 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:37 crc kubenswrapper[4824]: I0121 11:12:37.644446 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:37Z","lastTransitionTime":"2026-01-21T11:12:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:37 crc kubenswrapper[4824]: I0121 11:12:37.746706 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:37 crc kubenswrapper[4824]: I0121 11:12:37.747290 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:37 crc kubenswrapper[4824]: I0121 11:12:37.747304 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:37 crc kubenswrapper[4824]: I0121 11:12:37.748032 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:37 crc kubenswrapper[4824]: I0121 11:12:37.748051 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:37Z","lastTransitionTime":"2026-01-21T11:12:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:37 crc kubenswrapper[4824]: I0121 11:12:37.849788 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:37 crc kubenswrapper[4824]: I0121 11:12:37.849822 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:37 crc kubenswrapper[4824]: I0121 11:12:37.849831 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:37 crc kubenswrapper[4824]: I0121 11:12:37.849847 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:37 crc kubenswrapper[4824]: I0121 11:12:37.849858 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:37Z","lastTransitionTime":"2026-01-21T11:12:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:37 crc kubenswrapper[4824]: I0121 11:12:37.951715 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:37 crc kubenswrapper[4824]: I0121 11:12:37.951743 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:37 crc kubenswrapper[4824]: I0121 11:12:37.951751 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:37 crc kubenswrapper[4824]: I0121 11:12:37.951762 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:37 crc kubenswrapper[4824]: I0121 11:12:37.951772 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:37Z","lastTransitionTime":"2026-01-21T11:12:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:38 crc kubenswrapper[4824]: I0121 11:12:38.048363 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9rmnw" Jan 21 11:12:38 crc kubenswrapper[4824]: E0121 11:12:38.048493 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9rmnw" podUID="8f9bd12b-22ee-44ed-a8b9-c100d043f691" Jan 21 11:12:38 crc kubenswrapper[4824]: I0121 11:12:38.048549 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 11:12:38 crc kubenswrapper[4824]: E0121 11:12:38.048842 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 21 11:12:38 crc kubenswrapper[4824]: I0121 11:12:38.051261 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-15 15:21:44.248961456 +0000 UTC Jan 21 11:12:38 crc kubenswrapper[4824]: I0121 11:12:38.053707 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:38 crc kubenswrapper[4824]: I0121 11:12:38.053747 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:38 crc kubenswrapper[4824]: I0121 11:12:38.053756 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:38 crc kubenswrapper[4824]: I0121 11:12:38.053766 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:38 crc kubenswrapper[4824]: I0121 11:12:38.053775 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:38Z","lastTransitionTime":"2026-01-21T11:12:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:38 crc kubenswrapper[4824]: I0121 11:12:38.057104 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lv4rn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1267539-9adf-466f-8027-8ede009d19a0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eed1b884aa53e1ee318169a0cc2c9ad10130aa363ae4ac55e4bf065237756f53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbmfl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:15Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lv4rn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:38Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:38 crc kubenswrapper[4824]: I0121 11:12:38.064170 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-jr64s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9768c60d-5bee-4286-bced-076e584b04e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91158f14d2e98ac9f875a5e7101df3107ef28aa5fc5d62218f460af98ff920fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-flrnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:17Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-jr64s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:38Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:38 crc kubenswrapper[4824]: I0121 11:12:38.070715 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b44f8a34-1578-4634-b62e-8e222844b397\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fd58f2438b6f9cbb2c37687e39f02cdf96a029477b1ae33f6dc9f3b70ecc926\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef10a172a3279201375f7445673820033978ce1eb2bbef7a2c0ed297fd783875\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef10a172a3279201375f7445673820033978ce1eb2bbef7a2c0ed297fd783875\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:38Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:38 crc kubenswrapper[4824]: I0121 11:12:38.079309 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d31ab61-f94c-4f49-85d0-c9e91b83069b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a66b5c33884af00408dc35cf02f08ebd8570bd695f9f1fe558bf54fd06930431\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7514786602746dabfa690da48d48e233fe68d6dd6674aa5ecd489c0b3185f08\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61b14b6846f94cdeba8f1f61b3754e7663041f92c70356c4ee6cbb988a2e680f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a7023cfe4c098bcdd9c0477a8187f4ed7a0d2abdbe1e143436863948f441b4f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af11ad95bcf0c6bbd5fa4ab99bb592ce1c87f40d29c5ade7832414315820a569\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW0121 11:11:14.604290 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0121 11:11:14.604572 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0121 11:11:14.605875 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3284452394/tls.crt::/tmp/serving-cert-3284452394/tls.key\\\\\\\"\\\\nI0121 11:11:14.923427 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0121 11:11:14.928141 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0121 11:11:14.928162 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0121 11:11:14.928179 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0121 11:11:14.928184 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0121 11:11:14.933154 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0121 11:11:14.933230 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 11:11:14.933259 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0121 11:11:14.933282 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0121 11:11:14.933299 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0121 11:11:14.933315 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0121 11:11:14.933331 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0121 11:11:14.933177 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0121 11:11:14.934666 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48070c09fad511108ff22e44211cf26911509b8406d66948985cbfc3009d8233\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:38Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:38 crc kubenswrapper[4824]: I0121 11:12:38.088735 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9bb99c0a-683b-44da-9c0e-e8d284e87d53\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05e076febd697ed802ce5c7ee92a8c2c636a1a2c6609ad7eb01f17dace247c4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43529d1b9e1e5e22b0d56b023946119da8aacbae54b68c26ff0a1a82dc890c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ea232ad87393685aa323ce6393ff3869f8f8d423c2406a501c69c355535adfa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc6c7eff9e17b8a8ecc88bbfe703b5a01dbf1e02d342cede4f6a263a6526af43\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:38Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:38 crc kubenswrapper[4824]: I0121 11:12:38.096256 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e570fff1-6227-44d6-b71f-ee377b1099ae\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4d41492c3802a8168fffd249c15aeb9c902a5c7e01550e1aa6f9b769adfdc4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2690a1ecbc8fe1290e68bba2695a3edd833b5da933fb71e8ddc926a1340a9a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b85db961e1e078e1f54323c79a99c9413365256ac2fce411db7932bd0580ab6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://595783a1454735bd33f4862180b60d99a4e1b6fae1e735f2b7328391951bf422\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://595783a1454735bd33f4862180b60d99a4e1b6fae1e735f2b7328391951bf422\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:38Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:38 crc kubenswrapper[4824]: I0121 11:12:38.103653 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:38Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:38 crc kubenswrapper[4824]: I0121 11:12:38.115812 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"06f56d86-ec6f-4b57-b314-f508c0dd4d9f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d25c94f1ae0bbabb733f0e1ea1d1ba39562a92af73047a39aacc03013e239681\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca06234486b59cf990c69d9501e23f935328e2ba21013cd71ed64831ea749bcb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b63e45d66bcd11ff3479cc3ec6b450017da8d21d36e967b8d496ae0f24b0ee9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52758bcd31f2ba64edf03adfd4d1be6428015abff5b64c7c029fa787085f9c29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2736e73342d2adb175b0721dd4f40cc42d26fca06d5eb3334927518531c3b107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-21T11:11:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b35ad7dd0570c191eb0080739304f3cd43f17b3d1dddf895549441ebe50deaa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b35ad7dd0570c191eb0080739304f3cd43f17b3d1dddf895549441ebe50deaa6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://80cd5ea5e6fc19a88c1aab5c61f827f103ab798a5c84da8dc4d45cea86d5aacb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80cd5ea5e6fc19a88c1aab5c61f827f103ab798a5c84da8dc4d45cea86d5aacb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:10:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:10:59Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://234d8b914e6a88480021647ae65d74e9b21c55b6788312a094d00c8a42677d48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://234d8b914e6a88480021647ae65d74e9b21c55b6788312a094d00c8a42677d48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-21T11:11:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-21T11:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:10:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:38Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:38 crc kubenswrapper[4824]: I0121 11:12:38.123462 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:38Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:38 crc kubenswrapper[4824]: I0121 11:12:38.129771 4824 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-9rmnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f9bd12b-22ee-44ed-a8b9-c100d043f691\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-21T11:11:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2fgn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q2fgn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-21T11:11:28Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-9rmnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-21T11:12:38Z is after 2025-08-24T17:21:41Z" Jan 21 11:12:38 crc kubenswrapper[4824]: I0121 11:12:38.155643 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:38 crc kubenswrapper[4824]: I0121 11:12:38.155676 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:38 crc kubenswrapper[4824]: I0121 11:12:38.155685 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:38 crc kubenswrapper[4824]: I0121 11:12:38.155700 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:38 crc kubenswrapper[4824]: I0121 11:12:38.155709 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:38Z","lastTransitionTime":"2026-01-21T11:12:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:38 crc kubenswrapper[4824]: I0121 11:12:38.163201 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-x94nb" podStartSLOduration=84.163189144 podStartE2EDuration="1m24.163189144s" podCreationTimestamp="2026-01-21 11:11:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:12:38.162860426 +0000 UTC m=+100.455889717" watchObservedRunningTime="2026-01-21 11:12:38.163189144 +0000 UTC m=+100.456218436" Jan 21 11:12:38 crc kubenswrapper[4824]: I0121 11:12:38.181046 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-gjtnh" podStartSLOduration=84.181034891 podStartE2EDuration="1m24.181034891s" podCreationTimestamp="2026-01-21 11:11:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:12:38.180655107 +0000 UTC m=+100.473684399" watchObservedRunningTime="2026-01-21 11:12:38.181034891 +0000 UTC m=+100.474064184" Jan 21 11:12:38 crc kubenswrapper[4824]: I0121 11:12:38.181210 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podStartSLOduration=84.181205713 podStartE2EDuration="1m24.181205713s" podCreationTimestamp="2026-01-21 11:11:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:12:38.170652772 +0000 UTC m=+100.463682064" watchObservedRunningTime="2026-01-21 11:12:38.181205713 +0000 UTC m=+100.474235005" Jan 21 11:12:38 crc kubenswrapper[4824]: I0121 11:12:38.258945 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:38 crc kubenswrapper[4824]: I0121 11:12:38.258998 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:38 crc kubenswrapper[4824]: I0121 11:12:38.259007 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:38 crc kubenswrapper[4824]: I0121 11:12:38.259021 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:38 crc kubenswrapper[4824]: I0121 11:12:38.259029 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:38Z","lastTransitionTime":"2026-01-21T11:12:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:38 crc kubenswrapper[4824]: I0121 11:12:38.361290 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:38 crc kubenswrapper[4824]: I0121 11:12:38.361339 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:38 crc kubenswrapper[4824]: I0121 11:12:38.361350 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:38 crc kubenswrapper[4824]: I0121 11:12:38.361368 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:38 crc kubenswrapper[4824]: I0121 11:12:38.361379 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:38Z","lastTransitionTime":"2026-01-21T11:12:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:38 crc kubenswrapper[4824]: I0121 11:12:38.463198 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:38 crc kubenswrapper[4824]: I0121 11:12:38.463229 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:38 crc kubenswrapper[4824]: I0121 11:12:38.463238 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:38 crc kubenswrapper[4824]: I0121 11:12:38.463251 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:38 crc kubenswrapper[4824]: I0121 11:12:38.463258 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:38Z","lastTransitionTime":"2026-01-21T11:12:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:38 crc kubenswrapper[4824]: I0121 11:12:38.565360 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:38 crc kubenswrapper[4824]: I0121 11:12:38.565411 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:38 crc kubenswrapper[4824]: I0121 11:12:38.565421 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:38 crc kubenswrapper[4824]: I0121 11:12:38.565438 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:38 crc kubenswrapper[4824]: I0121 11:12:38.565450 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:38Z","lastTransitionTime":"2026-01-21T11:12:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:38 crc kubenswrapper[4824]: I0121 11:12:38.666993 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:38 crc kubenswrapper[4824]: I0121 11:12:38.667041 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:38 crc kubenswrapper[4824]: I0121 11:12:38.667050 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:38 crc kubenswrapper[4824]: I0121 11:12:38.667065 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:38 crc kubenswrapper[4824]: I0121 11:12:38.667075 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:38Z","lastTransitionTime":"2026-01-21T11:12:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:38 crc kubenswrapper[4824]: I0121 11:12:38.768879 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:38 crc kubenswrapper[4824]: I0121 11:12:38.768922 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:38 crc kubenswrapper[4824]: I0121 11:12:38.768932 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:38 crc kubenswrapper[4824]: I0121 11:12:38.768946 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:38 crc kubenswrapper[4824]: I0121 11:12:38.768969 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:38Z","lastTransitionTime":"2026-01-21T11:12:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:38 crc kubenswrapper[4824]: I0121 11:12:38.870866 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:38 crc kubenswrapper[4824]: I0121 11:12:38.870895 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:38 crc kubenswrapper[4824]: I0121 11:12:38.870903 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:38 crc kubenswrapper[4824]: I0121 11:12:38.870915 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:38 crc kubenswrapper[4824]: I0121 11:12:38.870925 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:38Z","lastTransitionTime":"2026-01-21T11:12:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:38 crc kubenswrapper[4824]: I0121 11:12:38.972907 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:38 crc kubenswrapper[4824]: I0121 11:12:38.972939 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:38 crc kubenswrapper[4824]: I0121 11:12:38.972948 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:38 crc kubenswrapper[4824]: I0121 11:12:38.972988 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:38 crc kubenswrapper[4824]: I0121 11:12:38.972998 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:38Z","lastTransitionTime":"2026-01-21T11:12:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:39 crc kubenswrapper[4824]: I0121 11:12:39.049040 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 11:12:39 crc kubenswrapper[4824]: I0121 11:12:39.049046 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 11:12:39 crc kubenswrapper[4824]: E0121 11:12:39.049174 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 21 11:12:39 crc kubenswrapper[4824]: E0121 11:12:39.049254 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 21 11:12:39 crc kubenswrapper[4824]: I0121 11:12:39.052265 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-13 04:23:10.736867186 +0000 UTC Jan 21 11:12:39 crc kubenswrapper[4824]: I0121 11:12:39.074922 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:39 crc kubenswrapper[4824]: I0121 11:12:39.074952 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:39 crc kubenswrapper[4824]: I0121 11:12:39.074978 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:39 crc kubenswrapper[4824]: I0121 11:12:39.074992 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:39 crc kubenswrapper[4824]: I0121 11:12:39.075003 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:39Z","lastTransitionTime":"2026-01-21T11:12:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:39 crc kubenswrapper[4824]: I0121 11:12:39.176903 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:39 crc kubenswrapper[4824]: I0121 11:12:39.176935 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:39 crc kubenswrapper[4824]: I0121 11:12:39.176943 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:39 crc kubenswrapper[4824]: I0121 11:12:39.176973 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:39 crc kubenswrapper[4824]: I0121 11:12:39.176989 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:39Z","lastTransitionTime":"2026-01-21T11:12:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:39 crc kubenswrapper[4824]: I0121 11:12:39.279003 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:39 crc kubenswrapper[4824]: I0121 11:12:39.279071 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:39 crc kubenswrapper[4824]: I0121 11:12:39.279082 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:39 crc kubenswrapper[4824]: I0121 11:12:39.279106 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:39 crc kubenswrapper[4824]: I0121 11:12:39.279120 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:39Z","lastTransitionTime":"2026-01-21T11:12:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:39 crc kubenswrapper[4824]: I0121 11:12:39.280662 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 21 11:12:39 crc kubenswrapper[4824]: I0121 11:12:39.280691 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 21 11:12:39 crc kubenswrapper[4824]: I0121 11:12:39.280700 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 21 11:12:39 crc kubenswrapper[4824]: I0121 11:12:39.280711 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 21 11:12:39 crc kubenswrapper[4824]: I0121 11:12:39.280720 4824 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-21T11:12:39Z","lastTransitionTime":"2026-01-21T11:12:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 21 11:12:39 crc kubenswrapper[4824]: I0121 11:12:39.306346 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-np2v8" podStartSLOduration=84.30632603 podStartE2EDuration="1m24.30632603s" podCreationTimestamp="2026-01-21 11:11:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:12:38.218257308 +0000 UTC m=+100.511286600" watchObservedRunningTime="2026-01-21 11:12:39.30632603 +0000 UTC m=+101.599355321" Jan 21 11:12:39 crc kubenswrapper[4824]: I0121 11:12:39.306905 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-xnwp2"] Jan 21 11:12:39 crc kubenswrapper[4824]: I0121 11:12:39.307292 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xnwp2" Jan 21 11:12:39 crc kubenswrapper[4824]: I0121 11:12:39.308570 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Jan 21 11:12:39 crc kubenswrapper[4824]: I0121 11:12:39.308610 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Jan 21 11:12:39 crc kubenswrapper[4824]: I0121 11:12:39.309520 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Jan 21 11:12:39 crc kubenswrapper[4824]: I0121 11:12:39.309939 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Jan 21 11:12:39 crc kubenswrapper[4824]: I0121 11:12:39.318073 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=85.318056528 podStartE2EDuration="1m25.318056528s" podCreationTimestamp="2026-01-21 11:11:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:12:39.317723341 +0000 UTC m=+101.610752633" watchObservedRunningTime="2026-01-21 11:12:39.318056528 +0000 UTC m=+101.611085820" Jan 21 11:12:39 crc kubenswrapper[4824]: I0121 11:12:39.327235 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=46.327217951 podStartE2EDuration="46.327217951s" podCreationTimestamp="2026-01-21 11:11:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:12:39.326564851 +0000 UTC m=+101.619594144" watchObservedRunningTime="2026-01-21 11:12:39.327217951 +0000 UTC m=+101.620247242" Jan 21 11:12:39 crc kubenswrapper[4824]: I0121 11:12:39.343589 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/3f2fe1cd-26f3-4484-86c5-d71cda217f5f-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-xnwp2\" (UID: \"3f2fe1cd-26f3-4484-86c5-d71cda217f5f\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xnwp2" Jan 21 11:12:39 crc kubenswrapper[4824]: I0121 11:12:39.343620 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3f2fe1cd-26f3-4484-86c5-d71cda217f5f-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-xnwp2\" (UID: \"3f2fe1cd-26f3-4484-86c5-d71cda217f5f\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xnwp2" Jan 21 11:12:39 crc kubenswrapper[4824]: I0121 11:12:39.343638 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3f2fe1cd-26f3-4484-86c5-d71cda217f5f-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-xnwp2\" (UID: \"3f2fe1cd-26f3-4484-86c5-d71cda217f5f\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xnwp2" Jan 21 11:12:39 crc kubenswrapper[4824]: I0121 11:12:39.343666 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/3f2fe1cd-26f3-4484-86c5-d71cda217f5f-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-xnwp2\" (UID: \"3f2fe1cd-26f3-4484-86c5-d71cda217f5f\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xnwp2" Jan 21 11:12:39 crc kubenswrapper[4824]: I0121 11:12:39.343759 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/3f2fe1cd-26f3-4484-86c5-d71cda217f5f-service-ca\") pod \"cluster-version-operator-5c965bbfc6-xnwp2\" (UID: \"3f2fe1cd-26f3-4484-86c5-d71cda217f5f\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xnwp2" Jan 21 11:12:39 crc kubenswrapper[4824]: I0121 11:12:39.344125 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-lv4rn" podStartSLOduration=85.344106837 podStartE2EDuration="1m25.344106837s" podCreationTimestamp="2026-01-21 11:11:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:12:39.344034691 +0000 UTC m=+101.637063983" watchObservedRunningTime="2026-01-21 11:12:39.344106837 +0000 UTC m=+101.637136129" Jan 21 11:12:39 crc kubenswrapper[4824]: I0121 11:12:39.352917 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-jr64s" podStartSLOduration=85.352901259 podStartE2EDuration="1m25.352901259s" podCreationTimestamp="2026-01-21 11:11:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:12:39.352323471 +0000 UTC m=+101.645352764" watchObservedRunningTime="2026-01-21 11:12:39.352901259 +0000 UTC m=+101.645930552" Jan 21 11:12:39 crc kubenswrapper[4824]: I0121 11:12:39.366075 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=25.366065917 podStartE2EDuration="25.366065917s" podCreationTimestamp="2026-01-21 11:12:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:12:39.365574161 +0000 UTC m=+101.658603463" watchObservedRunningTime="2026-01-21 11:12:39.366065917 +0000 UTC m=+101.659095209" Jan 21 11:12:39 crc kubenswrapper[4824]: I0121 11:12:39.375746 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=84.37572735 podStartE2EDuration="1m24.37572735s" podCreationTimestamp="2026-01-21 11:11:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:12:39.375543515 +0000 UTC m=+101.668572817" watchObservedRunningTime="2026-01-21 11:12:39.37572735 +0000 UTC m=+101.668756643" Jan 21 11:12:39 crc kubenswrapper[4824]: I0121 11:12:39.393837 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=84.393821465 podStartE2EDuration="1m24.393821465s" podCreationTimestamp="2026-01-21 11:11:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:12:39.391477915 +0000 UTC m=+101.684507207" watchObservedRunningTime="2026-01-21 11:12:39.393821465 +0000 UTC m=+101.686850757" Jan 21 11:12:39 crc kubenswrapper[4824]: I0121 11:12:39.445461 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3f2fe1cd-26f3-4484-86c5-d71cda217f5f-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-xnwp2\" (UID: \"3f2fe1cd-26f3-4484-86c5-d71cda217f5f\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xnwp2" Jan 21 11:12:39 crc kubenswrapper[4824]: I0121 11:12:39.445514 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/3f2fe1cd-26f3-4484-86c5-d71cda217f5f-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-xnwp2\" (UID: \"3f2fe1cd-26f3-4484-86c5-d71cda217f5f\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xnwp2" Jan 21 11:12:39 crc kubenswrapper[4824]: I0121 11:12:39.445539 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3f2fe1cd-26f3-4484-86c5-d71cda217f5f-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-xnwp2\" (UID: \"3f2fe1cd-26f3-4484-86c5-d71cda217f5f\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xnwp2" Jan 21 11:12:39 crc kubenswrapper[4824]: I0121 11:12:39.445572 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/3f2fe1cd-26f3-4484-86c5-d71cda217f5f-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-xnwp2\" (UID: \"3f2fe1cd-26f3-4484-86c5-d71cda217f5f\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xnwp2" Jan 21 11:12:39 crc kubenswrapper[4824]: I0121 11:12:39.445638 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/3f2fe1cd-26f3-4484-86c5-d71cda217f5f-service-ca\") pod \"cluster-version-operator-5c965bbfc6-xnwp2\" (UID: \"3f2fe1cd-26f3-4484-86c5-d71cda217f5f\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xnwp2" Jan 21 11:12:39 crc kubenswrapper[4824]: I0121 11:12:39.445697 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/3f2fe1cd-26f3-4484-86c5-d71cda217f5f-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-xnwp2\" (UID: \"3f2fe1cd-26f3-4484-86c5-d71cda217f5f\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xnwp2" Jan 21 11:12:39 crc kubenswrapper[4824]: I0121 11:12:39.445821 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/3f2fe1cd-26f3-4484-86c5-d71cda217f5f-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-xnwp2\" (UID: \"3f2fe1cd-26f3-4484-86c5-d71cda217f5f\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xnwp2" Jan 21 11:12:39 crc kubenswrapper[4824]: I0121 11:12:39.446477 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/3f2fe1cd-26f3-4484-86c5-d71cda217f5f-service-ca\") pod \"cluster-version-operator-5c965bbfc6-xnwp2\" (UID: \"3f2fe1cd-26f3-4484-86c5-d71cda217f5f\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xnwp2" Jan 21 11:12:39 crc kubenswrapper[4824]: I0121 11:12:39.450441 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3f2fe1cd-26f3-4484-86c5-d71cda217f5f-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-xnwp2\" (UID: \"3f2fe1cd-26f3-4484-86c5-d71cda217f5f\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xnwp2" Jan 21 11:12:39 crc kubenswrapper[4824]: I0121 11:12:39.462147 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3f2fe1cd-26f3-4484-86c5-d71cda217f5f-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-xnwp2\" (UID: \"3f2fe1cd-26f3-4484-86c5-d71cda217f5f\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xnwp2" Jan 21 11:12:39 crc kubenswrapper[4824]: I0121 11:12:39.617091 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xnwp2" Jan 21 11:12:40 crc kubenswrapper[4824]: I0121 11:12:40.048352 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9rmnw" Jan 21 11:12:40 crc kubenswrapper[4824]: E0121 11:12:40.048515 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9rmnw" podUID="8f9bd12b-22ee-44ed-a8b9-c100d043f691" Jan 21 11:12:40 crc kubenswrapper[4824]: I0121 11:12:40.048536 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 11:12:40 crc kubenswrapper[4824]: E0121 11:12:40.048654 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 21 11:12:40 crc kubenswrapper[4824]: I0121 11:12:40.053037 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-22 00:15:47.311527941 +0000 UTC Jan 21 11:12:40 crc kubenswrapper[4824]: I0121 11:12:40.053201 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Rotating certificates Jan 21 11:12:40 crc kubenswrapper[4824]: I0121 11:12:40.058353 4824 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Jan 21 11:12:40 crc kubenswrapper[4824]: I0121 11:12:40.393461 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xnwp2" event={"ID":"3f2fe1cd-26f3-4484-86c5-d71cda217f5f","Type":"ContainerStarted","Data":"ecaecae9b4699a601f5a9b20e08f49084d84f678317c3be0710aa3a5090e1411"} Jan 21 11:12:40 crc kubenswrapper[4824]: I0121 11:12:40.393501 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xnwp2" event={"ID":"3f2fe1cd-26f3-4484-86c5-d71cda217f5f","Type":"ContainerStarted","Data":"0208f03c73487747fdc2f6a321ec1667db4975ce87f6b1a134782fb4bc91eeb9"} Jan 21 11:12:41 crc kubenswrapper[4824]: I0121 11:12:41.048754 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 11:12:41 crc kubenswrapper[4824]: E0121 11:12:41.048879 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 21 11:12:41 crc kubenswrapper[4824]: I0121 11:12:41.049495 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 11:12:41 crc kubenswrapper[4824]: E0121 11:12:41.049682 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 21 11:12:42 crc kubenswrapper[4824]: I0121 11:12:42.049127 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 11:12:42 crc kubenswrapper[4824]: I0121 11:12:42.049150 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9rmnw" Jan 21 11:12:42 crc kubenswrapper[4824]: E0121 11:12:42.049274 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 21 11:12:42 crc kubenswrapper[4824]: E0121 11:12:42.049475 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9rmnw" podUID="8f9bd12b-22ee-44ed-a8b9-c100d043f691" Jan 21 11:12:43 crc kubenswrapper[4824]: I0121 11:12:43.048904 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 11:12:43 crc kubenswrapper[4824]: I0121 11:12:43.048904 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 11:12:43 crc kubenswrapper[4824]: E0121 11:12:43.049016 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 21 11:12:43 crc kubenswrapper[4824]: E0121 11:12:43.049092 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 21 11:12:43 crc kubenswrapper[4824]: I0121 11:12:43.049598 4824 scope.go:117] "RemoveContainer" containerID="86a6881f109dab1f910fe505a21c7c800e587de4dded2635147997c51c05c483" Jan 21 11:12:43 crc kubenswrapper[4824]: E0121 11:12:43.049733 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-htkvk_openshift-ovn-kubernetes(26b1108a-4b98-4e39-a3cd-e0d055089fd8)\"" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" podUID="26b1108a-4b98-4e39-a3cd-e0d055089fd8" Jan 21 11:12:44 crc kubenswrapper[4824]: I0121 11:12:44.049109 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9rmnw" Jan 21 11:12:44 crc kubenswrapper[4824]: I0121 11:12:44.049213 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 11:12:44 crc kubenswrapper[4824]: E0121 11:12:44.049361 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9rmnw" podUID="8f9bd12b-22ee-44ed-a8b9-c100d043f691" Jan 21 11:12:44 crc kubenswrapper[4824]: E0121 11:12:44.049475 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 21 11:12:45 crc kubenswrapper[4824]: I0121 11:12:45.048596 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 11:12:45 crc kubenswrapper[4824]: I0121 11:12:45.048676 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 11:12:45 crc kubenswrapper[4824]: E0121 11:12:45.048716 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 21 11:12:45 crc kubenswrapper[4824]: E0121 11:12:45.048809 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 21 11:12:46 crc kubenswrapper[4824]: I0121 11:12:46.048716 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9rmnw" Jan 21 11:12:46 crc kubenswrapper[4824]: I0121 11:12:46.048835 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 11:12:46 crc kubenswrapper[4824]: E0121 11:12:46.048975 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9rmnw" podUID="8f9bd12b-22ee-44ed-a8b9-c100d043f691" Jan 21 11:12:46 crc kubenswrapper[4824]: E0121 11:12:46.049065 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 21 11:12:47 crc kubenswrapper[4824]: I0121 11:12:47.048843 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 11:12:47 crc kubenswrapper[4824]: I0121 11:12:47.048876 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 11:12:47 crc kubenswrapper[4824]: E0121 11:12:47.048994 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 21 11:12:47 crc kubenswrapper[4824]: E0121 11:12:47.049080 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 21 11:12:48 crc kubenswrapper[4824]: I0121 11:12:48.048507 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9rmnw" Jan 21 11:12:48 crc kubenswrapper[4824]: I0121 11:12:48.048509 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 11:12:48 crc kubenswrapper[4824]: E0121 11:12:48.049423 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9rmnw" podUID="8f9bd12b-22ee-44ed-a8b9-c100d043f691" Jan 21 11:12:48 crc kubenswrapper[4824]: E0121 11:12:48.049503 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 21 11:12:49 crc kubenswrapper[4824]: I0121 11:12:49.048704 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 11:12:49 crc kubenswrapper[4824]: E0121 11:12:49.048817 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 21 11:12:49 crc kubenswrapper[4824]: I0121 11:12:49.048709 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 11:12:49 crc kubenswrapper[4824]: E0121 11:12:49.049128 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 21 11:12:49 crc kubenswrapper[4824]: I0121 11:12:49.414163 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-x94nb_6b9588d1-5876-4c2c-8de5-e5c0ad46f04e/kube-multus/1.log" Jan 21 11:12:49 crc kubenswrapper[4824]: I0121 11:12:49.414651 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-x94nb_6b9588d1-5876-4c2c-8de5-e5c0ad46f04e/kube-multus/0.log" Jan 21 11:12:49 crc kubenswrapper[4824]: I0121 11:12:49.414683 4824 generic.go:334] "Generic (PLEG): container finished" podID="6b9588d1-5876-4c2c-8de5-e5c0ad46f04e" containerID="49ce055ec8090d6957cb7e23fe75edb82c0c2d7f87e7203d280606368dd3179f" exitCode=1 Jan 21 11:12:49 crc kubenswrapper[4824]: I0121 11:12:49.414708 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-x94nb" event={"ID":"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e","Type":"ContainerDied","Data":"49ce055ec8090d6957cb7e23fe75edb82c0c2d7f87e7203d280606368dd3179f"} Jan 21 11:12:49 crc kubenswrapper[4824]: I0121 11:12:49.414744 4824 scope.go:117] "RemoveContainer" containerID="75d52224135b75133fc000611cd179240b9174f6492d3263341716c7e42ccb15" Jan 21 11:12:49 crc kubenswrapper[4824]: I0121 11:12:49.415695 4824 scope.go:117] "RemoveContainer" containerID="49ce055ec8090d6957cb7e23fe75edb82c0c2d7f87e7203d280606368dd3179f" Jan 21 11:12:49 crc kubenswrapper[4824]: E0121 11:12:49.416031 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-x94nb_openshift-multus(6b9588d1-5876-4c2c-8de5-e5c0ad46f04e)\"" pod="openshift-multus/multus-x94nb" podUID="6b9588d1-5876-4c2c-8de5-e5c0ad46f04e" Jan 21 11:12:49 crc kubenswrapper[4824]: I0121 11:12:49.428422 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-xnwp2" podStartSLOduration=95.428394389 podStartE2EDuration="1m35.428394389s" podCreationTimestamp="2026-01-21 11:11:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:12:40.403405701 +0000 UTC m=+102.696434993" watchObservedRunningTime="2026-01-21 11:12:49.428394389 +0000 UTC m=+111.721423681" Jan 21 11:12:50 crc kubenswrapper[4824]: I0121 11:12:50.049240 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9rmnw" Jan 21 11:12:50 crc kubenswrapper[4824]: I0121 11:12:50.049300 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 11:12:50 crc kubenswrapper[4824]: E0121 11:12:50.049373 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9rmnw" podUID="8f9bd12b-22ee-44ed-a8b9-c100d043f691" Jan 21 11:12:50 crc kubenswrapper[4824]: E0121 11:12:50.049476 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 21 11:12:50 crc kubenswrapper[4824]: I0121 11:12:50.417729 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-x94nb_6b9588d1-5876-4c2c-8de5-e5c0ad46f04e/kube-multus/1.log" Jan 21 11:12:51 crc kubenswrapper[4824]: I0121 11:12:51.048525 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 11:12:51 crc kubenswrapper[4824]: I0121 11:12:51.048570 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 11:12:51 crc kubenswrapper[4824]: E0121 11:12:51.048655 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 21 11:12:51 crc kubenswrapper[4824]: E0121 11:12:51.048733 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 21 11:12:52 crc kubenswrapper[4824]: I0121 11:12:52.048905 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9rmnw" Jan 21 11:12:52 crc kubenswrapper[4824]: I0121 11:12:52.049056 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 11:12:52 crc kubenswrapper[4824]: E0121 11:12:52.049685 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9rmnw" podUID="8f9bd12b-22ee-44ed-a8b9-c100d043f691" Jan 21 11:12:52 crc kubenswrapper[4824]: E0121 11:12:52.049821 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 21 11:12:53 crc kubenswrapper[4824]: I0121 11:12:53.049229 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 11:12:53 crc kubenswrapper[4824]: E0121 11:12:53.049360 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 21 11:12:53 crc kubenswrapper[4824]: I0121 11:12:53.049236 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 11:12:53 crc kubenswrapper[4824]: E0121 11:12:53.049545 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 21 11:12:54 crc kubenswrapper[4824]: I0121 11:12:54.049131 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9rmnw" Jan 21 11:12:54 crc kubenswrapper[4824]: I0121 11:12:54.049182 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 11:12:54 crc kubenswrapper[4824]: E0121 11:12:54.049279 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9rmnw" podUID="8f9bd12b-22ee-44ed-a8b9-c100d043f691" Jan 21 11:12:54 crc kubenswrapper[4824]: E0121 11:12:54.049416 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 21 11:12:55 crc kubenswrapper[4824]: I0121 11:12:55.048765 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 11:12:55 crc kubenswrapper[4824]: E0121 11:12:55.048860 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 21 11:12:55 crc kubenswrapper[4824]: I0121 11:12:55.048899 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 11:12:55 crc kubenswrapper[4824]: E0121 11:12:55.049218 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 21 11:12:56 crc kubenswrapper[4824]: I0121 11:12:56.048992 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 11:12:56 crc kubenswrapper[4824]: I0121 11:12:56.049025 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9rmnw" Jan 21 11:12:56 crc kubenswrapper[4824]: E0121 11:12:56.049112 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 21 11:12:56 crc kubenswrapper[4824]: E0121 11:12:56.049177 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9rmnw" podUID="8f9bd12b-22ee-44ed-a8b9-c100d043f691" Jan 21 11:12:56 crc kubenswrapper[4824]: I0121 11:12:56.049630 4824 scope.go:117] "RemoveContainer" containerID="86a6881f109dab1f910fe505a21c7c800e587de4dded2635147997c51c05c483" Jan 21 11:12:56 crc kubenswrapper[4824]: I0121 11:12:56.432315 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-htkvk_26b1108a-4b98-4e39-a3cd-e0d055089fd8/ovnkube-controller/3.log" Jan 21 11:12:56 crc kubenswrapper[4824]: I0121 11:12:56.434246 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" event={"ID":"26b1108a-4b98-4e39-a3cd-e0d055089fd8","Type":"ContainerStarted","Data":"99d1fdcb6aaacd8685fdad0e8e6e667a311f27767270fdc0f22419bc3b8af38a"} Jan 21 11:12:56 crc kubenswrapper[4824]: I0121 11:12:56.434587 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" Jan 21 11:12:56 crc kubenswrapper[4824]: I0121 11:12:56.641555 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" podStartSLOduration=102.641537931 podStartE2EDuration="1m42.641537931s" podCreationTimestamp="2026-01-21 11:11:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:12:56.456586283 +0000 UTC m=+118.749615575" watchObservedRunningTime="2026-01-21 11:12:56.641537931 +0000 UTC m=+118.934567223" Jan 21 11:12:56 crc kubenswrapper[4824]: I0121 11:12:56.641934 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-9rmnw"] Jan 21 11:12:56 crc kubenswrapper[4824]: I0121 11:12:56.642049 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9rmnw" Jan 21 11:12:56 crc kubenswrapper[4824]: E0121 11:12:56.642147 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9rmnw" podUID="8f9bd12b-22ee-44ed-a8b9-c100d043f691" Jan 21 11:12:57 crc kubenswrapper[4824]: I0121 11:12:57.048569 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 11:12:57 crc kubenswrapper[4824]: E0121 11:12:57.048860 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 21 11:12:57 crc kubenswrapper[4824]: I0121 11:12:57.048569 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 11:12:57 crc kubenswrapper[4824]: E0121 11:12:57.048931 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 21 11:12:58 crc kubenswrapper[4824]: E0121 11:12:58.028734 4824 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Jan 21 11:12:58 crc kubenswrapper[4824]: I0121 11:12:58.048558 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 11:12:58 crc kubenswrapper[4824]: E0121 11:12:58.050943 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 21 11:12:58 crc kubenswrapper[4824]: I0121 11:12:58.051009 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9rmnw" Jan 21 11:12:58 crc kubenswrapper[4824]: E0121 11:12:58.051084 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9rmnw" podUID="8f9bd12b-22ee-44ed-a8b9-c100d043f691" Jan 21 11:12:58 crc kubenswrapper[4824]: E0121 11:12:58.096949 4824 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Jan 21 11:12:59 crc kubenswrapper[4824]: I0121 11:12:59.048446 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 11:12:59 crc kubenswrapper[4824]: I0121 11:12:59.048495 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 11:12:59 crc kubenswrapper[4824]: E0121 11:12:59.048562 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 21 11:12:59 crc kubenswrapper[4824]: E0121 11:12:59.048700 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 21 11:13:00 crc kubenswrapper[4824]: I0121 11:13:00.049011 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9rmnw" Jan 21 11:13:00 crc kubenswrapper[4824]: I0121 11:13:00.049015 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 11:13:00 crc kubenswrapper[4824]: E0121 11:13:00.049135 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9rmnw" podUID="8f9bd12b-22ee-44ed-a8b9-c100d043f691" Jan 21 11:13:00 crc kubenswrapper[4824]: E0121 11:13:00.049327 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 21 11:13:00 crc kubenswrapper[4824]: I0121 11:13:00.049685 4824 scope.go:117] "RemoveContainer" containerID="49ce055ec8090d6957cb7e23fe75edb82c0c2d7f87e7203d280606368dd3179f" Jan 21 11:13:00 crc kubenswrapper[4824]: I0121 11:13:00.446741 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-x94nb_6b9588d1-5876-4c2c-8de5-e5c0ad46f04e/kube-multus/1.log" Jan 21 11:13:00 crc kubenswrapper[4824]: I0121 11:13:00.446793 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-x94nb" event={"ID":"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e","Type":"ContainerStarted","Data":"80258959697dd4e8883d646587a5ac979419b23c4e94eace61a8b23e0885b470"} Jan 21 11:13:01 crc kubenswrapper[4824]: I0121 11:13:01.049224 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 11:13:01 crc kubenswrapper[4824]: I0121 11:13:01.049257 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 11:13:01 crc kubenswrapper[4824]: E0121 11:13:01.049365 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 21 11:13:01 crc kubenswrapper[4824]: E0121 11:13:01.049456 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 21 11:13:02 crc kubenswrapper[4824]: I0121 11:13:02.048933 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9rmnw" Jan 21 11:13:02 crc kubenswrapper[4824]: I0121 11:13:02.048988 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 11:13:02 crc kubenswrapper[4824]: E0121 11:13:02.049120 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9rmnw" podUID="8f9bd12b-22ee-44ed-a8b9-c100d043f691" Jan 21 11:13:02 crc kubenswrapper[4824]: E0121 11:13:02.049210 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 21 11:13:03 crc kubenswrapper[4824]: I0121 11:13:03.048405 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 11:13:03 crc kubenswrapper[4824]: I0121 11:13:03.048415 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 11:13:03 crc kubenswrapper[4824]: E0121 11:13:03.048523 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 21 11:13:03 crc kubenswrapper[4824]: E0121 11:13:03.048620 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 21 11:13:04 crc kubenswrapper[4824]: I0121 11:13:04.049090 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9rmnw" Jan 21 11:13:04 crc kubenswrapper[4824]: I0121 11:13:04.049136 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 11:13:04 crc kubenswrapper[4824]: I0121 11:13:04.051449 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Jan 21 11:13:04 crc kubenswrapper[4824]: I0121 11:13:04.051634 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Jan 21 11:13:04 crc kubenswrapper[4824]: I0121 11:13:04.051709 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Jan 21 11:13:04 crc kubenswrapper[4824]: I0121 11:13:04.052247 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Jan 21 11:13:05 crc kubenswrapper[4824]: I0121 11:13:05.048200 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 11:13:05 crc kubenswrapper[4824]: I0121 11:13:05.048260 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 11:13:05 crc kubenswrapper[4824]: I0121 11:13:05.049843 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Jan 21 11:13:05 crc kubenswrapper[4824]: I0121 11:13:05.050009 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.709469 4824 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.731932 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-htg7q"] Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.732567 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-htg7q" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.732731 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-4dkd8"] Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.733037 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-4dkd8" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.737816 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.737949 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.738092 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.738202 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.738239 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.738376 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.738488 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.738499 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.738597 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-7grz2"] Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.738890 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.739151 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.739339 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-7grz2" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.739338 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.739504 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.739610 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.739871 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.739902 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.740289 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.740379 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-s2n22"] Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.747225 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-s2n22" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.747744 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.747864 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.748029 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-qx94f"] Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.748429 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-qx94f" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.748739 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-fbq6l"] Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.749113 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-fbq6l" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.750185 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-nwlwd"] Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.750379 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.750548 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-mh6ds"] Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.750828 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-mh6ds" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.750985 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.751108 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-nwlwd" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.751108 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-7s7wx"] Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.751545 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.751827 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7s7wx" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.752099 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-hzhhs"] Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.752505 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-hzhhs" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.753623 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-ct4lp"] Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.754005 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-ct4lp" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.754234 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.754317 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-fz7gx"] Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.754388 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.754462 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.754530 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.754612 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-fz7gx" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.754652 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.754725 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.754798 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.754930 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.755080 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.755415 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-f6c7b"] Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.755769 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-f6c7b" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.756326 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.756367 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.756640 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.757119 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2tjlq"] Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.757635 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2tjlq" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.767728 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.767837 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.780607 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.781174 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.781792 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.781948 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.782167 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.782307 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.782456 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.782589 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.782704 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.782823 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.783033 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.783171 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.783348 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.783450 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.783558 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.783706 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.783849 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.786192 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.786480 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.786504 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.786758 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.786818 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.786930 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.787013 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.787103 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.787262 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.787505 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.787608 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.787967 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.787985 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.788139 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.788162 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.788613 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.790200 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.790329 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.790340 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.790360 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.790475 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.790474 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.790557 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.790526 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.790617 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.790673 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.790525 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.790571 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.790766 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.796272 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.796437 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.796546 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.796641 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.796689 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.796733 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.796821 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.797268 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.797986 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.799468 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.799583 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-8g4qk"] Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.800042 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-8g4qk" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.800670 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-cnr8w"] Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.801145 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-cnr8w" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.802362 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-wxlxx"] Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.802733 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-wxlxx" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.803230 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.803401 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.804183 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.806666 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.806886 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-t2thx"] Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.807231 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-cpgg4"] Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.807481 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-j8dcx"] Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.807574 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.807710 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-t2thx" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.807753 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-j8dcx" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.807987 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.808064 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/fb8bdb2b-bdc7-4433-be10-296fc67dc1c3-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-7grz2\" (UID: \"fb8bdb2b-bdc7-4433-be10-296fc67dc1c3\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-7grz2" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.808083 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-cpgg4" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.808091 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/8364adda-37b5-4a9d-8abd-3daf12470f20-etcd-serving-ca\") pod \"apiserver-76f77b778f-htg7q\" (UID: \"8364adda-37b5-4a9d-8abd-3daf12470f20\") " pod="openshift-apiserver/apiserver-76f77b778f-htg7q" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.808109 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8364adda-37b5-4a9d-8abd-3daf12470f20-trusted-ca-bundle\") pod \"apiserver-76f77b778f-htg7q\" (UID: \"8364adda-37b5-4a9d-8abd-3daf12470f20\") " pod="openshift-apiserver/apiserver-76f77b778f-htg7q" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.808125 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/0475fd50-10a8-4195-8643-9408478e953c-encryption-config\") pod \"apiserver-7bbb656c7d-7s7wx\" (UID: \"0475fd50-10a8-4195-8643-9408478e953c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7s7wx" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.808141 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5zq9z\" (UniqueName: \"kubernetes.io/projected/fb8bdb2b-bdc7-4433-be10-296fc67dc1c3-kube-api-access-5zq9z\") pod \"machine-api-operator-5694c8668f-7grz2\" (UID: \"fb8bdb2b-bdc7-4433-be10-296fc67dc1c3\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-7grz2" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.808166 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/49ec79bd-9eb8-42c8-a262-e96c1457f686-config\") pod \"openshift-apiserver-operator-796bbdcf4f-qx94f\" (UID: \"49ec79bd-9eb8-42c8-a262-e96c1457f686\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-qx94f" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.808182 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dg7z5\" (UniqueName: \"kubernetes.io/projected/d962d0be-438a-4d08-8113-525238fa2156-kube-api-access-dg7z5\") pod \"controller-manager-879f6c89f-4dkd8\" (UID: \"d962d0be-438a-4d08-8113-525238fa2156\") " pod="openshift-controller-manager/controller-manager-879f6c89f-4dkd8" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.808196 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0475fd50-10a8-4195-8643-9408478e953c-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-7s7wx\" (UID: \"0475fd50-10a8-4195-8643-9408478e953c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7s7wx" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.808222 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-klfwh\" (UniqueName: \"kubernetes.io/projected/0475fd50-10a8-4195-8643-9408478e953c-kube-api-access-klfwh\") pod \"apiserver-7bbb656c7d-7s7wx\" (UID: \"0475fd50-10a8-4195-8643-9408478e953c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7s7wx" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.808249 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q8qjw\" (UniqueName: \"kubernetes.io/projected/c585ef2f-3e29-4b81-918f-11b075006ca0-kube-api-access-q8qjw\") pod \"console-f9d7485db-nwlwd\" (UID: \"c585ef2f-3e29-4b81-918f-11b075006ca0\") " pod="openshift-console/console-f9d7485db-nwlwd" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.808266 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/c585ef2f-3e29-4b81-918f-11b075006ca0-console-oauth-config\") pod \"console-f9d7485db-nwlwd\" (UID: \"c585ef2f-3e29-4b81-918f-11b075006ca0\") " pod="openshift-console/console-f9d7485db-nwlwd" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.808280 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/8364adda-37b5-4a9d-8abd-3daf12470f20-etcd-client\") pod \"apiserver-76f77b778f-htg7q\" (UID: \"8364adda-37b5-4a9d-8abd-3daf12470f20\") " pod="openshift-apiserver/apiserver-76f77b778f-htg7q" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.808296 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c585ef2f-3e29-4b81-918f-11b075006ca0-trusted-ca-bundle\") pod \"console-f9d7485db-nwlwd\" (UID: \"c585ef2f-3e29-4b81-918f-11b075006ca0\") " pod="openshift-console/console-f9d7485db-nwlwd" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.808310 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/8364adda-37b5-4a9d-8abd-3daf12470f20-encryption-config\") pod \"apiserver-76f77b778f-htg7q\" (UID: \"8364adda-37b5-4a9d-8abd-3daf12470f20\") " pod="openshift-apiserver/apiserver-76f77b778f-htg7q" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.808332 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8364adda-37b5-4a9d-8abd-3daf12470f20-config\") pod \"apiserver-76f77b778f-htg7q\" (UID: \"8364adda-37b5-4a9d-8abd-3daf12470f20\") " pod="openshift-apiserver/apiserver-76f77b778f-htg7q" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.808348 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gh2qf\" (UniqueName: \"kubernetes.io/projected/c4b9a738-6a66-467d-b8a5-bddb20cb7e98-kube-api-access-gh2qf\") pod \"authentication-operator-69f744f599-s2n22\" (UID: \"c4b9a738-6a66-467d-b8a5-bddb20cb7e98\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-s2n22" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.808393 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/8364adda-37b5-4a9d-8abd-3daf12470f20-audit\") pod \"apiserver-76f77b778f-htg7q\" (UID: \"8364adda-37b5-4a9d-8abd-3daf12470f20\") " pod="openshift-apiserver/apiserver-76f77b778f-htg7q" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.808431 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8364adda-37b5-4a9d-8abd-3daf12470f20-serving-cert\") pod \"apiserver-76f77b778f-htg7q\" (UID: \"8364adda-37b5-4a9d-8abd-3daf12470f20\") " pod="openshift-apiserver/apiserver-76f77b778f-htg7q" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.808452 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fb8bdb2b-bdc7-4433-be10-296fc67dc1c3-config\") pod \"machine-api-operator-5694c8668f-7grz2\" (UID: \"fb8bdb2b-bdc7-4433-be10-296fc67dc1c3\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-7grz2" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.808468 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/29d5a4fe-fd42-42ec-85b4-c79af6424296-config\") pod \"console-operator-58897d9998-mh6ds\" (UID: \"29d5a4fe-fd42-42ec-85b4-c79af6424296\") " pod="openshift-console-operator/console-operator-58897d9998-mh6ds" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.808480 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-7fl6h"] Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.808501 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c4b9a738-6a66-467d-b8a5-bddb20cb7e98-service-ca-bundle\") pod \"authentication-operator-69f744f599-s2n22\" (UID: \"c4b9a738-6a66-467d-b8a5-bddb20cb7e98\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-s2n22" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.808517 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/8d0d706a-5fe5-402d-9040-3c763e525554-available-featuregates\") pod \"openshift-config-operator-7777fb866f-fbq6l\" (UID: \"8d0d706a-5fe5-402d-9040-3c763e525554\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-fbq6l" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.808532 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/29d5a4fe-fd42-42ec-85b4-c79af6424296-trusted-ca\") pod \"console-operator-58897d9998-mh6ds\" (UID: \"29d5a4fe-fd42-42ec-85b4-c79af6424296\") " pod="openshift-console-operator/console-operator-58897d9998-mh6ds" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.808556 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/fb8bdb2b-bdc7-4433-be10-296fc67dc1c3-images\") pod \"machine-api-operator-5694c8668f-7grz2\" (UID: \"fb8bdb2b-bdc7-4433-be10-296fc67dc1c3\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-7grz2" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.808584 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/0475fd50-10a8-4195-8643-9408478e953c-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-7s7wx\" (UID: \"0475fd50-10a8-4195-8643-9408478e953c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7s7wx" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.808605 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/8364adda-37b5-4a9d-8abd-3daf12470f20-audit-dir\") pod \"apiserver-76f77b778f-htg7q\" (UID: \"8364adda-37b5-4a9d-8abd-3daf12470f20\") " pod="openshift-apiserver/apiserver-76f77b778f-htg7q" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.808623 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d962d0be-438a-4d08-8113-525238fa2156-config\") pod \"controller-manager-879f6c89f-4dkd8\" (UID: \"d962d0be-438a-4d08-8113-525238fa2156\") " pod="openshift-controller-manager/controller-manager-879f6c89f-4dkd8" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.808638 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-skznz\" (UniqueName: \"kubernetes.io/projected/49ec79bd-9eb8-42c8-a262-e96c1457f686-kube-api-access-skznz\") pod \"openshift-apiserver-operator-796bbdcf4f-qx94f\" (UID: \"49ec79bd-9eb8-42c8-a262-e96c1457f686\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-qx94f" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.808657 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c4b9a738-6a66-467d-b8a5-bddb20cb7e98-serving-cert\") pod \"authentication-operator-69f744f599-s2n22\" (UID: \"c4b9a738-6a66-467d-b8a5-bddb20cb7e98\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-s2n22" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.808673 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8d0d706a-5fe5-402d-9040-3c763e525554-serving-cert\") pod \"openshift-config-operator-7777fb866f-fbq6l\" (UID: \"8d0d706a-5fe5-402d-9040-3c763e525554\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-fbq6l" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.808689 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/8364adda-37b5-4a9d-8abd-3daf12470f20-node-pullsecrets\") pod \"apiserver-76f77b778f-htg7q\" (UID: \"8364adda-37b5-4a9d-8abd-3daf12470f20\") " pod="openshift-apiserver/apiserver-76f77b778f-htg7q" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.808703 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d962d0be-438a-4d08-8113-525238fa2156-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-4dkd8\" (UID: \"d962d0be-438a-4d08-8113-525238fa2156\") " pod="openshift-controller-manager/controller-manager-879f6c89f-4dkd8" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.808717 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rl46h\" (UniqueName: \"kubernetes.io/projected/29d5a4fe-fd42-42ec-85b4-c79af6424296-kube-api-access-rl46h\") pod \"console-operator-58897d9998-mh6ds\" (UID: \"29d5a4fe-fd42-42ec-85b4-c79af6424296\") " pod="openshift-console-operator/console-operator-58897d9998-mh6ds" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.808737 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/c585ef2f-3e29-4b81-918f-11b075006ca0-console-config\") pod \"console-f9d7485db-nwlwd\" (UID: \"c585ef2f-3e29-4b81-918f-11b075006ca0\") " pod="openshift-console/console-f9d7485db-nwlwd" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.808746 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-7fl6h" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.808752 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/0475fd50-10a8-4195-8643-9408478e953c-audit-policies\") pod \"apiserver-7bbb656c7d-7s7wx\" (UID: \"0475fd50-10a8-4195-8643-9408478e953c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7s7wx" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.808765 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/29d5a4fe-fd42-42ec-85b4-c79af6424296-serving-cert\") pod \"console-operator-58897d9998-mh6ds\" (UID: \"29d5a4fe-fd42-42ec-85b4-c79af6424296\") " pod="openshift-console-operator/console-operator-58897d9998-mh6ds" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.808780 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/c585ef2f-3e29-4b81-918f-11b075006ca0-console-serving-cert\") pod \"console-f9d7485db-nwlwd\" (UID: \"c585ef2f-3e29-4b81-918f-11b075006ca0\") " pod="openshift-console/console-f9d7485db-nwlwd" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.808794 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d962d0be-438a-4d08-8113-525238fa2156-serving-cert\") pod \"controller-manager-879f6c89f-4dkd8\" (UID: \"d962d0be-438a-4d08-8113-525238fa2156\") " pod="openshift-controller-manager/controller-manager-879f6c89f-4dkd8" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.808808 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/0475fd50-10a8-4195-8643-9408478e953c-audit-dir\") pod \"apiserver-7bbb656c7d-7s7wx\" (UID: \"0475fd50-10a8-4195-8643-9408478e953c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7s7wx" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.808823 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c4b9a738-6a66-467d-b8a5-bddb20cb7e98-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-s2n22\" (UID: \"c4b9a738-6a66-467d-b8a5-bddb20cb7e98\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-s2n22" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.808838 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c4b9a738-6a66-467d-b8a5-bddb20cb7e98-config\") pod \"authentication-operator-69f744f599-s2n22\" (UID: \"c4b9a738-6a66-467d-b8a5-bddb20cb7e98\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-s2n22" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.808863 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/49ec79bd-9eb8-42c8-a262-e96c1457f686-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-qx94f\" (UID: \"49ec79bd-9eb8-42c8-a262-e96c1457f686\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-qx94f" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.808876 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d962d0be-438a-4d08-8113-525238fa2156-client-ca\") pod \"controller-manager-879f6c89f-4dkd8\" (UID: \"d962d0be-438a-4d08-8113-525238fa2156\") " pod="openshift-controller-manager/controller-manager-879f6c89f-4dkd8" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.808891 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/8364adda-37b5-4a9d-8abd-3daf12470f20-image-import-ca\") pod \"apiserver-76f77b778f-htg7q\" (UID: \"8364adda-37b5-4a9d-8abd-3daf12470f20\") " pod="openshift-apiserver/apiserver-76f77b778f-htg7q" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.808905 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/0475fd50-10a8-4195-8643-9408478e953c-etcd-client\") pod \"apiserver-7bbb656c7d-7s7wx\" (UID: \"0475fd50-10a8-4195-8643-9408478e953c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7s7wx" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.808920 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fm7v7\" (UniqueName: \"kubernetes.io/projected/8364adda-37b5-4a9d-8abd-3daf12470f20-kube-api-access-fm7v7\") pod \"apiserver-76f77b778f-htg7q\" (UID: \"8364adda-37b5-4a9d-8abd-3daf12470f20\") " pod="openshift-apiserver/apiserver-76f77b778f-htg7q" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.808933 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0475fd50-10a8-4195-8643-9408478e953c-serving-cert\") pod \"apiserver-7bbb656c7d-7s7wx\" (UID: \"0475fd50-10a8-4195-8643-9408478e953c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7s7wx" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.808946 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n6nx5\" (UniqueName: \"kubernetes.io/projected/8d0d706a-5fe5-402d-9040-3c763e525554-kube-api-access-n6nx5\") pod \"openshift-config-operator-7777fb866f-fbq6l\" (UID: \"8d0d706a-5fe5-402d-9040-3c763e525554\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-fbq6l" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.808982 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/c585ef2f-3e29-4b81-918f-11b075006ca0-oauth-serving-cert\") pod \"console-f9d7485db-nwlwd\" (UID: \"c585ef2f-3e29-4b81-918f-11b075006ca0\") " pod="openshift-console/console-f9d7485db-nwlwd" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.808999 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/c585ef2f-3e29-4b81-918f-11b075006ca0-service-ca\") pod \"console-f9d7485db-nwlwd\" (UID: \"c585ef2f-3e29-4b81-918f-11b075006ca0\") " pod="openshift-console/console-f9d7485db-nwlwd" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.809248 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-cq5sn"] Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.809672 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-cq5sn" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.813317 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.813528 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.813653 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.815070 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-gjhp5"] Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.815479 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-rdcbg"] Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.815693 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-gjhp5" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.815787 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-rdcbg" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.816256 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.816333 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29483220-6r4fz"] Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.816744 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29483220-6r4fz" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.817599 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-bkkzh"] Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.818032 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-n8lql"] Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.818189 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-bkkzh" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.818414 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-n8lql" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.818648 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-qs8d6"] Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.819042 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qs8d6" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.820488 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5lmnb"] Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.821331 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5lmnb" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.821437 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-27d2q"] Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.822054 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-27d2q" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.824174 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-r65dm"] Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.824800 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-r65dm" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.827622 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-j6bst"] Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.835472 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-j6bst" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.836538 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-c5xks"] Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.837715 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-c5xks" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.838677 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.843087 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-s7wn9"] Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.844873 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-s7wn9" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.850886 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-bblkp"] Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.851948 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-bblkp" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.852514 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-9tldv"] Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.853185 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-9tldv" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.853637 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-kqtr5"] Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.854310 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kqtr5" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.854641 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.856304 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-cslqn"] Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.856715 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-cslqn" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.857361 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-7grz2"] Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.859187 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-qx94f"] Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.860431 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-4dkd8"] Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.861417 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-s2n22"] Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.863139 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-fbq6l"] Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.864336 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-htg7q"] Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.865191 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-7s7wx"] Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.865941 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-dlgmw"] Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.866805 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-5s88g"] Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.867036 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-dlgmw" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.867341 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-5s88g" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.867860 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-ct4lp"] Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.868482 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-cq5sn"] Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.869481 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-mh6ds"] Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.870541 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-c5xks"] Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.871394 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-nwlwd"] Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.872246 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-wxlxx"] Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.873230 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5lmnb"] Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.873732 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.874045 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-qs8d6"] Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.875034 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-j6bst"] Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.876236 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2tjlq"] Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.877289 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-j8dcx"] Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.878005 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-t2thx"] Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.879644 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-8g4qk"] Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.880488 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-cpgg4"] Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.881327 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-rdcbg"] Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.882149 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-7fl6h"] Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.883000 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-fz7gx"] Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.883797 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-r65dm"] Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.884594 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29483220-6r4fz"] Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.885494 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-bkkzh"] Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.886785 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-n8lql"] Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.887665 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-9tldv"] Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.888724 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-kqtr5"] Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.889592 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-s7wn9"] Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.890422 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-gjhp5"] Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.891266 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-cnr8w"] Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.892092 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-27d2q"] Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.892921 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-f6c7b"] Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.896840 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-jgvgl"] Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.897614 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.899687 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-dlgmw"] Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.899712 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-cslqn"] Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.899784 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-jgvgl" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.899981 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-jgvgl"] Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.910398 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8364adda-37b5-4a9d-8abd-3daf12470f20-serving-cert\") pod \"apiserver-76f77b778f-htg7q\" (UID: \"8364adda-37b5-4a9d-8abd-3daf12470f20\") " pod="openshift-apiserver/apiserver-76f77b778f-htg7q" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.910433 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fb8bdb2b-bdc7-4433-be10-296fc67dc1c3-config\") pod \"machine-api-operator-5694c8668f-7grz2\" (UID: \"fb8bdb2b-bdc7-4433-be10-296fc67dc1c3\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-7grz2" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.910475 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/29d5a4fe-fd42-42ec-85b4-c79af6424296-config\") pod \"console-operator-58897d9998-mh6ds\" (UID: \"29d5a4fe-fd42-42ec-85b4-c79af6424296\") " pod="openshift-console-operator/console-operator-58897d9998-mh6ds" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.910494 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c4b9a738-6a66-467d-b8a5-bddb20cb7e98-service-ca-bundle\") pod \"authentication-operator-69f744f599-s2n22\" (UID: \"c4b9a738-6a66-467d-b8a5-bddb20cb7e98\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-s2n22" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.911096 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/8d0d706a-5fe5-402d-9040-3c763e525554-available-featuregates\") pod \"openshift-config-operator-7777fb866f-fbq6l\" (UID: \"8d0d706a-5fe5-402d-9040-3c763e525554\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-fbq6l" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.911126 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/29d5a4fe-fd42-42ec-85b4-c79af6424296-trusted-ca\") pod \"console-operator-58897d9998-mh6ds\" (UID: \"29d5a4fe-fd42-42ec-85b4-c79af6424296\") " pod="openshift-console-operator/console-operator-58897d9998-mh6ds" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.911151 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/fb8bdb2b-bdc7-4433-be10-296fc67dc1c3-images\") pod \"machine-api-operator-5694c8668f-7grz2\" (UID: \"fb8bdb2b-bdc7-4433-be10-296fc67dc1c3\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-7grz2" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.911177 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/0475fd50-10a8-4195-8643-9408478e953c-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-7s7wx\" (UID: \"0475fd50-10a8-4195-8643-9408478e953c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7s7wx" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.911198 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-skznz\" (UniqueName: \"kubernetes.io/projected/49ec79bd-9eb8-42c8-a262-e96c1457f686-kube-api-access-skznz\") pod \"openshift-apiserver-operator-796bbdcf4f-qx94f\" (UID: \"49ec79bd-9eb8-42c8-a262-e96c1457f686\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-qx94f" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.911226 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/8364adda-37b5-4a9d-8abd-3daf12470f20-audit-dir\") pod \"apiserver-76f77b778f-htg7q\" (UID: \"8364adda-37b5-4a9d-8abd-3daf12470f20\") " pod="openshift-apiserver/apiserver-76f77b778f-htg7q" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.911243 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d962d0be-438a-4d08-8113-525238fa2156-config\") pod \"controller-manager-879f6c89f-4dkd8\" (UID: \"d962d0be-438a-4d08-8113-525238fa2156\") " pod="openshift-controller-manager/controller-manager-879f6c89f-4dkd8" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.911265 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c4b9a738-6a66-467d-b8a5-bddb20cb7e98-serving-cert\") pod \"authentication-operator-69f744f599-s2n22\" (UID: \"c4b9a738-6a66-467d-b8a5-bddb20cb7e98\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-s2n22" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.911284 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8d0d706a-5fe5-402d-9040-3c763e525554-serving-cert\") pod \"openshift-config-operator-7777fb866f-fbq6l\" (UID: \"8d0d706a-5fe5-402d-9040-3c763e525554\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-fbq6l" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.911304 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/8364adda-37b5-4a9d-8abd-3daf12470f20-node-pullsecrets\") pod \"apiserver-76f77b778f-htg7q\" (UID: \"8364adda-37b5-4a9d-8abd-3daf12470f20\") " pod="openshift-apiserver/apiserver-76f77b778f-htg7q" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.911321 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d962d0be-438a-4d08-8113-525238fa2156-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-4dkd8\" (UID: \"d962d0be-438a-4d08-8113-525238fa2156\") " pod="openshift-controller-manager/controller-manager-879f6c89f-4dkd8" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.911348 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rl46h\" (UniqueName: \"kubernetes.io/projected/29d5a4fe-fd42-42ec-85b4-c79af6424296-kube-api-access-rl46h\") pod \"console-operator-58897d9998-mh6ds\" (UID: \"29d5a4fe-fd42-42ec-85b4-c79af6424296\") " pod="openshift-console-operator/console-operator-58897d9998-mh6ds" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.911368 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/0475fd50-10a8-4195-8643-9408478e953c-audit-policies\") pod \"apiserver-7bbb656c7d-7s7wx\" (UID: \"0475fd50-10a8-4195-8643-9408478e953c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7s7wx" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.911385 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/29d5a4fe-fd42-42ec-85b4-c79af6424296-serving-cert\") pod \"console-operator-58897d9998-mh6ds\" (UID: \"29d5a4fe-fd42-42ec-85b4-c79af6424296\") " pod="openshift-console-operator/console-operator-58897d9998-mh6ds" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.911434 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fb8bdb2b-bdc7-4433-be10-296fc67dc1c3-config\") pod \"machine-api-operator-5694c8668f-7grz2\" (UID: \"fb8bdb2b-bdc7-4433-be10-296fc67dc1c3\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-7grz2" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.911501 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/c585ef2f-3e29-4b81-918f-11b075006ca0-console-config\") pod \"console-f9d7485db-nwlwd\" (UID: \"c585ef2f-3e29-4b81-918f-11b075006ca0\") " pod="openshift-console/console-f9d7485db-nwlwd" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.911519 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/c585ef2f-3e29-4b81-918f-11b075006ca0-console-serving-cert\") pod \"console-f9d7485db-nwlwd\" (UID: \"c585ef2f-3e29-4b81-918f-11b075006ca0\") " pod="openshift-console/console-f9d7485db-nwlwd" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.911538 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d962d0be-438a-4d08-8113-525238fa2156-serving-cert\") pod \"controller-manager-879f6c89f-4dkd8\" (UID: \"d962d0be-438a-4d08-8113-525238fa2156\") " pod="openshift-controller-manager/controller-manager-879f6c89f-4dkd8" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.911558 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c4b9a738-6a66-467d-b8a5-bddb20cb7e98-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-s2n22\" (UID: \"c4b9a738-6a66-467d-b8a5-bddb20cb7e98\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-s2n22" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.911573 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/0475fd50-10a8-4195-8643-9408478e953c-audit-dir\") pod \"apiserver-7bbb656c7d-7s7wx\" (UID: \"0475fd50-10a8-4195-8643-9408478e953c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7s7wx" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.911591 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c4b9a738-6a66-467d-b8a5-bddb20cb7e98-config\") pod \"authentication-operator-69f744f599-s2n22\" (UID: \"c4b9a738-6a66-467d-b8a5-bddb20cb7e98\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-s2n22" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.911618 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/49ec79bd-9eb8-42c8-a262-e96c1457f686-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-qx94f\" (UID: \"49ec79bd-9eb8-42c8-a262-e96c1457f686\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-qx94f" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.911633 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d962d0be-438a-4d08-8113-525238fa2156-client-ca\") pod \"controller-manager-879f6c89f-4dkd8\" (UID: \"d962d0be-438a-4d08-8113-525238fa2156\") " pod="openshift-controller-manager/controller-manager-879f6c89f-4dkd8" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.911651 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/8364adda-37b5-4a9d-8abd-3daf12470f20-image-import-ca\") pod \"apiserver-76f77b778f-htg7q\" (UID: \"8364adda-37b5-4a9d-8abd-3daf12470f20\") " pod="openshift-apiserver/apiserver-76f77b778f-htg7q" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.911667 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/0475fd50-10a8-4195-8643-9408478e953c-etcd-client\") pod \"apiserver-7bbb656c7d-7s7wx\" (UID: \"0475fd50-10a8-4195-8643-9408478e953c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7s7wx" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.911669 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/8d0d706a-5fe5-402d-9040-3c763e525554-available-featuregates\") pod \"openshift-config-operator-7777fb866f-fbq6l\" (UID: \"8d0d706a-5fe5-402d-9040-3c763e525554\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-fbq6l" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.911684 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/c585ef2f-3e29-4b81-918f-11b075006ca0-oauth-serving-cert\") pod \"console-f9d7485db-nwlwd\" (UID: \"c585ef2f-3e29-4b81-918f-11b075006ca0\") " pod="openshift-console/console-f9d7485db-nwlwd" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.911702 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fm7v7\" (UniqueName: \"kubernetes.io/projected/8364adda-37b5-4a9d-8abd-3daf12470f20-kube-api-access-fm7v7\") pod \"apiserver-76f77b778f-htg7q\" (UID: \"8364adda-37b5-4a9d-8abd-3daf12470f20\") " pod="openshift-apiserver/apiserver-76f77b778f-htg7q" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.911742 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/0475fd50-10a8-4195-8643-9408478e953c-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-7s7wx\" (UID: \"0475fd50-10a8-4195-8643-9408478e953c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7s7wx" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.911742 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0475fd50-10a8-4195-8643-9408478e953c-serving-cert\") pod \"apiserver-7bbb656c7d-7s7wx\" (UID: \"0475fd50-10a8-4195-8643-9408478e953c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7s7wx" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.911791 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n6nx5\" (UniqueName: \"kubernetes.io/projected/8d0d706a-5fe5-402d-9040-3c763e525554-kube-api-access-n6nx5\") pod \"openshift-config-operator-7777fb866f-fbq6l\" (UID: \"8d0d706a-5fe5-402d-9040-3c763e525554\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-fbq6l" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.911813 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/c585ef2f-3e29-4b81-918f-11b075006ca0-service-ca\") pod \"console-f9d7485db-nwlwd\" (UID: \"c585ef2f-3e29-4b81-918f-11b075006ca0\") " pod="openshift-console/console-f9d7485db-nwlwd" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.911830 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/8364adda-37b5-4a9d-8abd-3daf12470f20-etcd-serving-ca\") pod \"apiserver-76f77b778f-htg7q\" (UID: \"8364adda-37b5-4a9d-8abd-3daf12470f20\") " pod="openshift-apiserver/apiserver-76f77b778f-htg7q" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.911849 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/fb8bdb2b-bdc7-4433-be10-296fc67dc1c3-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-7grz2\" (UID: \"fb8bdb2b-bdc7-4433-be10-296fc67dc1c3\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-7grz2" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.911867 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/49ec79bd-9eb8-42c8-a262-e96c1457f686-config\") pod \"openshift-apiserver-operator-796bbdcf4f-qx94f\" (UID: \"49ec79bd-9eb8-42c8-a262-e96c1457f686\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-qx94f" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.911883 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8364adda-37b5-4a9d-8abd-3daf12470f20-trusted-ca-bundle\") pod \"apiserver-76f77b778f-htg7q\" (UID: \"8364adda-37b5-4a9d-8abd-3daf12470f20\") " pod="openshift-apiserver/apiserver-76f77b778f-htg7q" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.911897 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/0475fd50-10a8-4195-8643-9408478e953c-encryption-config\") pod \"apiserver-7bbb656c7d-7s7wx\" (UID: \"0475fd50-10a8-4195-8643-9408478e953c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7s7wx" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.911911 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5zq9z\" (UniqueName: \"kubernetes.io/projected/fb8bdb2b-bdc7-4433-be10-296fc67dc1c3-kube-api-access-5zq9z\") pod \"machine-api-operator-5694c8668f-7grz2\" (UID: \"fb8bdb2b-bdc7-4433-be10-296fc67dc1c3\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-7grz2" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.911930 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dg7z5\" (UniqueName: \"kubernetes.io/projected/d962d0be-438a-4d08-8113-525238fa2156-kube-api-access-dg7z5\") pod \"controller-manager-879f6c89f-4dkd8\" (UID: \"d962d0be-438a-4d08-8113-525238fa2156\") " pod="openshift-controller-manager/controller-manager-879f6c89f-4dkd8" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.911945 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0475fd50-10a8-4195-8643-9408478e953c-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-7s7wx\" (UID: \"0475fd50-10a8-4195-8643-9408478e953c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7s7wx" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.911977 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-klfwh\" (UniqueName: \"kubernetes.io/projected/0475fd50-10a8-4195-8643-9408478e953c-kube-api-access-klfwh\") pod \"apiserver-7bbb656c7d-7s7wx\" (UID: \"0475fd50-10a8-4195-8643-9408478e953c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7s7wx" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.912002 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q8qjw\" (UniqueName: \"kubernetes.io/projected/c585ef2f-3e29-4b81-918f-11b075006ca0-kube-api-access-q8qjw\") pod \"console-f9d7485db-nwlwd\" (UID: \"c585ef2f-3e29-4b81-918f-11b075006ca0\") " pod="openshift-console/console-f9d7485db-nwlwd" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.912014 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/8364adda-37b5-4a9d-8abd-3daf12470f20-node-pullsecrets\") pod \"apiserver-76f77b778f-htg7q\" (UID: \"8364adda-37b5-4a9d-8abd-3daf12470f20\") " pod="openshift-apiserver/apiserver-76f77b778f-htg7q" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.912021 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/c585ef2f-3e29-4b81-918f-11b075006ca0-console-oauth-config\") pod \"console-f9d7485db-nwlwd\" (UID: \"c585ef2f-3e29-4b81-918f-11b075006ca0\") " pod="openshift-console/console-f9d7485db-nwlwd" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.912059 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c585ef2f-3e29-4b81-918f-11b075006ca0-trusted-ca-bundle\") pod \"console-f9d7485db-nwlwd\" (UID: \"c585ef2f-3e29-4b81-918f-11b075006ca0\") " pod="openshift-console/console-f9d7485db-nwlwd" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.912079 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/8364adda-37b5-4a9d-8abd-3daf12470f20-etcd-client\") pod \"apiserver-76f77b778f-htg7q\" (UID: \"8364adda-37b5-4a9d-8abd-3daf12470f20\") " pod="openshift-apiserver/apiserver-76f77b778f-htg7q" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.912110 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8364adda-37b5-4a9d-8abd-3daf12470f20-config\") pod \"apiserver-76f77b778f-htg7q\" (UID: \"8364adda-37b5-4a9d-8abd-3daf12470f20\") " pod="openshift-apiserver/apiserver-76f77b778f-htg7q" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.912129 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/8364adda-37b5-4a9d-8abd-3daf12470f20-encryption-config\") pod \"apiserver-76f77b778f-htg7q\" (UID: \"8364adda-37b5-4a9d-8abd-3daf12470f20\") " pod="openshift-apiserver/apiserver-76f77b778f-htg7q" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.912152 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gh2qf\" (UniqueName: \"kubernetes.io/projected/c4b9a738-6a66-467d-b8a5-bddb20cb7e98-kube-api-access-gh2qf\") pod \"authentication-operator-69f744f599-s2n22\" (UID: \"c4b9a738-6a66-467d-b8a5-bddb20cb7e98\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-s2n22" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.912179 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/8364adda-37b5-4a9d-8abd-3daf12470f20-audit\") pod \"apiserver-76f77b778f-htg7q\" (UID: \"8364adda-37b5-4a9d-8abd-3daf12470f20\") " pod="openshift-apiserver/apiserver-76f77b778f-htg7q" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.912352 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/0475fd50-10a8-4195-8643-9408478e953c-audit-policies\") pod \"apiserver-7bbb656c7d-7s7wx\" (UID: \"0475fd50-10a8-4195-8643-9408478e953c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7s7wx" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.912381 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d962d0be-438a-4d08-8113-525238fa2156-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-4dkd8\" (UID: \"d962d0be-438a-4d08-8113-525238fa2156\") " pod="openshift-controller-manager/controller-manager-879f6c89f-4dkd8" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.911240 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/29d5a4fe-fd42-42ec-85b4-c79af6424296-config\") pod \"console-operator-58897d9998-mh6ds\" (UID: \"29d5a4fe-fd42-42ec-85b4-c79af6424296\") " pod="openshift-console-operator/console-operator-58897d9998-mh6ds" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.912755 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/29d5a4fe-fd42-42ec-85b4-c79af6424296-trusted-ca\") pod \"console-operator-58897d9998-mh6ds\" (UID: \"29d5a4fe-fd42-42ec-85b4-c79af6424296\") " pod="openshift-console-operator/console-operator-58897d9998-mh6ds" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.912944 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c4b9a738-6a66-467d-b8a5-bddb20cb7e98-service-ca-bundle\") pod \"authentication-operator-69f744f599-s2n22\" (UID: \"c4b9a738-6a66-467d-b8a5-bddb20cb7e98\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-s2n22" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.913194 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/8364adda-37b5-4a9d-8abd-3daf12470f20-audit\") pod \"apiserver-76f77b778f-htg7q\" (UID: \"8364adda-37b5-4a9d-8abd-3daf12470f20\") " pod="openshift-apiserver/apiserver-76f77b778f-htg7q" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.913216 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/8364adda-37b5-4a9d-8abd-3daf12470f20-audit-dir\") pod \"apiserver-76f77b778f-htg7q\" (UID: \"8364adda-37b5-4a9d-8abd-3daf12470f20\") " pod="openshift-apiserver/apiserver-76f77b778f-htg7q" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.913350 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/8364adda-37b5-4a9d-8abd-3daf12470f20-etcd-serving-ca\") pod \"apiserver-76f77b778f-htg7q\" (UID: \"8364adda-37b5-4a9d-8abd-3daf12470f20\") " pod="openshift-apiserver/apiserver-76f77b778f-htg7q" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.913659 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c4b9a738-6a66-467d-b8a5-bddb20cb7e98-config\") pod \"authentication-operator-69f744f599-s2n22\" (UID: \"c4b9a738-6a66-467d-b8a5-bddb20cb7e98\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-s2n22" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.913941 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/c585ef2f-3e29-4b81-918f-11b075006ca0-service-ca\") pod \"console-f9d7485db-nwlwd\" (UID: \"c585ef2f-3e29-4b81-918f-11b075006ca0\") " pod="openshift-console/console-f9d7485db-nwlwd" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.914357 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/c585ef2f-3e29-4b81-918f-11b075006ca0-console-config\") pod \"console-f9d7485db-nwlwd\" (UID: \"c585ef2f-3e29-4b81-918f-11b075006ca0\") " pod="openshift-console/console-f9d7485db-nwlwd" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.914571 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c585ef2f-3e29-4b81-918f-11b075006ca0-trusted-ca-bundle\") pod \"console-f9d7485db-nwlwd\" (UID: \"c585ef2f-3e29-4b81-918f-11b075006ca0\") " pod="openshift-console/console-f9d7485db-nwlwd" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.914652 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.914822 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/49ec79bd-9eb8-42c8-a262-e96c1457f686-config\") pod \"openshift-apiserver-operator-796bbdcf4f-qx94f\" (UID: \"49ec79bd-9eb8-42c8-a262-e96c1457f686\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-qx94f" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.915484 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d962d0be-438a-4d08-8113-525238fa2156-config\") pod \"controller-manager-879f6c89f-4dkd8\" (UID: \"d962d0be-438a-4d08-8113-525238fa2156\") " pod="openshift-controller-manager/controller-manager-879f6c89f-4dkd8" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.911935 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/fb8bdb2b-bdc7-4433-be10-296fc67dc1c3-images\") pod \"machine-api-operator-5694c8668f-7grz2\" (UID: \"fb8bdb2b-bdc7-4433-be10-296fc67dc1c3\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-7grz2" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.915695 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/8364adda-37b5-4a9d-8abd-3daf12470f20-image-import-ca\") pod \"apiserver-76f77b778f-htg7q\" (UID: \"8364adda-37b5-4a9d-8abd-3daf12470f20\") " pod="openshift-apiserver/apiserver-76f77b778f-htg7q" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.915730 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0475fd50-10a8-4195-8643-9408478e953c-serving-cert\") pod \"apiserver-7bbb656c7d-7s7wx\" (UID: \"0475fd50-10a8-4195-8643-9408478e953c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7s7wx" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.915782 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/0475fd50-10a8-4195-8643-9408478e953c-audit-dir\") pod \"apiserver-7bbb656c7d-7s7wx\" (UID: \"0475fd50-10a8-4195-8643-9408478e953c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7s7wx" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.915826 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8d0d706a-5fe5-402d-9040-3c763e525554-serving-cert\") pod \"openshift-config-operator-7777fb866f-fbq6l\" (UID: \"8d0d706a-5fe5-402d-9040-3c763e525554\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-fbq6l" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.916131 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0475fd50-10a8-4195-8643-9408478e953c-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-7s7wx\" (UID: \"0475fd50-10a8-4195-8643-9408478e953c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7s7wx" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.916383 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8364adda-37b5-4a9d-8abd-3daf12470f20-config\") pod \"apiserver-76f77b778f-htg7q\" (UID: \"8364adda-37b5-4a9d-8abd-3daf12470f20\") " pod="openshift-apiserver/apiserver-76f77b778f-htg7q" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.916432 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8364adda-37b5-4a9d-8abd-3daf12470f20-trusted-ca-bundle\") pod \"apiserver-76f77b778f-htg7q\" (UID: \"8364adda-37b5-4a9d-8abd-3daf12470f20\") " pod="openshift-apiserver/apiserver-76f77b778f-htg7q" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.916895 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/c585ef2f-3e29-4b81-918f-11b075006ca0-oauth-serving-cert\") pod \"console-f9d7485db-nwlwd\" (UID: \"c585ef2f-3e29-4b81-918f-11b075006ca0\") " pod="openshift-console/console-f9d7485db-nwlwd" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.917496 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/0475fd50-10a8-4195-8643-9408478e953c-encryption-config\") pod \"apiserver-7bbb656c7d-7s7wx\" (UID: \"0475fd50-10a8-4195-8643-9408478e953c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7s7wx" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.917591 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/8364adda-37b5-4a9d-8abd-3daf12470f20-etcd-client\") pod \"apiserver-76f77b778f-htg7q\" (UID: \"8364adda-37b5-4a9d-8abd-3daf12470f20\") " pod="openshift-apiserver/apiserver-76f77b778f-htg7q" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.917650 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/c585ef2f-3e29-4b81-918f-11b075006ca0-console-oauth-config\") pod \"console-f9d7485db-nwlwd\" (UID: \"c585ef2f-3e29-4b81-918f-11b075006ca0\") " pod="openshift-console/console-f9d7485db-nwlwd" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.917660 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/29d5a4fe-fd42-42ec-85b4-c79af6424296-serving-cert\") pod \"console-operator-58897d9998-mh6ds\" (UID: \"29d5a4fe-fd42-42ec-85b4-c79af6424296\") " pod="openshift-console-operator/console-operator-58897d9998-mh6ds" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.917757 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c4b9a738-6a66-467d-b8a5-bddb20cb7e98-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-s2n22\" (UID: \"c4b9a738-6a66-467d-b8a5-bddb20cb7e98\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-s2n22" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.917782 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/0475fd50-10a8-4195-8643-9408478e953c-etcd-client\") pod \"apiserver-7bbb656c7d-7s7wx\" (UID: \"0475fd50-10a8-4195-8643-9408478e953c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7s7wx" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.917884 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8364adda-37b5-4a9d-8abd-3daf12470f20-serving-cert\") pod \"apiserver-76f77b778f-htg7q\" (UID: \"8364adda-37b5-4a9d-8abd-3daf12470f20\") " pod="openshift-apiserver/apiserver-76f77b778f-htg7q" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.917834 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c4b9a738-6a66-467d-b8a5-bddb20cb7e98-serving-cert\") pod \"authentication-operator-69f744f599-s2n22\" (UID: \"c4b9a738-6a66-467d-b8a5-bddb20cb7e98\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-s2n22" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.918024 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d962d0be-438a-4d08-8113-525238fa2156-client-ca\") pod \"controller-manager-879f6c89f-4dkd8\" (UID: \"d962d0be-438a-4d08-8113-525238fa2156\") " pod="openshift-controller-manager/controller-manager-879f6c89f-4dkd8" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.918696 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/49ec79bd-9eb8-42c8-a262-e96c1457f686-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-qx94f\" (UID: \"49ec79bd-9eb8-42c8-a262-e96c1457f686\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-qx94f" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.919428 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/c585ef2f-3e29-4b81-918f-11b075006ca0-console-serving-cert\") pod \"console-f9d7485db-nwlwd\" (UID: \"c585ef2f-3e29-4b81-918f-11b075006ca0\") " pod="openshift-console/console-f9d7485db-nwlwd" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.919760 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d962d0be-438a-4d08-8113-525238fa2156-serving-cert\") pod \"controller-manager-879f6c89f-4dkd8\" (UID: \"d962d0be-438a-4d08-8113-525238fa2156\") " pod="openshift-controller-manager/controller-manager-879f6c89f-4dkd8" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.919894 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/8364adda-37b5-4a9d-8abd-3daf12470f20-encryption-config\") pod \"apiserver-76f77b778f-htg7q\" (UID: \"8364adda-37b5-4a9d-8abd-3daf12470f20\") " pod="openshift-apiserver/apiserver-76f77b778f-htg7q" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.932535 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/fb8bdb2b-bdc7-4433-be10-296fc67dc1c3-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-7grz2\" (UID: \"fb8bdb2b-bdc7-4433-be10-296fc67dc1c3\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-7grz2" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.945712 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-b7tsx"] Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.946260 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-b7tsx" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.951654 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-b7tsx"] Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.954010 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.974078 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Jan 21 11:13:09 crc kubenswrapper[4824]: I0121 11:13:09.994134 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Jan 21 11:13:10 crc kubenswrapper[4824]: I0121 11:13:10.014089 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Jan 21 11:13:10 crc kubenswrapper[4824]: I0121 11:13:10.035409 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Jan 21 11:13:10 crc kubenswrapper[4824]: I0121 11:13:10.054160 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Jan 21 11:13:10 crc kubenswrapper[4824]: I0121 11:13:10.075114 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Jan 21 11:13:10 crc kubenswrapper[4824]: I0121 11:13:10.095173 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Jan 21 11:13:10 crc kubenswrapper[4824]: I0121 11:13:10.114216 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Jan 21 11:13:10 crc kubenswrapper[4824]: I0121 11:13:10.134675 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Jan 21 11:13:10 crc kubenswrapper[4824]: I0121 11:13:10.154078 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Jan 21 11:13:10 crc kubenswrapper[4824]: I0121 11:13:10.174105 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Jan 21 11:13:10 crc kubenswrapper[4824]: I0121 11:13:10.194883 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Jan 21 11:13:10 crc kubenswrapper[4824]: I0121 11:13:10.214157 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Jan 21 11:13:10 crc kubenswrapper[4824]: I0121 11:13:10.234647 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Jan 21 11:13:10 crc kubenswrapper[4824]: I0121 11:13:10.254213 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Jan 21 11:13:10 crc kubenswrapper[4824]: I0121 11:13:10.274700 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Jan 21 11:13:10 crc kubenswrapper[4824]: I0121 11:13:10.294223 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Jan 21 11:13:10 crc kubenswrapper[4824]: I0121 11:13:10.314797 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Jan 21 11:13:10 crc kubenswrapper[4824]: I0121 11:13:10.334873 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Jan 21 11:13:10 crc kubenswrapper[4824]: I0121 11:13:10.354472 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Jan 21 11:13:10 crc kubenswrapper[4824]: I0121 11:13:10.374432 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Jan 21 11:13:10 crc kubenswrapper[4824]: I0121 11:13:10.394649 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Jan 21 11:13:10 crc kubenswrapper[4824]: I0121 11:13:10.414543 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Jan 21 11:13:10 crc kubenswrapper[4824]: I0121 11:13:10.435818 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Jan 21 11:13:10 crc kubenswrapper[4824]: I0121 11:13:10.454279 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Jan 21 11:13:10 crc kubenswrapper[4824]: I0121 11:13:10.474891 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Jan 21 11:13:10 crc kubenswrapper[4824]: I0121 11:13:10.504934 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Jan 21 11:13:10 crc kubenswrapper[4824]: I0121 11:13:10.514313 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Jan 21 11:13:10 crc kubenswrapper[4824]: I0121 11:13:10.533938 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Jan 21 11:13:10 crc kubenswrapper[4824]: I0121 11:13:10.554633 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Jan 21 11:13:10 crc kubenswrapper[4824]: I0121 11:13:10.574338 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Jan 21 11:13:10 crc kubenswrapper[4824]: I0121 11:13:10.594684 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Jan 21 11:13:10 crc kubenswrapper[4824]: I0121 11:13:10.614019 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Jan 21 11:13:10 crc kubenswrapper[4824]: I0121 11:13:10.634642 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Jan 21 11:13:10 crc kubenswrapper[4824]: I0121 11:13:10.653906 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Jan 21 11:13:10 crc kubenswrapper[4824]: I0121 11:13:10.674712 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Jan 21 11:13:10 crc kubenswrapper[4824]: I0121 11:13:10.699321 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Jan 21 11:13:10 crc kubenswrapper[4824]: I0121 11:13:10.714210 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Jan 21 11:13:10 crc kubenswrapper[4824]: I0121 11:13:10.734752 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Jan 21 11:13:10 crc kubenswrapper[4824]: I0121 11:13:10.755830 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Jan 21 11:13:10 crc kubenswrapper[4824]: I0121 11:13:10.774930 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Jan 21 11:13:10 crc kubenswrapper[4824]: I0121 11:13:10.794438 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Jan 21 11:13:10 crc kubenswrapper[4824]: I0121 11:13:10.814750 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Jan 21 11:13:10 crc kubenswrapper[4824]: I0121 11:13:10.833819 4824 request.go:700] Waited for 1.012475708s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/configmaps?fieldSelector=metadata.name%3Dopenshift-service-ca.crt&limit=500&resourceVersion=0 Jan 21 11:13:10 crc kubenswrapper[4824]: I0121 11:13:10.834781 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Jan 21 11:13:10 crc kubenswrapper[4824]: I0121 11:13:10.853861 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Jan 21 11:13:10 crc kubenswrapper[4824]: I0121 11:13:10.874430 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Jan 21 11:13:10 crc kubenswrapper[4824]: I0121 11:13:10.895060 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Jan 21 11:13:10 crc kubenswrapper[4824]: I0121 11:13:10.914525 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Jan 21 11:13:10 crc kubenswrapper[4824]: I0121 11:13:10.934131 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Jan 21 11:13:10 crc kubenswrapper[4824]: I0121 11:13:10.954468 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Jan 21 11:13:10 crc kubenswrapper[4824]: I0121 11:13:10.974319 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Jan 21 11:13:10 crc kubenswrapper[4824]: I0121 11:13:10.993897 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Jan 21 11:13:11 crc kubenswrapper[4824]: I0121 11:13:11.013941 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Jan 21 11:13:11 crc kubenswrapper[4824]: I0121 11:13:11.034359 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Jan 21 11:13:11 crc kubenswrapper[4824]: I0121 11:13:11.055232 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Jan 21 11:13:11 crc kubenswrapper[4824]: I0121 11:13:11.073947 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Jan 21 11:13:11 crc kubenswrapper[4824]: I0121 11:13:11.094645 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Jan 21 11:13:11 crc kubenswrapper[4824]: I0121 11:13:11.114180 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Jan 21 11:13:11 crc kubenswrapper[4824]: I0121 11:13:11.134764 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Jan 21 11:13:11 crc kubenswrapper[4824]: I0121 11:13:11.154045 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Jan 21 11:13:11 crc kubenswrapper[4824]: I0121 11:13:11.175337 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Jan 21 11:13:11 crc kubenswrapper[4824]: I0121 11:13:11.194033 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Jan 21 11:13:11 crc kubenswrapper[4824]: I0121 11:13:11.214311 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Jan 21 11:13:11 crc kubenswrapper[4824]: I0121 11:13:11.234830 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Jan 21 11:13:11 crc kubenswrapper[4824]: I0121 11:13:11.254064 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Jan 21 11:13:11 crc kubenswrapper[4824]: I0121 11:13:11.273931 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Jan 21 11:13:11 crc kubenswrapper[4824]: I0121 11:13:11.294349 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Jan 21 11:13:11 crc kubenswrapper[4824]: I0121 11:13:11.314547 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Jan 21 11:13:11 crc kubenswrapper[4824]: I0121 11:13:11.334793 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Jan 21 11:13:11 crc kubenswrapper[4824]: I0121 11:13:11.353916 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Jan 21 11:13:11 crc kubenswrapper[4824]: I0121 11:13:11.374368 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Jan 21 11:13:11 crc kubenswrapper[4824]: I0121 11:13:11.393877 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Jan 21 11:13:11 crc kubenswrapper[4824]: I0121 11:13:11.434380 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Jan 21 11:13:11 crc kubenswrapper[4824]: I0121 11:13:11.454681 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Jan 21 11:13:11 crc kubenswrapper[4824]: I0121 11:13:11.474535 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Jan 21 11:13:11 crc kubenswrapper[4824]: I0121 11:13:11.494196 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Jan 21 11:13:11 crc kubenswrapper[4824]: I0121 11:13:11.514863 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Jan 21 11:13:11 crc kubenswrapper[4824]: I0121 11:13:11.533717 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Jan 21 11:13:11 crc kubenswrapper[4824]: I0121 11:13:11.554263 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Jan 21 11:13:11 crc kubenswrapper[4824]: I0121 11:13:11.574361 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Jan 21 11:13:11 crc kubenswrapper[4824]: I0121 11:13:11.594504 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Jan 21 11:13:11 crc kubenswrapper[4824]: I0121 11:13:11.614299 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Jan 21 11:13:11 crc kubenswrapper[4824]: I0121 11:13:11.634024 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Jan 21 11:13:11 crc kubenswrapper[4824]: I0121 11:13:11.654227 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Jan 21 11:13:11 crc kubenswrapper[4824]: I0121 11:13:11.674174 4824 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Jan 21 11:13:11 crc kubenswrapper[4824]: I0121 11:13:11.694277 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Jan 21 11:13:11 crc kubenswrapper[4824]: I0121 11:13:11.714530 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Jan 21 11:13:11 crc kubenswrapper[4824]: I0121 11:13:11.746270 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-skznz\" (UniqueName: \"kubernetes.io/projected/49ec79bd-9eb8-42c8-a262-e96c1457f686-kube-api-access-skznz\") pod \"openshift-apiserver-operator-796bbdcf4f-qx94f\" (UID: \"49ec79bd-9eb8-42c8-a262-e96c1457f686\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-qx94f" Jan 21 11:13:11 crc kubenswrapper[4824]: I0121 11:13:11.764473 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fm7v7\" (UniqueName: \"kubernetes.io/projected/8364adda-37b5-4a9d-8abd-3daf12470f20-kube-api-access-fm7v7\") pod \"apiserver-76f77b778f-htg7q\" (UID: \"8364adda-37b5-4a9d-8abd-3daf12470f20\") " pod="openshift-apiserver/apiserver-76f77b778f-htg7q" Jan 21 11:13:11 crc kubenswrapper[4824]: I0121 11:13:11.785821 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rl46h\" (UniqueName: \"kubernetes.io/projected/29d5a4fe-fd42-42ec-85b4-c79af6424296-kube-api-access-rl46h\") pod \"console-operator-58897d9998-mh6ds\" (UID: \"29d5a4fe-fd42-42ec-85b4-c79af6424296\") " pod="openshift-console-operator/console-operator-58897d9998-mh6ds" Jan 21 11:13:11 crc kubenswrapper[4824]: I0121 11:13:11.805156 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n6nx5\" (UniqueName: \"kubernetes.io/projected/8d0d706a-5fe5-402d-9040-3c763e525554-kube-api-access-n6nx5\") pod \"openshift-config-operator-7777fb866f-fbq6l\" (UID: \"8d0d706a-5fe5-402d-9040-3c763e525554\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-fbq6l" Jan 21 11:13:11 crc kubenswrapper[4824]: I0121 11:13:11.824426 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dg7z5\" (UniqueName: \"kubernetes.io/projected/d962d0be-438a-4d08-8113-525238fa2156-kube-api-access-dg7z5\") pod \"controller-manager-879f6c89f-4dkd8\" (UID: \"d962d0be-438a-4d08-8113-525238fa2156\") " pod="openshift-controller-manager/controller-manager-879f6c89f-4dkd8" Jan 21 11:13:11 crc kubenswrapper[4824]: I0121 11:13:11.844862 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gh2qf\" (UniqueName: \"kubernetes.io/projected/c4b9a738-6a66-467d-b8a5-bddb20cb7e98-kube-api-access-gh2qf\") pod \"authentication-operator-69f744f599-s2n22\" (UID: \"c4b9a738-6a66-467d-b8a5-bddb20cb7e98\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-s2n22" Jan 21 11:13:11 crc kubenswrapper[4824]: I0121 11:13:11.853733 4824 request.go:700] Waited for 1.937689926s due to client-side throttling, not priority and fairness, request: POST:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-oauth-apiserver/serviceaccounts/oauth-apiserver-sa/token Jan 21 11:13:11 crc kubenswrapper[4824]: I0121 11:13:11.853867 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-htg7q" Jan 21 11:13:11 crc kubenswrapper[4824]: I0121 11:13:11.863359 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-4dkd8" Jan 21 11:13:11 crc kubenswrapper[4824]: I0121 11:13:11.869342 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-klfwh\" (UniqueName: \"kubernetes.io/projected/0475fd50-10a8-4195-8643-9408478e953c-kube-api-access-klfwh\") pod \"apiserver-7bbb656c7d-7s7wx\" (UID: \"0475fd50-10a8-4195-8643-9408478e953c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7s7wx" Jan 21 11:13:11 crc kubenswrapper[4824]: I0121 11:13:11.887743 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q8qjw\" (UniqueName: \"kubernetes.io/projected/c585ef2f-3e29-4b81-918f-11b075006ca0-kube-api-access-q8qjw\") pod \"console-f9d7485db-nwlwd\" (UID: \"c585ef2f-3e29-4b81-918f-11b075006ca0\") " pod="openshift-console/console-f9d7485db-nwlwd" Jan 21 11:13:11 crc kubenswrapper[4824]: I0121 11:13:11.890316 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-s2n22" Jan 21 11:13:11 crc kubenswrapper[4824]: I0121 11:13:11.908834 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5zq9z\" (UniqueName: \"kubernetes.io/projected/fb8bdb2b-bdc7-4433-be10-296fc67dc1c3-kube-api-access-5zq9z\") pod \"machine-api-operator-5694c8668f-7grz2\" (UID: \"fb8bdb2b-bdc7-4433-be10-296fc67dc1c3\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-7grz2" Jan 21 11:13:11 crc kubenswrapper[4824]: I0121 11:13:11.925350 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-qx94f" Jan 21 11:13:11 crc kubenswrapper[4824]: I0121 11:13:11.935070 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Jan 21 11:13:11 crc kubenswrapper[4824]: I0121 11:13:11.936886 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-fbq6l" Jan 21 11:13:11 crc kubenswrapper[4824]: I0121 11:13:11.955328 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-mh6ds" Jan 21 11:13:11 crc kubenswrapper[4824]: I0121 11:13:11.955736 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Jan 21 11:13:11 crc kubenswrapper[4824]: I0121 11:13:11.962686 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-nwlwd" Jan 21 11:13:11 crc kubenswrapper[4824]: I0121 11:13:11.968493 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7s7wx" Jan 21 11:13:11 crc kubenswrapper[4824]: I0121 11:13:11.975206 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.002156 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-htg7q"] Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.002891 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Jan 21 11:13:12 crc kubenswrapper[4824]: W0121 11:13:12.019875 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8364adda_37b5_4a9d_8abd_3daf12470f20.slice/crio-57ced44b4bd056825dfdc76b73c5ce663d44e6ce7c97b61af5207421cc94a24c WatchSource:0}: Error finding container 57ced44b4bd056825dfdc76b73c5ce663d44e6ce7c97b61af5207421cc94a24c: Status 404 returned error can't find the container with id 57ced44b4bd056825dfdc76b73c5ce663d44e6ce7c97b61af5207421cc94a24c Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.025661 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-4dkd8"] Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.033533 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xgr59\" (UniqueName: \"kubernetes.io/projected/024ed20d-cf91-4660-998b-3d06d3464607-kube-api-access-xgr59\") pod \"etcd-operator-b45778765-cpgg4\" (UID: \"024ed20d-cf91-4660-998b-3d06d3464607\") " pod="openshift-etcd-operator/etcd-operator-b45778765-cpgg4" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.033564 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/51e27630-d5c1-4eed-8232-3698ab99fc19-config\") pod \"route-controller-manager-6576b87f9c-qs8d6\" (UID: \"51e27630-d5c1-4eed-8232-3698ab99fc19\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qs8d6" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.033597 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d132606d-b1ae-4cb7-865e-e1065c2eb4a9-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-t2thx\" (UID: \"d132606d-b1ae-4cb7-865e-e1065c2eb4a9\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-t2thx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.033613 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q27ql\" (UniqueName: \"kubernetes.io/projected/5dab0fba-f35e-42bd-9a7e-913b70c2dd2b-kube-api-access-q27ql\") pod \"ingress-operator-5b745b69d9-n8lql\" (UID: \"5dab0fba-f35e-42bd-9a7e-913b70c2dd2b\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-n8lql" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.033627 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/13490666-4efb-492b-be88-1c6fde3f378e-serving-cert\") pod \"service-ca-operator-777779d784-c5xks\" (UID: \"13490666-4efb-492b-be88-1c6fde3f378e\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-c5xks" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.033675 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h6sxp\" (UniqueName: \"kubernetes.io/projected/16be0b45-eab8-4d0a-823b-5d877031dba5-kube-api-access-h6sxp\") pod \"olm-operator-6b444d44fb-cq5sn\" (UID: \"16be0b45-eab8-4d0a-823b-5d877031dba5\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-cq5sn" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.033691 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fh969\" (UniqueName: \"kubernetes.io/projected/74172695-d17f-4a47-b013-9e025e534916-kube-api-access-fh969\") pod \"kube-storage-version-migrator-operator-b67b599dd-s7wn9\" (UID: \"74172695-d17f-4a47-b013-9e025e534916\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-s7wn9" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.033723 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/5dab0fba-f35e-42bd-9a7e-913b70c2dd2b-bound-sa-token\") pod \"ingress-operator-5b745b69d9-n8lql\" (UID: \"5dab0fba-f35e-42bd-9a7e-913b70c2dd2b\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-n8lql" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.033740 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5ksj6\" (UniqueName: \"kubernetes.io/projected/51e27630-d5c1-4eed-8232-3698ab99fc19-kube-api-access-5ksj6\") pod \"route-controller-manager-6576b87f9c-qs8d6\" (UID: \"51e27630-d5c1-4eed-8232-3698ab99fc19\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qs8d6" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.033756 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c873221f-c31d-411a-bbf5-c59e45b473a5-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-5lmnb\" (UID: \"c873221f-c31d-411a-bbf5-c59e45b473a5\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5lmnb" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.033773 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/b79b805a-7ab1-467f-90a5-ac7944cc564e-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-2tjlq\" (UID: \"b79b805a-7ab1-467f-90a5-ac7944cc564e\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2tjlq" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.033839 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/024ed20d-cf91-4660-998b-3d06d3464607-etcd-service-ca\") pod \"etcd-operator-b45778765-cpgg4\" (UID: \"024ed20d-cf91-4660-998b-3d06d3464607\") " pod="openshift-etcd-operator/etcd-operator-b45778765-cpgg4" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.033869 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/86dc1658-d302-449f-9ae3-d7bc5e795c57-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-bkkzh\" (UID: \"86dc1658-d302-449f-9ae3-d7bc5e795c57\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-bkkzh" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.033885 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/3358782c-3435-4c68-9f9c-c1040a6ada32-signing-key\") pod \"service-ca-9c57cc56f-rdcbg\" (UID: \"3358782c-3435-4c68-9f9c-c1040a6ada32\") " pod="openshift-service-ca/service-ca-9c57cc56f-rdcbg" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.033900 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c873221f-c31d-411a-bbf5-c59e45b473a5-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-5lmnb\" (UID: \"c873221f-c31d-411a-bbf5-c59e45b473a5\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5lmnb" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.033924 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/14d99a1f-fca3-421a-8b47-ab4cda2f77fc-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-gjhp5\" (UID: \"14d99a1f-fca3-421a-8b47-ab4cda2f77fc\") " pod="openshift-marketplace/marketplace-operator-79b997595-gjhp5" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.033948 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/024ed20d-cf91-4660-998b-3d06d3464607-serving-cert\") pod \"etcd-operator-b45778765-cpgg4\" (UID: \"024ed20d-cf91-4660-998b-3d06d3464607\") " pod="openshift-etcd-operator/etcd-operator-b45778765-cpgg4" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.034045 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/3358782c-3435-4c68-9f9c-c1040a6ada32-signing-cabundle\") pod \"service-ca-9c57cc56f-rdcbg\" (UID: \"3358782c-3435-4c68-9f9c-c1040a6ada32\") " pod="openshift-service-ca/service-ca-9c57cc56f-rdcbg" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.034073 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/94e71514-0af6-4254-8815-c933ade6c9da-installation-pull-secrets\") pod \"image-registry-697d97f7c8-wxlxx\" (UID: \"94e71514-0af6-4254-8815-c933ade6c9da\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxlxx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.034090 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kjjkx\" (UniqueName: \"kubernetes.io/projected/14d99a1f-fca3-421a-8b47-ab4cda2f77fc-kube-api-access-kjjkx\") pod \"marketplace-operator-79b997595-gjhp5\" (UID: \"14d99a1f-fca3-421a-8b47-ab4cda2f77fc\") " pod="openshift-marketplace/marketplace-operator-79b997595-gjhp5" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.034107 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b79b805a-7ab1-467f-90a5-ac7944cc564e-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-2tjlq\" (UID: \"b79b805a-7ab1-467f-90a5-ac7944cc564e\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2tjlq" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.034121 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/07ad02bb-f06a-4619-87d1-112c78642603-config\") pod \"kube-controller-manager-operator-78b949d7b-7fl6h\" (UID: \"07ad02bb-f06a-4619-87d1-112c78642603\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-7fl6h" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.034137 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6q9tc\" (UniqueName: \"kubernetes.io/projected/4b22a7aa-bcc0-4876-a374-a9943a6040bd-kube-api-access-6q9tc\") pod \"router-default-5444994796-bblkp\" (UID: \"4b22a7aa-bcc0-4876-a374-a9943a6040bd\") " pod="openshift-ingress/router-default-5444994796-bblkp" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.034159 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/94e71514-0af6-4254-8815-c933ade6c9da-bound-sa-token\") pod \"image-registry-697d97f7c8-wxlxx\" (UID: \"94e71514-0af6-4254-8815-c933ade6c9da\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxlxx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.034172 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/86b5dca2-989f-4b6b-983b-87763ce0d469-audit-policies\") pod \"oauth-openshift-558db77b4-fz7gx\" (UID: \"86b5dca2-989f-4b6b-983b-87763ce0d469\") " pod="openshift-authentication/oauth-openshift-558db77b4-fz7gx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.034186 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/37d2c136-5241-41da-a630-3d70273f1670-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-cnr8w\" (UID: \"37d2c136-5241-41da-a630-3d70273f1670\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-cnr8w" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.034214 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2552x\" (UniqueName: \"kubernetes.io/projected/27b93806-10a5-46ba-8a81-20b05fd990c1-kube-api-access-2552x\") pod \"migrator-59844c95c7-j6bst\" (UID: \"27b93806-10a5-46ba-8a81-20b05fd990c1\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-j6bst" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.034239 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/7352d717-b2a3-4cd2-9e35-3ae2fe2d7b34-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-27d2q\" (UID: \"7352d717-b2a3-4cd2-9e35-3ae2fe2d7b34\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-27d2q" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.034255 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6qg69\" (UniqueName: \"kubernetes.io/projected/ae7ce20e-9b65-4dd4-9e00-d96cf91803cb-kube-api-access-6qg69\") pod \"dns-operator-744455d44c-f6c7b\" (UID: \"ae7ce20e-9b65-4dd4-9e00-d96cf91803cb\") " pod="openshift-dns-operator/dns-operator-744455d44c-f6c7b" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.034272 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/5dab0fba-f35e-42bd-9a7e-913b70c2dd2b-metrics-tls\") pod \"ingress-operator-5b745b69d9-n8lql\" (UID: \"5dab0fba-f35e-42bd-9a7e-913b70c2dd2b\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-n8lql" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.034285 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/ae7ce20e-9b65-4dd4-9e00-d96cf91803cb-metrics-tls\") pod \"dns-operator-744455d44c-f6c7b\" (UID: \"ae7ce20e-9b65-4dd4-9e00-d96cf91803cb\") " pod="openshift-dns-operator/dns-operator-744455d44c-f6c7b" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.034309 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/86b5dca2-989f-4b6b-983b-87763ce0d469-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-fz7gx\" (UID: \"86b5dca2-989f-4b6b-983b-87763ce0d469\") " pod="openshift-authentication/oauth-openshift-558db77b4-fz7gx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.034349 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/86b5dca2-989f-4b6b-983b-87763ce0d469-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-fz7gx\" (UID: \"86b5dca2-989f-4b6b-983b-87763ce0d469\") " pod="openshift-authentication/oauth-openshift-558db77b4-fz7gx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.034382 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nqqww\" (UniqueName: \"kubernetes.io/projected/09bd0b9c-46fb-433f-bd30-014c69c90d9f-kube-api-access-nqqww\") pod \"collect-profiles-29483220-6r4fz\" (UID: \"09bd0b9c-46fb-433f-bd30-014c69c90d9f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29483220-6r4fz" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.034426 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pczg5\" (UniqueName: \"kubernetes.io/projected/86dc1658-d302-449f-9ae3-d7bc5e795c57-kube-api-access-pczg5\") pod \"package-server-manager-789f6589d5-bkkzh\" (UID: \"86dc1658-d302-449f-9ae3-d7bc5e795c57\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-bkkzh" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.034441 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/14d99a1f-fca3-421a-8b47-ab4cda2f77fc-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-gjhp5\" (UID: \"14d99a1f-fca3-421a-8b47-ab4cda2f77fc\") " pod="openshift-marketplace/marketplace-operator-79b997595-gjhp5" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.034462 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/024ed20d-cf91-4660-998b-3d06d3464607-etcd-client\") pod \"etcd-operator-b45778765-cpgg4\" (UID: \"024ed20d-cf91-4660-998b-3d06d3464607\") " pod="openshift-etcd-operator/etcd-operator-b45778765-cpgg4" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.034477 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/09bd0b9c-46fb-433f-bd30-014c69c90d9f-config-volume\") pod \"collect-profiles-29483220-6r4fz\" (UID: \"09bd0b9c-46fb-433f-bd30-014c69c90d9f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29483220-6r4fz" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.035226 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/94e71514-0af6-4254-8815-c933ade6c9da-registry-certificates\") pod \"image-registry-697d97f7c8-wxlxx\" (UID: \"94e71514-0af6-4254-8815-c933ade6c9da\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxlxx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.035249 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/86b5dca2-989f-4b6b-983b-87763ce0d469-audit-dir\") pod \"oauth-openshift-558db77b4-fz7gx\" (UID: \"86b5dca2-989f-4b6b-983b-87763ce0d469\") " pod="openshift-authentication/oauth-openshift-558db77b4-fz7gx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.035280 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxlxx\" (UID: \"94e71514-0af6-4254-8815-c933ade6c9da\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxlxx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.035308 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/74172695-d17f-4a47-b013-9e025e534916-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-s7wn9\" (UID: \"74172695-d17f-4a47-b013-9e025e534916\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-s7wn9" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.035420 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hlp6l\" (UniqueName: \"kubernetes.io/projected/6a0a5723-a022-41f3-9222-c62cac06397e-kube-api-access-hlp6l\") pod \"cluster-samples-operator-665b6dd947-ct4lp\" (UID: \"6a0a5723-a022-41f3-9222-c62cac06397e\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-ct4lp" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.035452 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dzrmn\" (UniqueName: \"kubernetes.io/projected/7352d717-b2a3-4cd2-9e35-3ae2fe2d7b34-kube-api-access-dzrmn\") pod \"multus-admission-controller-857f4d67dd-27d2q\" (UID: \"7352d717-b2a3-4cd2-9e35-3ae2fe2d7b34\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-27d2q" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.035481 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/86b5dca2-989f-4b6b-983b-87763ce0d469-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-fz7gx\" (UID: \"86b5dca2-989f-4b6b-983b-87763ce0d469\") " pod="openshift-authentication/oauth-openshift-558db77b4-fz7gx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.035537 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/02c3b11b-dbc4-445d-94fa-95c5ccc03cb8-srv-cert\") pod \"catalog-operator-68c6474976-j8dcx\" (UID: \"02c3b11b-dbc4-445d-94fa-95c5ccc03cb8\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-j8dcx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.035555 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j8xhg\" (UniqueName: \"kubernetes.io/projected/37d2c136-5241-41da-a630-3d70273f1670-kube-api-access-j8xhg\") pod \"openshift-controller-manager-operator-756b6f6bc6-cnr8w\" (UID: \"37d2c136-5241-41da-a630-3d70273f1670\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-cnr8w" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.035585 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/86b5dca2-989f-4b6b-983b-87763ce0d469-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-fz7gx\" (UID: \"86b5dca2-989f-4b6b-983b-87763ce0d469\") " pod="openshift-authentication/oauth-openshift-558db77b4-fz7gx" Jan 21 11:13:12 crc kubenswrapper[4824]: E0121 11:13:12.035758 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-21 11:13:12.535744472 +0000 UTC m=+134.828773765 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxlxx" (UID: "94e71514-0af6-4254-8815-c933ade6c9da") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.035878 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/07ad02bb-f06a-4619-87d1-112c78642603-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-7fl6h\" (UID: \"07ad02bb-f06a-4619-87d1-112c78642603\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-7fl6h" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.035966 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bhtz6\" (UniqueName: \"kubernetes.io/projected/86b5dca2-989f-4b6b-983b-87763ce0d469-kube-api-access-bhtz6\") pod \"oauth-openshift-558db77b4-fz7gx\" (UID: \"86b5dca2-989f-4b6b-983b-87763ce0d469\") " pod="openshift-authentication/oauth-openshift-558db77b4-fz7gx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.035989 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/37d2c136-5241-41da-a630-3d70273f1670-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-cnr8w\" (UID: \"37d2c136-5241-41da-a630-3d70273f1670\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-cnr8w" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.036188 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/86b5dca2-989f-4b6b-983b-87763ce0d469-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-fz7gx\" (UID: \"86b5dca2-989f-4b6b-983b-87763ce0d469\") " pod="openshift-authentication/oauth-openshift-558db77b4-fz7gx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.036374 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/86b5dca2-989f-4b6b-983b-87763ce0d469-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-fz7gx\" (UID: \"86b5dca2-989f-4b6b-983b-87763ce0d469\") " pod="openshift-authentication/oauth-openshift-558db77b4-fz7gx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.036430 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d132606d-b1ae-4cb7-865e-e1065c2eb4a9-config\") pod \"kube-apiserver-operator-766d6c64bb-t2thx\" (UID: \"d132606d-b1ae-4cb7-865e-e1065c2eb4a9\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-t2thx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.036471 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/07ad02bb-f06a-4619-87d1-112c78642603-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-7fl6h\" (UID: \"07ad02bb-f06a-4619-87d1-112c78642603\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-7fl6h" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.036501 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/51e27630-d5c1-4eed-8232-3698ab99fc19-serving-cert\") pod \"route-controller-manager-6576b87f9c-qs8d6\" (UID: \"51e27630-d5c1-4eed-8232-3698ab99fc19\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qs8d6" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.036593 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/02c3b11b-dbc4-445d-94fa-95c5ccc03cb8-profile-collector-cert\") pod \"catalog-operator-68c6474976-j8dcx\" (UID: \"02c3b11b-dbc4-445d-94fa-95c5ccc03cb8\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-j8dcx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.036764 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/94e71514-0af6-4254-8815-c933ade6c9da-registry-tls\") pod \"image-registry-697d97f7c8-wxlxx\" (UID: \"94e71514-0af6-4254-8815-c933ade6c9da\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxlxx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.036787 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/86b5dca2-989f-4b6b-983b-87763ce0d469-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-fz7gx\" (UID: \"86b5dca2-989f-4b6b-983b-87763ce0d469\") " pod="openshift-authentication/oauth-openshift-558db77b4-fz7gx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.036846 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/7a5d2d02-347e-4799-a28f-f303d51758d0-auth-proxy-config\") pod \"machine-approver-56656f9798-hzhhs\" (UID: \"7a5d2d02-347e-4799-a28f-f303d51758d0\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-hzhhs" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.037016 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/6a0a5723-a022-41f3-9222-c62cac06397e-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-ct4lp\" (UID: \"6a0a5723-a022-41f3-9222-c62cac06397e\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-ct4lp" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.037298 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/16be0b45-eab8-4d0a-823b-5d877031dba5-profile-collector-cert\") pod \"olm-operator-6b444d44fb-cq5sn\" (UID: \"16be0b45-eab8-4d0a-823b-5d877031dba5\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-cq5sn" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.037335 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nk6sw\" (UniqueName: \"kubernetes.io/projected/b79b805a-7ab1-467f-90a5-ac7944cc564e-kube-api-access-nk6sw\") pod \"cluster-image-registry-operator-dc59b4c8b-2tjlq\" (UID: \"b79b805a-7ab1-467f-90a5-ac7944cc564e\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2tjlq" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.037377 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/74172695-d17f-4a47-b013-9e025e534916-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-s7wn9\" (UID: \"74172695-d17f-4a47-b013-9e025e534916\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-s7wn9" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.037421 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7a5d2d02-347e-4799-a28f-f303d51758d0-config\") pod \"machine-approver-56656f9798-hzhhs\" (UID: \"7a5d2d02-347e-4799-a28f-f303d51758d0\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-hzhhs" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.037486 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tkwb6\" (UniqueName: \"kubernetes.io/projected/c117483a-fe5c-423b-995e-9c2e7f59566a-kube-api-access-tkwb6\") pod \"control-plane-machine-set-operator-78cbb6b69f-r65dm\" (UID: \"c117483a-fe5c-423b-995e-9c2e7f59566a\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-r65dm" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.037538 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pwmbq\" (UniqueName: \"kubernetes.io/projected/7a5d2d02-347e-4799-a28f-f303d51758d0-kube-api-access-pwmbq\") pod \"machine-approver-56656f9798-hzhhs\" (UID: \"7a5d2d02-347e-4799-a28f-f303d51758d0\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-hzhhs" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.037554 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4b22a7aa-bcc0-4876-a374-a9943a6040bd-service-ca-bundle\") pod \"router-default-5444994796-bblkp\" (UID: \"4b22a7aa-bcc0-4876-a374-a9943a6040bd\") " pod="openshift-ingress/router-default-5444994796-bblkp" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.037579 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/13490666-4efb-492b-be88-1c6fde3f378e-config\") pod \"service-ca-operator-777779d784-c5xks\" (UID: \"13490666-4efb-492b-be88-1c6fde3f378e\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-c5xks" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.037658 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zcznm\" (UniqueName: \"kubernetes.io/projected/94e71514-0af6-4254-8815-c933ade6c9da-kube-api-access-zcznm\") pod \"image-registry-697d97f7c8-wxlxx\" (UID: \"94e71514-0af6-4254-8815-c933ade6c9da\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxlxx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.037678 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/86b5dca2-989f-4b6b-983b-87763ce0d469-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-fz7gx\" (UID: \"86b5dca2-989f-4b6b-983b-87763ce0d469\") " pod="openshift-authentication/oauth-openshift-558db77b4-fz7gx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.038375 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/c117483a-fe5c-423b-995e-9c2e7f59566a-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-r65dm\" (UID: \"c117483a-fe5c-423b-995e-9c2e7f59566a\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-r65dm" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.038447 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/86b5dca2-989f-4b6b-983b-87763ce0d469-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-fz7gx\" (UID: \"86b5dca2-989f-4b6b-983b-87763ce0d469\") " pod="openshift-authentication/oauth-openshift-558db77b4-fz7gx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.038469 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/024ed20d-cf91-4660-998b-3d06d3464607-config\") pod \"etcd-operator-b45778765-cpgg4\" (UID: \"024ed20d-cf91-4660-998b-3d06d3464607\") " pod="openshift-etcd-operator/etcd-operator-b45778765-cpgg4" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.038520 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5dab0fba-f35e-42bd-9a7e-913b70c2dd2b-trusted-ca\") pod \"ingress-operator-5b745b69d9-n8lql\" (UID: \"5dab0fba-f35e-42bd-9a7e-913b70c2dd2b\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-n8lql" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.038538 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4b22a7aa-bcc0-4876-a374-a9943a6040bd-metrics-certs\") pod \"router-default-5444994796-bblkp\" (UID: \"4b22a7aa-bcc0-4876-a374-a9943a6040bd\") " pod="openshift-ingress/router-default-5444994796-bblkp" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.038558 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c873221f-c31d-411a-bbf5-c59e45b473a5-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-5lmnb\" (UID: \"c873221f-c31d-411a-bbf5-c59e45b473a5\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5lmnb" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.038585 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/b79b805a-7ab1-467f-90a5-ac7944cc564e-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-2tjlq\" (UID: \"b79b805a-7ab1-467f-90a5-ac7944cc564e\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2tjlq" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.038604 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pmsvd\" (UniqueName: \"kubernetes.io/projected/3358782c-3435-4c68-9f9c-c1040a6ada32-kube-api-access-pmsvd\") pod \"service-ca-9c57cc56f-rdcbg\" (UID: \"3358782c-3435-4c68-9f9c-c1040a6ada32\") " pod="openshift-service-ca/service-ca-9c57cc56f-rdcbg" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.038669 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/7a5d2d02-347e-4799-a28f-f303d51758d0-machine-approver-tls\") pod \"machine-approver-56656f9798-hzhhs\" (UID: \"7a5d2d02-347e-4799-a28f-f303d51758d0\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-hzhhs" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.038722 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/51e27630-d5c1-4eed-8232-3698ab99fc19-client-ca\") pod \"route-controller-manager-6576b87f9c-qs8d6\" (UID: \"51e27630-d5c1-4eed-8232-3698ab99fc19\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qs8d6" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.038745 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/024ed20d-cf91-4660-998b-3d06d3464607-etcd-ca\") pod \"etcd-operator-b45778765-cpgg4\" (UID: \"024ed20d-cf91-4660-998b-3d06d3464607\") " pod="openshift-etcd-operator/etcd-operator-b45778765-cpgg4" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.039081 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/86b5dca2-989f-4b6b-983b-87763ce0d469-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-fz7gx\" (UID: \"86b5dca2-989f-4b6b-983b-87763ce0d469\") " pod="openshift-authentication/oauth-openshift-558db77b4-fz7gx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.039103 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d132606d-b1ae-4cb7-865e-e1065c2eb4a9-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-t2thx\" (UID: \"d132606d-b1ae-4cb7-865e-e1065c2eb4a9\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-t2thx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.039129 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/94e71514-0af6-4254-8815-c933ade6c9da-trusted-ca\") pod \"image-registry-697d97f7c8-wxlxx\" (UID: \"94e71514-0af6-4254-8815-c933ade6c9da\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxlxx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.039143 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/09bd0b9c-46fb-433f-bd30-014c69c90d9f-secret-volume\") pod \"collect-profiles-29483220-6r4fz\" (UID: \"09bd0b9c-46fb-433f-bd30-014c69c90d9f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29483220-6r4fz" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.039157 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/4b22a7aa-bcc0-4876-a374-a9943a6040bd-default-certificate\") pod \"router-default-5444994796-bblkp\" (UID: \"4b22a7aa-bcc0-4876-a374-a9943a6040bd\") " pod="openshift-ingress/router-default-5444994796-bblkp" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.039186 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-md94v\" (UniqueName: \"kubernetes.io/projected/02c3b11b-dbc4-445d-94fa-95c5ccc03cb8-kube-api-access-md94v\") pod \"catalog-operator-68c6474976-j8dcx\" (UID: \"02c3b11b-dbc4-445d-94fa-95c5ccc03cb8\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-j8dcx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.039221 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/86b5dca2-989f-4b6b-983b-87763ce0d469-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-fz7gx\" (UID: \"86b5dca2-989f-4b6b-983b-87763ce0d469\") " pod="openshift-authentication/oauth-openshift-558db77b4-fz7gx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.039237 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jsn9x\" (UniqueName: \"kubernetes.io/projected/2279132e-c274-424b-8c8b-80ae5bd7dfe5-kube-api-access-jsn9x\") pod \"downloads-7954f5f757-8g4qk\" (UID: \"2279132e-c274-424b-8c8b-80ae5bd7dfe5\") " pod="openshift-console/downloads-7954f5f757-8g4qk" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.039250 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/4b22a7aa-bcc0-4876-a374-a9943a6040bd-stats-auth\") pod \"router-default-5444994796-bblkp\" (UID: \"4b22a7aa-bcc0-4876-a374-a9943a6040bd\") " pod="openshift-ingress/router-default-5444994796-bblkp" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.039279 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/94e71514-0af6-4254-8815-c933ade6c9da-ca-trust-extracted\") pod \"image-registry-697d97f7c8-wxlxx\" (UID: \"94e71514-0af6-4254-8815-c933ade6c9da\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxlxx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.039294 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/16be0b45-eab8-4d0a-823b-5d877031dba5-srv-cert\") pod \"olm-operator-6b444d44fb-cq5sn\" (UID: \"16be0b45-eab8-4d0a-823b-5d877031dba5\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-cq5sn" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.039309 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qrlgj\" (UniqueName: \"kubernetes.io/projected/13490666-4efb-492b-be88-1c6fde3f378e-kube-api-access-qrlgj\") pod \"service-ca-operator-777779d784-c5xks\" (UID: \"13490666-4efb-492b-be88-1c6fde3f378e\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-c5xks" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.061790 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-s2n22"] Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.140651 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.140821 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/74172695-d17f-4a47-b013-9e025e534916-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-s7wn9\" (UID: \"74172695-d17f-4a47-b013-9e025e534916\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-s7wn9" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.140860 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/d93236c5-17ab-4cd0-88a3-f4b3479067e5-proxy-tls\") pod \"machine-config-controller-84d6567774-9tldv\" (UID: \"d93236c5-17ab-4cd0-88a3-f4b3479067e5\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-9tldv" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.140877 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c7fe95fc-b7d1-4ad1-8acf-e7ee4564babe-config-volume\") pod \"dns-default-dlgmw\" (UID: \"c7fe95fc-b7d1-4ad1-8acf-e7ee4564babe\") " pod="openshift-dns/dns-default-dlgmw" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.140898 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hlp6l\" (UniqueName: \"kubernetes.io/projected/6a0a5723-a022-41f3-9222-c62cac06397e-kube-api-access-hlp6l\") pod \"cluster-samples-operator-665b6dd947-ct4lp\" (UID: \"6a0a5723-a022-41f3-9222-c62cac06397e\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-ct4lp" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.140916 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/86b5dca2-989f-4b6b-983b-87763ce0d469-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-fz7gx\" (UID: \"86b5dca2-989f-4b6b-983b-87763ce0d469\") " pod="openshift-authentication/oauth-openshift-558db77b4-fz7gx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.140932 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/02c3b11b-dbc4-445d-94fa-95c5ccc03cb8-srv-cert\") pod \"catalog-operator-68c6474976-j8dcx\" (UID: \"02c3b11b-dbc4-445d-94fa-95c5ccc03cb8\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-j8dcx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.140946 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j8xhg\" (UniqueName: \"kubernetes.io/projected/37d2c136-5241-41da-a630-3d70273f1670-kube-api-access-j8xhg\") pod \"openshift-controller-manager-operator-756b6f6bc6-cnr8w\" (UID: \"37d2c136-5241-41da-a630-3d70273f1670\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-cnr8w" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.140975 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dzrmn\" (UniqueName: \"kubernetes.io/projected/7352d717-b2a3-4cd2-9e35-3ae2fe2d7b34-kube-api-access-dzrmn\") pod \"multus-admission-controller-857f4d67dd-27d2q\" (UID: \"7352d717-b2a3-4cd2-9e35-3ae2fe2d7b34\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-27d2q" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.140992 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/86b5dca2-989f-4b6b-983b-87763ce0d469-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-fz7gx\" (UID: \"86b5dca2-989f-4b6b-983b-87763ce0d469\") " pod="openshift-authentication/oauth-openshift-558db77b4-fz7gx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.141102 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/07ad02bb-f06a-4619-87d1-112c78642603-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-7fl6h\" (UID: \"07ad02bb-f06a-4619-87d1-112c78642603\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-7fl6h" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.141126 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/63c09cd8-12ad-40a1-9877-3fc1500acb7a-auth-proxy-config\") pod \"machine-config-operator-74547568cd-kqtr5\" (UID: \"63c09cd8-12ad-40a1-9877-3fc1500acb7a\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kqtr5" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.141142 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/37d2c136-5241-41da-a630-3d70273f1670-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-cnr8w\" (UID: \"37d2c136-5241-41da-a630-3d70273f1670\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-cnr8w" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.141159 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/faede743-203b-4bd1-b17a-776a4744bdf1-cert\") pod \"ingress-canary-b7tsx\" (UID: \"faede743-203b-4bd1-b17a-776a4744bdf1\") " pod="openshift-ingress-canary/ingress-canary-b7tsx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.141173 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bhtz6\" (UniqueName: \"kubernetes.io/projected/86b5dca2-989f-4b6b-983b-87763ce0d469-kube-api-access-bhtz6\") pod \"oauth-openshift-558db77b4-fz7gx\" (UID: \"86b5dca2-989f-4b6b-983b-87763ce0d469\") " pod="openshift-authentication/oauth-openshift-558db77b4-fz7gx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.141188 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j48kx\" (UniqueName: \"kubernetes.io/projected/23d7bb25-5d2c-4530-86dd-25a956953a4f-kube-api-access-j48kx\") pod \"packageserver-d55dfcdfc-cslqn\" (UID: \"23d7bb25-5d2c-4530-86dd-25a956953a4f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-cslqn" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.141215 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/86b5dca2-989f-4b6b-983b-87763ce0d469-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-fz7gx\" (UID: \"86b5dca2-989f-4b6b-983b-87763ce0d469\") " pod="openshift-authentication/oauth-openshift-558db77b4-fz7gx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.141233 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/86b5dca2-989f-4b6b-983b-87763ce0d469-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-fz7gx\" (UID: \"86b5dca2-989f-4b6b-983b-87763ce0d469\") " pod="openshift-authentication/oauth-openshift-558db77b4-fz7gx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.141248 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d132606d-b1ae-4cb7-865e-e1065c2eb4a9-config\") pod \"kube-apiserver-operator-766d6c64bb-t2thx\" (UID: \"d132606d-b1ae-4cb7-865e-e1065c2eb4a9\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-t2thx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.141275 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/02c3b11b-dbc4-445d-94fa-95c5ccc03cb8-profile-collector-cert\") pod \"catalog-operator-68c6474976-j8dcx\" (UID: \"02c3b11b-dbc4-445d-94fa-95c5ccc03cb8\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-j8dcx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.141291 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/07ad02bb-f06a-4619-87d1-112c78642603-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-7fl6h\" (UID: \"07ad02bb-f06a-4619-87d1-112c78642603\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-7fl6h" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.141307 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/51e27630-d5c1-4eed-8232-3698ab99fc19-serving-cert\") pod \"route-controller-manager-6576b87f9c-qs8d6\" (UID: \"51e27630-d5c1-4eed-8232-3698ab99fc19\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qs8d6" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.141322 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/86b5dca2-989f-4b6b-983b-87763ce0d469-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-fz7gx\" (UID: \"86b5dca2-989f-4b6b-983b-87763ce0d469\") " pod="openshift-authentication/oauth-openshift-558db77b4-fz7gx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.141336 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6nqck\" (UniqueName: \"kubernetes.io/projected/c7fe95fc-b7d1-4ad1-8acf-e7ee4564babe-kube-api-access-6nqck\") pod \"dns-default-dlgmw\" (UID: \"c7fe95fc-b7d1-4ad1-8acf-e7ee4564babe\") " pod="openshift-dns/dns-default-dlgmw" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.141350 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/94e71514-0af6-4254-8815-c933ade6c9da-registry-tls\") pod \"image-registry-697d97f7c8-wxlxx\" (UID: \"94e71514-0af6-4254-8815-c933ade6c9da\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxlxx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.141365 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nz5wq\" (UniqueName: \"kubernetes.io/projected/26739e4d-344c-4e25-bd0a-019f11456897-kube-api-access-nz5wq\") pod \"csi-hostpathplugin-jgvgl\" (UID: \"26739e4d-344c-4e25-bd0a-019f11456897\") " pod="hostpath-provisioner/csi-hostpathplugin-jgvgl" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.141380 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/7a5d2d02-347e-4799-a28f-f303d51758d0-auth-proxy-config\") pod \"machine-approver-56656f9798-hzhhs\" (UID: \"7a5d2d02-347e-4799-a28f-f303d51758d0\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-hzhhs" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.141416 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/6a0a5723-a022-41f3-9222-c62cac06397e-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-ct4lp\" (UID: \"6a0a5723-a022-41f3-9222-c62cac06397e\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-ct4lp" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.141431 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/23d7bb25-5d2c-4530-86dd-25a956953a4f-apiservice-cert\") pod \"packageserver-d55dfcdfc-cslqn\" (UID: \"23d7bb25-5d2c-4530-86dd-25a956953a4f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-cslqn" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.141458 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/16be0b45-eab8-4d0a-823b-5d877031dba5-profile-collector-cert\") pod \"olm-operator-6b444d44fb-cq5sn\" (UID: \"16be0b45-eab8-4d0a-823b-5d877031dba5\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-cq5sn" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.141474 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nk6sw\" (UniqueName: \"kubernetes.io/projected/b79b805a-7ab1-467f-90a5-ac7944cc564e-kube-api-access-nk6sw\") pod \"cluster-image-registry-operator-dc59b4c8b-2tjlq\" (UID: \"b79b805a-7ab1-467f-90a5-ac7944cc564e\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2tjlq" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.141491 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/74172695-d17f-4a47-b013-9e025e534916-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-s7wn9\" (UID: \"74172695-d17f-4a47-b013-9e025e534916\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-s7wn9" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.141505 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/26739e4d-344c-4e25-bd0a-019f11456897-csi-data-dir\") pod \"csi-hostpathplugin-jgvgl\" (UID: \"26739e4d-344c-4e25-bd0a-019f11456897\") " pod="hostpath-provisioner/csi-hostpathplugin-jgvgl" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.141520 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7a5d2d02-347e-4799-a28f-f303d51758d0-config\") pod \"machine-approver-56656f9798-hzhhs\" (UID: \"7a5d2d02-347e-4799-a28f-f303d51758d0\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-hzhhs" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.141537 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tkwb6\" (UniqueName: \"kubernetes.io/projected/c117483a-fe5c-423b-995e-9c2e7f59566a-kube-api-access-tkwb6\") pod \"control-plane-machine-set-operator-78cbb6b69f-r65dm\" (UID: \"c117483a-fe5c-423b-995e-9c2e7f59566a\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-r65dm" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.141552 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pwmbq\" (UniqueName: \"kubernetes.io/projected/7a5d2d02-347e-4799-a28f-f303d51758d0-kube-api-access-pwmbq\") pod \"machine-approver-56656f9798-hzhhs\" (UID: \"7a5d2d02-347e-4799-a28f-f303d51758d0\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-hzhhs" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.141566 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4b22a7aa-bcc0-4876-a374-a9943a6040bd-service-ca-bundle\") pod \"router-default-5444994796-bblkp\" (UID: \"4b22a7aa-bcc0-4876-a374-a9943a6040bd\") " pod="openshift-ingress/router-default-5444994796-bblkp" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.141624 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/13490666-4efb-492b-be88-1c6fde3f378e-config\") pod \"service-ca-operator-777779d784-c5xks\" (UID: \"13490666-4efb-492b-be88-1c6fde3f378e\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-c5xks" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.141641 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/63c09cd8-12ad-40a1-9877-3fc1500acb7a-images\") pod \"machine-config-operator-74547568cd-kqtr5\" (UID: \"63c09cd8-12ad-40a1-9877-3fc1500acb7a\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kqtr5" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.141658 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h5dpw\" (UniqueName: \"kubernetes.io/projected/faede743-203b-4bd1-b17a-776a4744bdf1-kube-api-access-h5dpw\") pod \"ingress-canary-b7tsx\" (UID: \"faede743-203b-4bd1-b17a-776a4744bdf1\") " pod="openshift-ingress-canary/ingress-canary-b7tsx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.141672 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zcznm\" (UniqueName: \"kubernetes.io/projected/94e71514-0af6-4254-8815-c933ade6c9da-kube-api-access-zcznm\") pod \"image-registry-697d97f7c8-wxlxx\" (UID: \"94e71514-0af6-4254-8815-c933ade6c9da\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxlxx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.141687 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/86b5dca2-989f-4b6b-983b-87763ce0d469-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-fz7gx\" (UID: \"86b5dca2-989f-4b6b-983b-87763ce0d469\") " pod="openshift-authentication/oauth-openshift-558db77b4-fz7gx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.141703 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/63c09cd8-12ad-40a1-9877-3fc1500acb7a-proxy-tls\") pod \"machine-config-operator-74547568cd-kqtr5\" (UID: \"63c09cd8-12ad-40a1-9877-3fc1500acb7a\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kqtr5" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.141719 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/c117483a-fe5c-423b-995e-9c2e7f59566a-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-r65dm\" (UID: \"c117483a-fe5c-423b-995e-9c2e7f59566a\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-r65dm" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.141737 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/86b5dca2-989f-4b6b-983b-87763ce0d469-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-fz7gx\" (UID: \"86b5dca2-989f-4b6b-983b-87763ce0d469\") " pod="openshift-authentication/oauth-openshift-558db77b4-fz7gx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.141752 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kxstx\" (UniqueName: \"kubernetes.io/projected/d93236c5-17ab-4cd0-88a3-f4b3479067e5-kube-api-access-kxstx\") pod \"machine-config-controller-84d6567774-9tldv\" (UID: \"d93236c5-17ab-4cd0-88a3-f4b3479067e5\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-9tldv" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.141768 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/024ed20d-cf91-4660-998b-3d06d3464607-config\") pod \"etcd-operator-b45778765-cpgg4\" (UID: \"024ed20d-cf91-4660-998b-3d06d3464607\") " pod="openshift-etcd-operator/etcd-operator-b45778765-cpgg4" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.141782 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5dab0fba-f35e-42bd-9a7e-913b70c2dd2b-trusted-ca\") pod \"ingress-operator-5b745b69d9-n8lql\" (UID: \"5dab0fba-f35e-42bd-9a7e-913b70c2dd2b\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-n8lql" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.141796 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4b22a7aa-bcc0-4876-a374-a9943a6040bd-metrics-certs\") pod \"router-default-5444994796-bblkp\" (UID: \"4b22a7aa-bcc0-4876-a374-a9943a6040bd\") " pod="openshift-ingress/router-default-5444994796-bblkp" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.141811 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/d93236c5-17ab-4cd0-88a3-f4b3479067e5-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-9tldv\" (UID: \"d93236c5-17ab-4cd0-88a3-f4b3479067e5\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-9tldv" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.141826 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/b79b805a-7ab1-467f-90a5-ac7944cc564e-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-2tjlq\" (UID: \"b79b805a-7ab1-467f-90a5-ac7944cc564e\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2tjlq" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.141846 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pmsvd\" (UniqueName: \"kubernetes.io/projected/3358782c-3435-4c68-9f9c-c1040a6ada32-kube-api-access-pmsvd\") pod \"service-ca-9c57cc56f-rdcbg\" (UID: \"3358782c-3435-4c68-9f9c-c1040a6ada32\") " pod="openshift-service-ca/service-ca-9c57cc56f-rdcbg" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.141860 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/c7fe95fc-b7d1-4ad1-8acf-e7ee4564babe-metrics-tls\") pod \"dns-default-dlgmw\" (UID: \"c7fe95fc-b7d1-4ad1-8acf-e7ee4564babe\") " pod="openshift-dns/dns-default-dlgmw" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.141877 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c873221f-c31d-411a-bbf5-c59e45b473a5-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-5lmnb\" (UID: \"c873221f-c31d-411a-bbf5-c59e45b473a5\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5lmnb" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.141892 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/7a5d2d02-347e-4799-a28f-f303d51758d0-machine-approver-tls\") pod \"machine-approver-56656f9798-hzhhs\" (UID: \"7a5d2d02-347e-4799-a28f-f303d51758d0\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-hzhhs" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.141905 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/51e27630-d5c1-4eed-8232-3698ab99fc19-client-ca\") pod \"route-controller-manager-6576b87f9c-qs8d6\" (UID: \"51e27630-d5c1-4eed-8232-3698ab99fc19\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qs8d6" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.141921 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/86b5dca2-989f-4b6b-983b-87763ce0d469-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-fz7gx\" (UID: \"86b5dca2-989f-4b6b-983b-87763ce0d469\") " pod="openshift-authentication/oauth-openshift-558db77b4-fz7gx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.141935 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/024ed20d-cf91-4660-998b-3d06d3464607-etcd-ca\") pod \"etcd-operator-b45778765-cpgg4\" (UID: \"024ed20d-cf91-4660-998b-3d06d3464607\") " pod="openshift-etcd-operator/etcd-operator-b45778765-cpgg4" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.141950 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d132606d-b1ae-4cb7-865e-e1065c2eb4a9-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-t2thx\" (UID: \"d132606d-b1ae-4cb7-865e-e1065c2eb4a9\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-t2thx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.141996 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lk25w\" (UniqueName: \"kubernetes.io/projected/920ef22d-4a57-455c-a30f-c6d5f3a419e2-kube-api-access-lk25w\") pod \"machine-config-server-5s88g\" (UID: \"920ef22d-4a57-455c-a30f-c6d5f3a419e2\") " pod="openshift-machine-config-operator/machine-config-server-5s88g" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.142015 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/94e71514-0af6-4254-8815-c933ade6c9da-trusted-ca\") pod \"image-registry-697d97f7c8-wxlxx\" (UID: \"94e71514-0af6-4254-8815-c933ade6c9da\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxlxx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.142030 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/26739e4d-344c-4e25-bd0a-019f11456897-registration-dir\") pod \"csi-hostpathplugin-jgvgl\" (UID: \"26739e4d-344c-4e25-bd0a-019f11456897\") " pod="hostpath-provisioner/csi-hostpathplugin-jgvgl" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.142046 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/4b22a7aa-bcc0-4876-a374-a9943a6040bd-default-certificate\") pod \"router-default-5444994796-bblkp\" (UID: \"4b22a7aa-bcc0-4876-a374-a9943a6040bd\") " pod="openshift-ingress/router-default-5444994796-bblkp" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.142060 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/23d7bb25-5d2c-4530-86dd-25a956953a4f-webhook-cert\") pod \"packageserver-d55dfcdfc-cslqn\" (UID: \"23d7bb25-5d2c-4530-86dd-25a956953a4f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-cslqn" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.142074 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/09bd0b9c-46fb-433f-bd30-014c69c90d9f-secret-volume\") pod \"collect-profiles-29483220-6r4fz\" (UID: \"09bd0b9c-46fb-433f-bd30-014c69c90d9f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29483220-6r4fz" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.142088 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-md94v\" (UniqueName: \"kubernetes.io/projected/02c3b11b-dbc4-445d-94fa-95c5ccc03cb8-kube-api-access-md94v\") pod \"catalog-operator-68c6474976-j8dcx\" (UID: \"02c3b11b-dbc4-445d-94fa-95c5ccc03cb8\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-j8dcx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.142105 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/86b5dca2-989f-4b6b-983b-87763ce0d469-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-fz7gx\" (UID: \"86b5dca2-989f-4b6b-983b-87763ce0d469\") " pod="openshift-authentication/oauth-openshift-558db77b4-fz7gx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.142119 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jsn9x\" (UniqueName: \"kubernetes.io/projected/2279132e-c274-424b-8c8b-80ae5bd7dfe5-kube-api-access-jsn9x\") pod \"downloads-7954f5f757-8g4qk\" (UID: \"2279132e-c274-424b-8c8b-80ae5bd7dfe5\") " pod="openshift-console/downloads-7954f5f757-8g4qk" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.142131 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/4b22a7aa-bcc0-4876-a374-a9943a6040bd-stats-auth\") pod \"router-default-5444994796-bblkp\" (UID: \"4b22a7aa-bcc0-4876-a374-a9943a6040bd\") " pod="openshift-ingress/router-default-5444994796-bblkp" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.142147 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/16be0b45-eab8-4d0a-823b-5d877031dba5-srv-cert\") pod \"olm-operator-6b444d44fb-cq5sn\" (UID: \"16be0b45-eab8-4d0a-823b-5d877031dba5\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-cq5sn" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.142160 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qrlgj\" (UniqueName: \"kubernetes.io/projected/13490666-4efb-492b-be88-1c6fde3f378e-kube-api-access-qrlgj\") pod \"service-ca-operator-777779d784-c5xks\" (UID: \"13490666-4efb-492b-be88-1c6fde3f378e\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-c5xks" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.142175 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/94e71514-0af6-4254-8815-c933ade6c9da-ca-trust-extracted\") pod \"image-registry-697d97f7c8-wxlxx\" (UID: \"94e71514-0af6-4254-8815-c933ade6c9da\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxlxx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.142191 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xgr59\" (UniqueName: \"kubernetes.io/projected/024ed20d-cf91-4660-998b-3d06d3464607-kube-api-access-xgr59\") pod \"etcd-operator-b45778765-cpgg4\" (UID: \"024ed20d-cf91-4660-998b-3d06d3464607\") " pod="openshift-etcd-operator/etcd-operator-b45778765-cpgg4" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.142217 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/51e27630-d5c1-4eed-8232-3698ab99fc19-config\") pod \"route-controller-manager-6576b87f9c-qs8d6\" (UID: \"51e27630-d5c1-4eed-8232-3698ab99fc19\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qs8d6" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.142232 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d132606d-b1ae-4cb7-865e-e1065c2eb4a9-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-t2thx\" (UID: \"d132606d-b1ae-4cb7-865e-e1065c2eb4a9\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-t2thx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.142247 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q27ql\" (UniqueName: \"kubernetes.io/projected/5dab0fba-f35e-42bd-9a7e-913b70c2dd2b-kube-api-access-q27ql\") pod \"ingress-operator-5b745b69d9-n8lql\" (UID: \"5dab0fba-f35e-42bd-9a7e-913b70c2dd2b\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-n8lql" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.142261 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/26739e4d-344c-4e25-bd0a-019f11456897-plugins-dir\") pod \"csi-hostpathplugin-jgvgl\" (UID: \"26739e4d-344c-4e25-bd0a-019f11456897\") " pod="hostpath-provisioner/csi-hostpathplugin-jgvgl" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.142278 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h6sxp\" (UniqueName: \"kubernetes.io/projected/16be0b45-eab8-4d0a-823b-5d877031dba5-kube-api-access-h6sxp\") pod \"olm-operator-6b444d44fb-cq5sn\" (UID: \"16be0b45-eab8-4d0a-823b-5d877031dba5\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-cq5sn" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.142293 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/13490666-4efb-492b-be88-1c6fde3f378e-serving-cert\") pod \"service-ca-operator-777779d784-c5xks\" (UID: \"13490666-4efb-492b-be88-1c6fde3f378e\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-c5xks" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.142308 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5ksj6\" (UniqueName: \"kubernetes.io/projected/51e27630-d5c1-4eed-8232-3698ab99fc19-kube-api-access-5ksj6\") pod \"route-controller-manager-6576b87f9c-qs8d6\" (UID: \"51e27630-d5c1-4eed-8232-3698ab99fc19\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qs8d6" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.142323 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fh969\" (UniqueName: \"kubernetes.io/projected/74172695-d17f-4a47-b013-9e025e534916-kube-api-access-fh969\") pod \"kube-storage-version-migrator-operator-b67b599dd-s7wn9\" (UID: \"74172695-d17f-4a47-b013-9e025e534916\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-s7wn9" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.142336 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/5dab0fba-f35e-42bd-9a7e-913b70c2dd2b-bound-sa-token\") pod \"ingress-operator-5b745b69d9-n8lql\" (UID: \"5dab0fba-f35e-42bd-9a7e-913b70c2dd2b\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-n8lql" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.142350 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/b79b805a-7ab1-467f-90a5-ac7944cc564e-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-2tjlq\" (UID: \"b79b805a-7ab1-467f-90a5-ac7944cc564e\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2tjlq" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.142364 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c873221f-c31d-411a-bbf5-c59e45b473a5-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-5lmnb\" (UID: \"c873221f-c31d-411a-bbf5-c59e45b473a5\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5lmnb" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.142379 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s4hmd\" (UniqueName: \"kubernetes.io/projected/63c09cd8-12ad-40a1-9877-3fc1500acb7a-kube-api-access-s4hmd\") pod \"machine-config-operator-74547568cd-kqtr5\" (UID: \"63c09cd8-12ad-40a1-9877-3fc1500acb7a\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kqtr5" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.142397 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/024ed20d-cf91-4660-998b-3d06d3464607-etcd-service-ca\") pod \"etcd-operator-b45778765-cpgg4\" (UID: \"024ed20d-cf91-4660-998b-3d06d3464607\") " pod="openshift-etcd-operator/etcd-operator-b45778765-cpgg4" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.142411 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/26739e4d-344c-4e25-bd0a-019f11456897-socket-dir\") pod \"csi-hostpathplugin-jgvgl\" (UID: \"26739e4d-344c-4e25-bd0a-019f11456897\") " pod="hostpath-provisioner/csi-hostpathplugin-jgvgl" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.142427 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/86dc1658-d302-449f-9ae3-d7bc5e795c57-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-bkkzh\" (UID: \"86dc1658-d302-449f-9ae3-d7bc5e795c57\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-bkkzh" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.142442 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/3358782c-3435-4c68-9f9c-c1040a6ada32-signing-key\") pod \"service-ca-9c57cc56f-rdcbg\" (UID: \"3358782c-3435-4c68-9f9c-c1040a6ada32\") " pod="openshift-service-ca/service-ca-9c57cc56f-rdcbg" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.142456 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/920ef22d-4a57-455c-a30f-c6d5f3a419e2-node-bootstrap-token\") pod \"machine-config-server-5s88g\" (UID: \"920ef22d-4a57-455c-a30f-c6d5f3a419e2\") " pod="openshift-machine-config-operator/machine-config-server-5s88g" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.142472 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c873221f-c31d-411a-bbf5-c59e45b473a5-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-5lmnb\" (UID: \"c873221f-c31d-411a-bbf5-c59e45b473a5\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5lmnb" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.142486 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/26739e4d-344c-4e25-bd0a-019f11456897-mountpoint-dir\") pod \"csi-hostpathplugin-jgvgl\" (UID: \"26739e4d-344c-4e25-bd0a-019f11456897\") " pod="hostpath-provisioner/csi-hostpathplugin-jgvgl" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.142502 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/024ed20d-cf91-4660-998b-3d06d3464607-serving-cert\") pod \"etcd-operator-b45778765-cpgg4\" (UID: \"024ed20d-cf91-4660-998b-3d06d3464607\") " pod="openshift-etcd-operator/etcd-operator-b45778765-cpgg4" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.142517 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/14d99a1f-fca3-421a-8b47-ab4cda2f77fc-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-gjhp5\" (UID: \"14d99a1f-fca3-421a-8b47-ab4cda2f77fc\") " pod="openshift-marketplace/marketplace-operator-79b997595-gjhp5" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.142532 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/23d7bb25-5d2c-4530-86dd-25a956953a4f-tmpfs\") pod \"packageserver-d55dfcdfc-cslqn\" (UID: \"23d7bb25-5d2c-4530-86dd-25a956953a4f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-cslqn" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.142546 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/920ef22d-4a57-455c-a30f-c6d5f3a419e2-certs\") pod \"machine-config-server-5s88g\" (UID: \"920ef22d-4a57-455c-a30f-c6d5f3a419e2\") " pod="openshift-machine-config-operator/machine-config-server-5s88g" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.142560 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/3358782c-3435-4c68-9f9c-c1040a6ada32-signing-cabundle\") pod \"service-ca-9c57cc56f-rdcbg\" (UID: \"3358782c-3435-4c68-9f9c-c1040a6ada32\") " pod="openshift-service-ca/service-ca-9c57cc56f-rdcbg" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.142577 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kjjkx\" (UniqueName: \"kubernetes.io/projected/14d99a1f-fca3-421a-8b47-ab4cda2f77fc-kube-api-access-kjjkx\") pod \"marketplace-operator-79b997595-gjhp5\" (UID: \"14d99a1f-fca3-421a-8b47-ab4cda2f77fc\") " pod="openshift-marketplace/marketplace-operator-79b997595-gjhp5" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.142600 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/94e71514-0af6-4254-8815-c933ade6c9da-installation-pull-secrets\") pod \"image-registry-697d97f7c8-wxlxx\" (UID: \"94e71514-0af6-4254-8815-c933ade6c9da\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxlxx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.142621 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b79b805a-7ab1-467f-90a5-ac7944cc564e-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-2tjlq\" (UID: \"b79b805a-7ab1-467f-90a5-ac7944cc564e\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2tjlq" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.142637 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/94e71514-0af6-4254-8815-c933ade6c9da-bound-sa-token\") pod \"image-registry-697d97f7c8-wxlxx\" (UID: \"94e71514-0af6-4254-8815-c933ade6c9da\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxlxx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.142653 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/86b5dca2-989f-4b6b-983b-87763ce0d469-audit-policies\") pod \"oauth-openshift-558db77b4-fz7gx\" (UID: \"86b5dca2-989f-4b6b-983b-87763ce0d469\") " pod="openshift-authentication/oauth-openshift-558db77b4-fz7gx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.142669 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/37d2c136-5241-41da-a630-3d70273f1670-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-cnr8w\" (UID: \"37d2c136-5241-41da-a630-3d70273f1670\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-cnr8w" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.142683 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2552x\" (UniqueName: \"kubernetes.io/projected/27b93806-10a5-46ba-8a81-20b05fd990c1-kube-api-access-2552x\") pod \"migrator-59844c95c7-j6bst\" (UID: \"27b93806-10a5-46ba-8a81-20b05fd990c1\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-j6bst" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.142698 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/07ad02bb-f06a-4619-87d1-112c78642603-config\") pod \"kube-controller-manager-operator-78b949d7b-7fl6h\" (UID: \"07ad02bb-f06a-4619-87d1-112c78642603\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-7fl6h" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.142712 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6q9tc\" (UniqueName: \"kubernetes.io/projected/4b22a7aa-bcc0-4876-a374-a9943a6040bd-kube-api-access-6q9tc\") pod \"router-default-5444994796-bblkp\" (UID: \"4b22a7aa-bcc0-4876-a374-a9943a6040bd\") " pod="openshift-ingress/router-default-5444994796-bblkp" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.142734 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/7352d717-b2a3-4cd2-9e35-3ae2fe2d7b34-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-27d2q\" (UID: \"7352d717-b2a3-4cd2-9e35-3ae2fe2d7b34\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-27d2q" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.142749 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6qg69\" (UniqueName: \"kubernetes.io/projected/ae7ce20e-9b65-4dd4-9e00-d96cf91803cb-kube-api-access-6qg69\") pod \"dns-operator-744455d44c-f6c7b\" (UID: \"ae7ce20e-9b65-4dd4-9e00-d96cf91803cb\") " pod="openshift-dns-operator/dns-operator-744455d44c-f6c7b" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.142765 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/5dab0fba-f35e-42bd-9a7e-913b70c2dd2b-metrics-tls\") pod \"ingress-operator-5b745b69d9-n8lql\" (UID: \"5dab0fba-f35e-42bd-9a7e-913b70c2dd2b\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-n8lql" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.142778 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/ae7ce20e-9b65-4dd4-9e00-d96cf91803cb-metrics-tls\") pod \"dns-operator-744455d44c-f6c7b\" (UID: \"ae7ce20e-9b65-4dd4-9e00-d96cf91803cb\") " pod="openshift-dns-operator/dns-operator-744455d44c-f6c7b" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.142794 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/86b5dca2-989f-4b6b-983b-87763ce0d469-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-fz7gx\" (UID: \"86b5dca2-989f-4b6b-983b-87763ce0d469\") " pod="openshift-authentication/oauth-openshift-558db77b4-fz7gx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.142809 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/86b5dca2-989f-4b6b-983b-87763ce0d469-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-fz7gx\" (UID: \"86b5dca2-989f-4b6b-983b-87763ce0d469\") " pod="openshift-authentication/oauth-openshift-558db77b4-fz7gx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.142825 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nqqww\" (UniqueName: \"kubernetes.io/projected/09bd0b9c-46fb-433f-bd30-014c69c90d9f-kube-api-access-nqqww\") pod \"collect-profiles-29483220-6r4fz\" (UID: \"09bd0b9c-46fb-433f-bd30-014c69c90d9f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29483220-6r4fz" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.142847 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pczg5\" (UniqueName: \"kubernetes.io/projected/86dc1658-d302-449f-9ae3-d7bc5e795c57-kube-api-access-pczg5\") pod \"package-server-manager-789f6589d5-bkkzh\" (UID: \"86dc1658-d302-449f-9ae3-d7bc5e795c57\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-bkkzh" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.142862 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/14d99a1f-fca3-421a-8b47-ab4cda2f77fc-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-gjhp5\" (UID: \"14d99a1f-fca3-421a-8b47-ab4cda2f77fc\") " pod="openshift-marketplace/marketplace-operator-79b997595-gjhp5" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.142879 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/94e71514-0af6-4254-8815-c933ade6c9da-registry-certificates\") pod \"image-registry-697d97f7c8-wxlxx\" (UID: \"94e71514-0af6-4254-8815-c933ade6c9da\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxlxx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.142893 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/86b5dca2-989f-4b6b-983b-87763ce0d469-audit-dir\") pod \"oauth-openshift-558db77b4-fz7gx\" (UID: \"86b5dca2-989f-4b6b-983b-87763ce0d469\") " pod="openshift-authentication/oauth-openshift-558db77b4-fz7gx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.142908 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/024ed20d-cf91-4660-998b-3d06d3464607-etcd-client\") pod \"etcd-operator-b45778765-cpgg4\" (UID: \"024ed20d-cf91-4660-998b-3d06d3464607\") " pod="openshift-etcd-operator/etcd-operator-b45778765-cpgg4" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.142923 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/09bd0b9c-46fb-433f-bd30-014c69c90d9f-config-volume\") pod \"collect-profiles-29483220-6r4fz\" (UID: \"09bd0b9c-46fb-433f-bd30-014c69c90d9f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29483220-6r4fz" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.143611 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/09bd0b9c-46fb-433f-bd30-014c69c90d9f-config-volume\") pod \"collect-profiles-29483220-6r4fz\" (UID: \"09bd0b9c-46fb-433f-bd30-014c69c90d9f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29483220-6r4fz" Jan 21 11:13:12 crc kubenswrapper[4824]: E0121 11:13:12.143680 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-21 11:13:12.643668409 +0000 UTC m=+134.936697701 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.144181 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c873221f-c31d-411a-bbf5-c59e45b473a5-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-5lmnb\" (UID: \"c873221f-c31d-411a-bbf5-c59e45b473a5\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5lmnb" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.144774 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/86b5dca2-989f-4b6b-983b-87763ce0d469-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-fz7gx\" (UID: \"86b5dca2-989f-4b6b-983b-87763ce0d469\") " pod="openshift-authentication/oauth-openshift-558db77b4-fz7gx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.144938 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/024ed20d-cf91-4660-998b-3d06d3464607-config\") pod \"etcd-operator-b45778765-cpgg4\" (UID: \"024ed20d-cf91-4660-998b-3d06d3464607\") " pod="openshift-etcd-operator/etcd-operator-b45778765-cpgg4" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.145168 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/07ad02bb-f06a-4619-87d1-112c78642603-config\") pod \"kube-controller-manager-operator-78b949d7b-7fl6h\" (UID: \"07ad02bb-f06a-4619-87d1-112c78642603\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-7fl6h" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.146162 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/13490666-4efb-492b-be88-1c6fde3f378e-config\") pod \"service-ca-operator-777779d784-c5xks\" (UID: \"13490666-4efb-492b-be88-1c6fde3f378e\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-c5xks" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.146410 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4b22a7aa-bcc0-4876-a374-a9943a6040bd-service-ca-bundle\") pod \"router-default-5444994796-bblkp\" (UID: \"4b22a7aa-bcc0-4876-a374-a9943a6040bd\") " pod="openshift-ingress/router-default-5444994796-bblkp" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.146579 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/86b5dca2-989f-4b6b-983b-87763ce0d469-audit-dir\") pod \"oauth-openshift-558db77b4-fz7gx\" (UID: \"86b5dca2-989f-4b6b-983b-87763ce0d469\") " pod="openshift-authentication/oauth-openshift-558db77b4-fz7gx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.147131 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/14d99a1f-fca3-421a-8b47-ab4cda2f77fc-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-gjhp5\" (UID: \"14d99a1f-fca3-421a-8b47-ab4cda2f77fc\") " pod="openshift-marketplace/marketplace-operator-79b997595-gjhp5" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.147306 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7a5d2d02-347e-4799-a28f-f303d51758d0-config\") pod \"machine-approver-56656f9798-hzhhs\" (UID: \"7a5d2d02-347e-4799-a28f-f303d51758d0\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-hzhhs" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.148487 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/51e27630-d5c1-4eed-8232-3698ab99fc19-config\") pod \"route-controller-manager-6576b87f9c-qs8d6\" (UID: \"51e27630-d5c1-4eed-8232-3698ab99fc19\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qs8d6" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.148750 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/4b22a7aa-bcc0-4876-a374-a9943a6040bd-default-certificate\") pod \"router-default-5444994796-bblkp\" (UID: \"4b22a7aa-bcc0-4876-a374-a9943a6040bd\") " pod="openshift-ingress/router-default-5444994796-bblkp" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.149072 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/86b5dca2-989f-4b6b-983b-87763ce0d469-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-fz7gx\" (UID: \"86b5dca2-989f-4b6b-983b-87763ce0d469\") " pod="openshift-authentication/oauth-openshift-558db77b4-fz7gx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.149601 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/94e71514-0af6-4254-8815-c933ade6c9da-registry-certificates\") pod \"image-registry-697d97f7c8-wxlxx\" (UID: \"94e71514-0af6-4254-8815-c933ade6c9da\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxlxx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.150817 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b79b805a-7ab1-467f-90a5-ac7944cc564e-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-2tjlq\" (UID: \"b79b805a-7ab1-467f-90a5-ac7944cc564e\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2tjlq" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.152791 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/024ed20d-cf91-4660-998b-3d06d3464607-etcd-ca\") pod \"etcd-operator-b45778765-cpgg4\" (UID: \"024ed20d-cf91-4660-998b-3d06d3464607\") " pod="openshift-etcd-operator/etcd-operator-b45778765-cpgg4" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.153053 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/94e71514-0af6-4254-8815-c933ade6c9da-ca-trust-extracted\") pod \"image-registry-697d97f7c8-wxlxx\" (UID: \"94e71514-0af6-4254-8815-c933ade6c9da\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxlxx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.153128 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4b22a7aa-bcc0-4876-a374-a9943a6040bd-metrics-certs\") pod \"router-default-5444994796-bblkp\" (UID: \"4b22a7aa-bcc0-4876-a374-a9943a6040bd\") " pod="openshift-ingress/router-default-5444994796-bblkp" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.153214 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/94e71514-0af6-4254-8815-c933ade6c9da-trusted-ca\") pod \"image-registry-697d97f7c8-wxlxx\" (UID: \"94e71514-0af6-4254-8815-c933ade6c9da\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxlxx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.153456 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/b79b805a-7ab1-467f-90a5-ac7944cc564e-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-2tjlq\" (UID: \"b79b805a-7ab1-467f-90a5-ac7944cc564e\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2tjlq" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.153518 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/37d2c136-5241-41da-a630-3d70273f1670-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-cnr8w\" (UID: \"37d2c136-5241-41da-a630-3d70273f1670\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-cnr8w" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.153884 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/86b5dca2-989f-4b6b-983b-87763ce0d469-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-fz7gx\" (UID: \"86b5dca2-989f-4b6b-983b-87763ce0d469\") " pod="openshift-authentication/oauth-openshift-558db77b4-fz7gx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.154172 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/09bd0b9c-46fb-433f-bd30-014c69c90d9f-secret-volume\") pod \"collect-profiles-29483220-6r4fz\" (UID: \"09bd0b9c-46fb-433f-bd30-014c69c90d9f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29483220-6r4fz" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.154318 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/74172695-d17f-4a47-b013-9e025e534916-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-s7wn9\" (UID: \"74172695-d17f-4a47-b013-9e025e534916\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-s7wn9" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.154360 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5dab0fba-f35e-42bd-9a7e-913b70c2dd2b-trusted-ca\") pod \"ingress-operator-5b745b69d9-n8lql\" (UID: \"5dab0fba-f35e-42bd-9a7e-913b70c2dd2b\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-n8lql" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.154574 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/86b5dca2-989f-4b6b-983b-87763ce0d469-audit-policies\") pod \"oauth-openshift-558db77b4-fz7gx\" (UID: \"86b5dca2-989f-4b6b-983b-87763ce0d469\") " pod="openshift-authentication/oauth-openshift-558db77b4-fz7gx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.154972 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/6a0a5723-a022-41f3-9222-c62cac06397e-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-ct4lp\" (UID: \"6a0a5723-a022-41f3-9222-c62cac06397e\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-ct4lp" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.155046 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/51e27630-d5c1-4eed-8232-3698ab99fc19-client-ca\") pod \"route-controller-manager-6576b87f9c-qs8d6\" (UID: \"51e27630-d5c1-4eed-8232-3698ab99fc19\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qs8d6" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.155284 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/94e71514-0af6-4254-8815-c933ade6c9da-installation-pull-secrets\") pod \"image-registry-697d97f7c8-wxlxx\" (UID: \"94e71514-0af6-4254-8815-c933ade6c9da\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxlxx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.155668 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/16be0b45-eab8-4d0a-823b-5d877031dba5-srv-cert\") pod \"olm-operator-6b444d44fb-cq5sn\" (UID: \"16be0b45-eab8-4d0a-823b-5d877031dba5\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-cq5sn" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.155722 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/7a5d2d02-347e-4799-a28f-f303d51758d0-machine-approver-tls\") pod \"machine-approver-56656f9798-hzhhs\" (UID: \"7a5d2d02-347e-4799-a28f-f303d51758d0\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-hzhhs" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.155725 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/51e27630-d5c1-4eed-8232-3698ab99fc19-serving-cert\") pod \"route-controller-manager-6576b87f9c-qs8d6\" (UID: \"51e27630-d5c1-4eed-8232-3698ab99fc19\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qs8d6" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.155991 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/024ed20d-cf91-4660-998b-3d06d3464607-etcd-client\") pod \"etcd-operator-b45778765-cpgg4\" (UID: \"024ed20d-cf91-4660-998b-3d06d3464607\") " pod="openshift-etcd-operator/etcd-operator-b45778765-cpgg4" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.156032 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/024ed20d-cf91-4660-998b-3d06d3464607-etcd-service-ca\") pod \"etcd-operator-b45778765-cpgg4\" (UID: \"024ed20d-cf91-4660-998b-3d06d3464607\") " pod="openshift-etcd-operator/etcd-operator-b45778765-cpgg4" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.156093 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/86b5dca2-989f-4b6b-983b-87763ce0d469-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-fz7gx\" (UID: \"86b5dca2-989f-4b6b-983b-87763ce0d469\") " pod="openshift-authentication/oauth-openshift-558db77b4-fz7gx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.156242 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/07ad02bb-f06a-4619-87d1-112c78642603-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-7fl6h\" (UID: \"07ad02bb-f06a-4619-87d1-112c78642603\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-7fl6h" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.156345 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/74172695-d17f-4a47-b013-9e025e534916-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-s7wn9\" (UID: \"74172695-d17f-4a47-b013-9e025e534916\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-s7wn9" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.156356 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/86b5dca2-989f-4b6b-983b-87763ce0d469-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-fz7gx\" (UID: \"86b5dca2-989f-4b6b-983b-87763ce0d469\") " pod="openshift-authentication/oauth-openshift-558db77b4-fz7gx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.156826 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/5dab0fba-f35e-42bd-9a7e-913b70c2dd2b-metrics-tls\") pod \"ingress-operator-5b745b69d9-n8lql\" (UID: \"5dab0fba-f35e-42bd-9a7e-913b70c2dd2b\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-n8lql" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.156864 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/86b5dca2-989f-4b6b-983b-87763ce0d469-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-fz7gx\" (UID: \"86b5dca2-989f-4b6b-983b-87763ce0d469\") " pod="openshift-authentication/oauth-openshift-558db77b4-fz7gx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.155460 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/3358782c-3435-4c68-9f9c-c1040a6ada32-signing-cabundle\") pod \"service-ca-9c57cc56f-rdcbg\" (UID: \"3358782c-3435-4c68-9f9c-c1040a6ada32\") " pod="openshift-service-ca/service-ca-9c57cc56f-rdcbg" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.157517 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d132606d-b1ae-4cb7-865e-e1065c2eb4a9-config\") pod \"kube-apiserver-operator-766d6c64bb-t2thx\" (UID: \"d132606d-b1ae-4cb7-865e-e1065c2eb4a9\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-t2thx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.157735 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/7a5d2d02-347e-4799-a28f-f303d51758d0-auth-proxy-config\") pod \"machine-approver-56656f9798-hzhhs\" (UID: \"7a5d2d02-347e-4799-a28f-f303d51758d0\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-hzhhs" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.158090 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/02c3b11b-dbc4-445d-94fa-95c5ccc03cb8-profile-collector-cert\") pod \"catalog-operator-68c6474976-j8dcx\" (UID: \"02c3b11b-dbc4-445d-94fa-95c5ccc03cb8\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-j8dcx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.158439 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/86b5dca2-989f-4b6b-983b-87763ce0d469-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-fz7gx\" (UID: \"86b5dca2-989f-4b6b-983b-87763ce0d469\") " pod="openshift-authentication/oauth-openshift-558db77b4-fz7gx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.158513 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/024ed20d-cf91-4660-998b-3d06d3464607-serving-cert\") pod \"etcd-operator-b45778765-cpgg4\" (UID: \"024ed20d-cf91-4660-998b-3d06d3464607\") " pod="openshift-etcd-operator/etcd-operator-b45778765-cpgg4" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.158829 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/14d99a1f-fca3-421a-8b47-ab4cda2f77fc-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-gjhp5\" (UID: \"14d99a1f-fca3-421a-8b47-ab4cda2f77fc\") " pod="openshift-marketplace/marketplace-operator-79b997595-gjhp5" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.158902 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/94e71514-0af6-4254-8815-c933ade6c9da-registry-tls\") pod \"image-registry-697d97f7c8-wxlxx\" (UID: \"94e71514-0af6-4254-8815-c933ade6c9da\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxlxx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.158926 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/86b5dca2-989f-4b6b-983b-87763ce0d469-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-fz7gx\" (UID: \"86b5dca2-989f-4b6b-983b-87763ce0d469\") " pod="openshift-authentication/oauth-openshift-558db77b4-fz7gx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.159160 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/86b5dca2-989f-4b6b-983b-87763ce0d469-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-fz7gx\" (UID: \"86b5dca2-989f-4b6b-983b-87763ce0d469\") " pod="openshift-authentication/oauth-openshift-558db77b4-fz7gx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.159240 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/7352d717-b2a3-4cd2-9e35-3ae2fe2d7b34-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-27d2q\" (UID: \"7352d717-b2a3-4cd2-9e35-3ae2fe2d7b34\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-27d2q" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.159265 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d132606d-b1ae-4cb7-865e-e1065c2eb4a9-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-t2thx\" (UID: \"d132606d-b1ae-4cb7-865e-e1065c2eb4a9\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-t2thx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.159431 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/37d2c136-5241-41da-a630-3d70273f1670-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-cnr8w\" (UID: \"37d2c136-5241-41da-a630-3d70273f1670\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-cnr8w" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.159478 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/86b5dca2-989f-4b6b-983b-87763ce0d469-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-fz7gx\" (UID: \"86b5dca2-989f-4b6b-983b-87763ce0d469\") " pod="openshift-authentication/oauth-openshift-558db77b4-fz7gx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.159775 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/4b22a7aa-bcc0-4876-a374-a9943a6040bd-stats-auth\") pod \"router-default-5444994796-bblkp\" (UID: \"4b22a7aa-bcc0-4876-a374-a9943a6040bd\") " pod="openshift-ingress/router-default-5444994796-bblkp" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.159928 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/86b5dca2-989f-4b6b-983b-87763ce0d469-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-fz7gx\" (UID: \"86b5dca2-989f-4b6b-983b-87763ce0d469\") " pod="openshift-authentication/oauth-openshift-558db77b4-fz7gx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.160070 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/13490666-4efb-492b-be88-1c6fde3f378e-serving-cert\") pod \"service-ca-operator-777779d784-c5xks\" (UID: \"13490666-4efb-492b-be88-1c6fde3f378e\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-c5xks" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.160117 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/ae7ce20e-9b65-4dd4-9e00-d96cf91803cb-metrics-tls\") pod \"dns-operator-744455d44c-f6c7b\" (UID: \"ae7ce20e-9b65-4dd4-9e00-d96cf91803cb\") " pod="openshift-dns-operator/dns-operator-744455d44c-f6c7b" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.160324 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/16be0b45-eab8-4d0a-823b-5d877031dba5-profile-collector-cert\") pod \"olm-operator-6b444d44fb-cq5sn\" (UID: \"16be0b45-eab8-4d0a-823b-5d877031dba5\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-cq5sn" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.160616 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/02c3b11b-dbc4-445d-94fa-95c5ccc03cb8-srv-cert\") pod \"catalog-operator-68c6474976-j8dcx\" (UID: \"02c3b11b-dbc4-445d-94fa-95c5ccc03cb8\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-j8dcx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.161423 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/3358782c-3435-4c68-9f9c-c1040a6ada32-signing-key\") pod \"service-ca-9c57cc56f-rdcbg\" (UID: \"3358782c-3435-4c68-9f9c-c1040a6ada32\") " pod="openshift-service-ca/service-ca-9c57cc56f-rdcbg" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.161984 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c873221f-c31d-411a-bbf5-c59e45b473a5-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-5lmnb\" (UID: \"c873221f-c31d-411a-bbf5-c59e45b473a5\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5lmnb" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.164375 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/c117483a-fe5c-423b-995e-9c2e7f59566a-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-r65dm\" (UID: \"c117483a-fe5c-423b-995e-9c2e7f59566a\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-r65dm" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.164760 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/86dc1658-d302-449f-9ae3-d7bc5e795c57-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-bkkzh\" (UID: \"86dc1658-d302-449f-9ae3-d7bc5e795c57\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-bkkzh" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.168523 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-7grz2" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.182629 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-7s7wx"] Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.186288 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nk6sw\" (UniqueName: \"kubernetes.io/projected/b79b805a-7ab1-467f-90a5-ac7944cc564e-kube-api-access-nk6sw\") pod \"cluster-image-registry-operator-dc59b4c8b-2tjlq\" (UID: \"b79b805a-7ab1-467f-90a5-ac7944cc564e\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2tjlq" Jan 21 11:13:12 crc kubenswrapper[4824]: W0121 11:13:12.186371 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0475fd50_10a8_4195_8643_9408478e953c.slice/crio-75c7fa16af4bca084c4cc4bde49251029592fd9bed3decc1ddef1b9a08515ea4 WatchSource:0}: Error finding container 75c7fa16af4bca084c4cc4bde49251029592fd9bed3decc1ddef1b9a08515ea4: Status 404 returned error can't find the container with id 75c7fa16af4bca084c4cc4bde49251029592fd9bed3decc1ddef1b9a08515ea4 Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.206338 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hlp6l\" (UniqueName: \"kubernetes.io/projected/6a0a5723-a022-41f3-9222-c62cac06397e-kube-api-access-hlp6l\") pod \"cluster-samples-operator-665b6dd947-ct4lp\" (UID: \"6a0a5723-a022-41f3-9222-c62cac06397e\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-ct4lp" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.228171 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2552x\" (UniqueName: \"kubernetes.io/projected/27b93806-10a5-46ba-8a81-20b05fd990c1-kube-api-access-2552x\") pod \"migrator-59844c95c7-j6bst\" (UID: \"27b93806-10a5-46ba-8a81-20b05fd990c1\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-j6bst" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.245044 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/26739e4d-344c-4e25-bd0a-019f11456897-mountpoint-dir\") pod \"csi-hostpathplugin-jgvgl\" (UID: \"26739e4d-344c-4e25-bd0a-019f11456897\") " pod="hostpath-provisioner/csi-hostpathplugin-jgvgl" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.245095 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/23d7bb25-5d2c-4530-86dd-25a956953a4f-tmpfs\") pod \"packageserver-d55dfcdfc-cslqn\" (UID: \"23d7bb25-5d2c-4530-86dd-25a956953a4f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-cslqn" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.245130 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/920ef22d-4a57-455c-a30f-c6d5f3a419e2-certs\") pod \"machine-config-server-5s88g\" (UID: \"920ef22d-4a57-455c-a30f-c6d5f3a419e2\") " pod="openshift-machine-config-operator/machine-config-server-5s88g" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.245285 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxlxx\" (UID: \"94e71514-0af6-4254-8815-c933ade6c9da\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxlxx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.245309 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/d93236c5-17ab-4cd0-88a3-f4b3479067e5-proxy-tls\") pod \"machine-config-controller-84d6567774-9tldv\" (UID: \"d93236c5-17ab-4cd0-88a3-f4b3479067e5\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-9tldv" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.245331 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c7fe95fc-b7d1-4ad1-8acf-e7ee4564babe-config-volume\") pod \"dns-default-dlgmw\" (UID: \"c7fe95fc-b7d1-4ad1-8acf-e7ee4564babe\") " pod="openshift-dns/dns-default-dlgmw" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.245372 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/63c09cd8-12ad-40a1-9877-3fc1500acb7a-auth-proxy-config\") pod \"machine-config-operator-74547568cd-kqtr5\" (UID: \"63c09cd8-12ad-40a1-9877-3fc1500acb7a\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kqtr5" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.245400 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/faede743-203b-4bd1-b17a-776a4744bdf1-cert\") pod \"ingress-canary-b7tsx\" (UID: \"faede743-203b-4bd1-b17a-776a4744bdf1\") " pod="openshift-ingress-canary/ingress-canary-b7tsx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.245421 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j48kx\" (UniqueName: \"kubernetes.io/projected/23d7bb25-5d2c-4530-86dd-25a956953a4f-kube-api-access-j48kx\") pod \"packageserver-d55dfcdfc-cslqn\" (UID: \"23d7bb25-5d2c-4530-86dd-25a956953a4f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-cslqn" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.245471 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6nqck\" (UniqueName: \"kubernetes.io/projected/c7fe95fc-b7d1-4ad1-8acf-e7ee4564babe-kube-api-access-6nqck\") pod \"dns-default-dlgmw\" (UID: \"c7fe95fc-b7d1-4ad1-8acf-e7ee4564babe\") " pod="openshift-dns/dns-default-dlgmw" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.245496 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nz5wq\" (UniqueName: \"kubernetes.io/projected/26739e4d-344c-4e25-bd0a-019f11456897-kube-api-access-nz5wq\") pod \"csi-hostpathplugin-jgvgl\" (UID: \"26739e4d-344c-4e25-bd0a-019f11456897\") " pod="hostpath-provisioner/csi-hostpathplugin-jgvgl" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.245520 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/23d7bb25-5d2c-4530-86dd-25a956953a4f-apiservice-cert\") pod \"packageserver-d55dfcdfc-cslqn\" (UID: \"23d7bb25-5d2c-4530-86dd-25a956953a4f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-cslqn" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.245550 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/26739e4d-344c-4e25-bd0a-019f11456897-csi-data-dir\") pod \"csi-hostpathplugin-jgvgl\" (UID: \"26739e4d-344c-4e25-bd0a-019f11456897\") " pod="hostpath-provisioner/csi-hostpathplugin-jgvgl" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.245600 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/63c09cd8-12ad-40a1-9877-3fc1500acb7a-images\") pod \"machine-config-operator-74547568cd-kqtr5\" (UID: \"63c09cd8-12ad-40a1-9877-3fc1500acb7a\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kqtr5" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.245619 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h5dpw\" (UniqueName: \"kubernetes.io/projected/faede743-203b-4bd1-b17a-776a4744bdf1-kube-api-access-h5dpw\") pod \"ingress-canary-b7tsx\" (UID: \"faede743-203b-4bd1-b17a-776a4744bdf1\") " pod="openshift-ingress-canary/ingress-canary-b7tsx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.245641 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/63c09cd8-12ad-40a1-9877-3fc1500acb7a-proxy-tls\") pod \"machine-config-operator-74547568cd-kqtr5\" (UID: \"63c09cd8-12ad-40a1-9877-3fc1500acb7a\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kqtr5" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.245661 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kxstx\" (UniqueName: \"kubernetes.io/projected/d93236c5-17ab-4cd0-88a3-f4b3479067e5-kube-api-access-kxstx\") pod \"machine-config-controller-84d6567774-9tldv\" (UID: \"d93236c5-17ab-4cd0-88a3-f4b3479067e5\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-9tldv" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.245682 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/d93236c5-17ab-4cd0-88a3-f4b3479067e5-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-9tldv\" (UID: \"d93236c5-17ab-4cd0-88a3-f4b3479067e5\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-9tldv" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.245712 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/c7fe95fc-b7d1-4ad1-8acf-e7ee4564babe-metrics-tls\") pod \"dns-default-dlgmw\" (UID: \"c7fe95fc-b7d1-4ad1-8acf-e7ee4564babe\") " pod="openshift-dns/dns-default-dlgmw" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.245735 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lk25w\" (UniqueName: \"kubernetes.io/projected/920ef22d-4a57-455c-a30f-c6d5f3a419e2-kube-api-access-lk25w\") pod \"machine-config-server-5s88g\" (UID: \"920ef22d-4a57-455c-a30f-c6d5f3a419e2\") " pod="openshift-machine-config-operator/machine-config-server-5s88g" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.245755 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/26739e4d-344c-4e25-bd0a-019f11456897-registration-dir\") pod \"csi-hostpathplugin-jgvgl\" (UID: \"26739e4d-344c-4e25-bd0a-019f11456897\") " pod="hostpath-provisioner/csi-hostpathplugin-jgvgl" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.245777 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/23d7bb25-5d2c-4530-86dd-25a956953a4f-webhook-cert\") pod \"packageserver-d55dfcdfc-cslqn\" (UID: \"23d7bb25-5d2c-4530-86dd-25a956953a4f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-cslqn" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.245847 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/26739e4d-344c-4e25-bd0a-019f11456897-plugins-dir\") pod \"csi-hostpathplugin-jgvgl\" (UID: \"26739e4d-344c-4e25-bd0a-019f11456897\") " pod="hostpath-provisioner/csi-hostpathplugin-jgvgl" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.245917 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s4hmd\" (UniqueName: \"kubernetes.io/projected/63c09cd8-12ad-40a1-9877-3fc1500acb7a-kube-api-access-s4hmd\") pod \"machine-config-operator-74547568cd-kqtr5\" (UID: \"63c09cd8-12ad-40a1-9877-3fc1500acb7a\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kqtr5" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.245971 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/26739e4d-344c-4e25-bd0a-019f11456897-socket-dir\") pod \"csi-hostpathplugin-jgvgl\" (UID: \"26739e4d-344c-4e25-bd0a-019f11456897\") " pod="hostpath-provisioner/csi-hostpathplugin-jgvgl" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.245996 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/920ef22d-4a57-455c-a30f-c6d5f3a419e2-node-bootstrap-token\") pod \"machine-config-server-5s88g\" (UID: \"920ef22d-4a57-455c-a30f-c6d5f3a419e2\") " pod="openshift-machine-config-operator/machine-config-server-5s88g" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.246091 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/26739e4d-344c-4e25-bd0a-019f11456897-csi-data-dir\") pod \"csi-hostpathplugin-jgvgl\" (UID: \"26739e4d-344c-4e25-bd0a-019f11456897\") " pod="hostpath-provisioner/csi-hostpathplugin-jgvgl" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.246424 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/26739e4d-344c-4e25-bd0a-019f11456897-registration-dir\") pod \"csi-hostpathplugin-jgvgl\" (UID: \"26739e4d-344c-4e25-bd0a-019f11456897\") " pod="hostpath-provisioner/csi-hostpathplugin-jgvgl" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.246523 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/26739e4d-344c-4e25-bd0a-019f11456897-mountpoint-dir\") pod \"csi-hostpathplugin-jgvgl\" (UID: \"26739e4d-344c-4e25-bd0a-019f11456897\") " pod="hostpath-provisioner/csi-hostpathplugin-jgvgl" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.247264 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/d93236c5-17ab-4cd0-88a3-f4b3479067e5-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-9tldv\" (UID: \"d93236c5-17ab-4cd0-88a3-f4b3479067e5\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-9tldv" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.247416 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/23d7bb25-5d2c-4530-86dd-25a956953a4f-tmpfs\") pod \"packageserver-d55dfcdfc-cslqn\" (UID: \"23d7bb25-5d2c-4530-86dd-25a956953a4f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-cslqn" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.247449 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/63c09cd8-12ad-40a1-9877-3fc1500acb7a-auth-proxy-config\") pod \"machine-config-operator-74547568cd-kqtr5\" (UID: \"63c09cd8-12ad-40a1-9877-3fc1500acb7a\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kqtr5" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.247497 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/26739e4d-344c-4e25-bd0a-019f11456897-plugins-dir\") pod \"csi-hostpathplugin-jgvgl\" (UID: \"26739e4d-344c-4e25-bd0a-019f11456897\") " pod="hostpath-provisioner/csi-hostpathplugin-jgvgl" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.248341 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/63c09cd8-12ad-40a1-9877-3fc1500acb7a-proxy-tls\") pod \"machine-config-operator-74547568cd-kqtr5\" (UID: \"63c09cd8-12ad-40a1-9877-3fc1500acb7a\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kqtr5" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.248384 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c7fe95fc-b7d1-4ad1-8acf-e7ee4564babe-config-volume\") pod \"dns-default-dlgmw\" (UID: \"c7fe95fc-b7d1-4ad1-8acf-e7ee4564babe\") " pod="openshift-dns/dns-default-dlgmw" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.248542 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/26739e4d-344c-4e25-bd0a-019f11456897-socket-dir\") pod \"csi-hostpathplugin-jgvgl\" (UID: \"26739e4d-344c-4e25-bd0a-019f11456897\") " pod="hostpath-provisioner/csi-hostpathplugin-jgvgl" Jan 21 11:13:12 crc kubenswrapper[4824]: E0121 11:13:12.248671 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-21 11:13:12.748655901 +0000 UTC m=+135.041685193 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxlxx" (UID: "94e71514-0af6-4254-8815-c933ade6c9da") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.251527 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/63c09cd8-12ad-40a1-9877-3fc1500acb7a-images\") pod \"machine-config-operator-74547568cd-kqtr5\" (UID: \"63c09cd8-12ad-40a1-9877-3fc1500acb7a\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kqtr5" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.253587 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/faede743-203b-4bd1-b17a-776a4744bdf1-cert\") pod \"ingress-canary-b7tsx\" (UID: \"faede743-203b-4bd1-b17a-776a4744bdf1\") " pod="openshift-ingress-canary/ingress-canary-b7tsx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.255332 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/23d7bb25-5d2c-4530-86dd-25a956953a4f-apiservice-cert\") pod \"packageserver-d55dfcdfc-cslqn\" (UID: \"23d7bb25-5d2c-4530-86dd-25a956953a4f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-cslqn" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.255332 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/920ef22d-4a57-455c-a30f-c6d5f3a419e2-certs\") pod \"machine-config-server-5s88g\" (UID: \"920ef22d-4a57-455c-a30f-c6d5f3a419e2\") " pod="openshift-machine-config-operator/machine-config-server-5s88g" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.255760 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/d93236c5-17ab-4cd0-88a3-f4b3479067e5-proxy-tls\") pod \"machine-config-controller-84d6567774-9tldv\" (UID: \"d93236c5-17ab-4cd0-88a3-f4b3479067e5\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-9tldv" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.255866 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/920ef22d-4a57-455c-a30f-c6d5f3a419e2-node-bootstrap-token\") pod \"machine-config-server-5s88g\" (UID: \"920ef22d-4a57-455c-a30f-c6d5f3a419e2\") " pod="openshift-machine-config-operator/machine-config-server-5s88g" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.256083 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nqqww\" (UniqueName: \"kubernetes.io/projected/09bd0b9c-46fb-433f-bd30-014c69c90d9f-kube-api-access-nqqww\") pod \"collect-profiles-29483220-6r4fz\" (UID: \"09bd0b9c-46fb-433f-bd30-014c69c90d9f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29483220-6r4fz" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.256176 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/c7fe95fc-b7d1-4ad1-8acf-e7ee4564babe-metrics-tls\") pod \"dns-default-dlgmw\" (UID: \"c7fe95fc-b7d1-4ad1-8acf-e7ee4564babe\") " pod="openshift-dns/dns-default-dlgmw" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.258467 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/23d7bb25-5d2c-4530-86dd-25a956953a4f-webhook-cert\") pod \"packageserver-d55dfcdfc-cslqn\" (UID: \"23d7bb25-5d2c-4530-86dd-25a956953a4f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-cslqn" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.265531 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pczg5\" (UniqueName: \"kubernetes.io/projected/86dc1658-d302-449f-9ae3-d7bc5e795c57-kube-api-access-pczg5\") pod \"package-server-manager-789f6589d5-bkkzh\" (UID: \"86dc1658-d302-449f-9ae3-d7bc5e795c57\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-bkkzh" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.280218 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-ct4lp" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.285702 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tkwb6\" (UniqueName: \"kubernetes.io/projected/c117483a-fe5c-423b-995e-9c2e7f59566a-kube-api-access-tkwb6\") pod \"control-plane-machine-set-operator-78cbb6b69f-r65dm\" (UID: \"c117483a-fe5c-423b-995e-9c2e7f59566a\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-r65dm" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.295377 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-7grz2"] Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.308407 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pwmbq\" (UniqueName: \"kubernetes.io/projected/7a5d2d02-347e-4799-a28f-f303d51758d0-kube-api-access-pwmbq\") pod \"machine-approver-56656f9798-hzhhs\" (UID: \"7a5d2d02-347e-4799-a28f-f303d51758d0\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-hzhhs" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.317900 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-fbq6l"] Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.318976 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-qx94f"] Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.330761 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d132606d-b1ae-4cb7-865e-e1065c2eb4a9-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-t2thx\" (UID: \"d132606d-b1ae-4cb7-865e-e1065c2eb4a9\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-t2thx" Jan 21 11:13:12 crc kubenswrapper[4824]: W0121 11:13:12.338142 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8d0d706a_5fe5_402d_9040_3c763e525554.slice/crio-c63ddc91317112389bd667f708219c27c993346cd6fab5c172fde1b836cc4835 WatchSource:0}: Error finding container c63ddc91317112389bd667f708219c27c993346cd6fab5c172fde1b836cc4835: Status 404 returned error can't find the container with id c63ddc91317112389bd667f708219c27c993346cd6fab5c172fde1b836cc4835 Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.347648 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 21 11:13:12 crc kubenswrapper[4824]: E0121 11:13:12.348683 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-21 11:13:12.848648883 +0000 UTC m=+135.141678175 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.350629 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dzrmn\" (UniqueName: \"kubernetes.io/projected/7352d717-b2a3-4cd2-9e35-3ae2fe2d7b34-kube-api-access-dzrmn\") pod \"multus-admission-controller-857f4d67dd-27d2q\" (UID: \"7352d717-b2a3-4cd2-9e35-3ae2fe2d7b34\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-27d2q" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.365544 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29483220-6r4fz" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.372723 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zcznm\" (UniqueName: \"kubernetes.io/projected/94e71514-0af6-4254-8815-c933ade6c9da-kube-api-access-zcznm\") pod \"image-registry-697d97f7c8-wxlxx\" (UID: \"94e71514-0af6-4254-8815-c933ade6c9da\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxlxx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.373561 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-mh6ds"] Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.374929 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-nwlwd"] Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.387918 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/94e71514-0af6-4254-8815-c933ade6c9da-bound-sa-token\") pod \"image-registry-697d97f7c8-wxlxx\" (UID: \"94e71514-0af6-4254-8815-c933ade6c9da\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxlxx" Jan 21 11:13:12 crc kubenswrapper[4824]: W0121 11:13:12.388978 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc585ef2f_3e29_4b81_918f_11b075006ca0.slice/crio-0b16adc0798e89f22e1034c180951bc349e7c3b376c2bea592b22d2474b2bc5e WatchSource:0}: Error finding container 0b16adc0798e89f22e1034c180951bc349e7c3b376c2bea592b22d2474b2bc5e: Status 404 returned error can't find the container with id 0b16adc0798e89f22e1034c180951bc349e7c3b376c2bea592b22d2474b2bc5e Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.390560 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-t2thx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.411068 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6qg69\" (UniqueName: \"kubernetes.io/projected/ae7ce20e-9b65-4dd4-9e00-d96cf91803cb-kube-api-access-6qg69\") pod \"dns-operator-744455d44c-f6c7b\" (UID: \"ae7ce20e-9b65-4dd4-9e00-d96cf91803cb\") " pod="openshift-dns-operator/dns-operator-744455d44c-f6c7b" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.415571 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-bkkzh" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.429187 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-27d2q" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.431386 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6q9tc\" (UniqueName: \"kubernetes.io/projected/4b22a7aa-bcc0-4876-a374-a9943a6040bd-kube-api-access-6q9tc\") pod \"router-default-5444994796-bblkp\" (UID: \"4b22a7aa-bcc0-4876-a374-a9943a6040bd\") " pod="openshift-ingress/router-default-5444994796-bblkp" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.434094 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-r65dm" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.441397 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-j6bst" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.448598 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kjjkx\" (UniqueName: \"kubernetes.io/projected/14d99a1f-fca3-421a-8b47-ab4cda2f77fc-kube-api-access-kjjkx\") pod \"marketplace-operator-79b997595-gjhp5\" (UID: \"14d99a1f-fca3-421a-8b47-ab4cda2f77fc\") " pod="openshift-marketplace/marketplace-operator-79b997595-gjhp5" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.449736 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxlxx\" (UID: \"94e71514-0af6-4254-8815-c933ade6c9da\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxlxx" Jan 21 11:13:12 crc kubenswrapper[4824]: E0121 11:13:12.453178 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-21 11:13:12.953164449 +0000 UTC m=+135.246193741 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxlxx" (UID: "94e71514-0af6-4254-8815-c933ade6c9da") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.459805 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-bblkp" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.467802 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c873221f-c31d-411a-bbf5-c59e45b473a5-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-5lmnb\" (UID: \"c873221f-c31d-411a-bbf5-c59e45b473a5\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5lmnb" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.470240 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-ct4lp"] Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.479922 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-nwlwd" event={"ID":"c585ef2f-3e29-4b81-918f-11b075006ca0","Type":"ContainerStarted","Data":"0b16adc0798e89f22e1034c180951bc349e7c3b376c2bea592b22d2474b2bc5e"} Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.487719 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-s2n22" event={"ID":"c4b9a738-6a66-467d-b8a5-bddb20cb7e98","Type":"ContainerStarted","Data":"a4408a690301e0b3fde8a526b30b6f1091bb7fa680fdd7aff2e9a023eba39e40"} Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.487750 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-s2n22" event={"ID":"c4b9a738-6a66-467d-b8a5-bddb20cb7e98","Type":"ContainerStarted","Data":"94a67f23ca38cf50838a0b0ca9ef557570584d59359aa614c529d238934cbf0e"} Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.490324 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/b79b805a-7ab1-467f-90a5-ac7944cc564e-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-2tjlq\" (UID: \"b79b805a-7ab1-467f-90a5-ac7944cc564e\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2tjlq" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.499426 4824 generic.go:334] "Generic (PLEG): container finished" podID="0475fd50-10a8-4195-8643-9408478e953c" containerID="8e7f98988a1c14c829902e8582cad543e1a6a6f7682b17a425252e44a603a94a" exitCode=0 Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.499493 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7s7wx" event={"ID":"0475fd50-10a8-4195-8643-9408478e953c","Type":"ContainerDied","Data":"8e7f98988a1c14c829902e8582cad543e1a6a6f7682b17a425252e44a603a94a"} Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.499519 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7s7wx" event={"ID":"0475fd50-10a8-4195-8643-9408478e953c","Type":"ContainerStarted","Data":"75c7fa16af4bca084c4cc4bde49251029592fd9bed3decc1ddef1b9a08515ea4"} Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.504351 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-qx94f" event={"ID":"49ec79bd-9eb8-42c8-a262-e96c1457f686","Type":"ContainerStarted","Data":"24756c15affe3985216342ed2ac894d5b6bf6b2cb876e0f10ccb2500b9bad960"} Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.504388 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-qx94f" event={"ID":"49ec79bd-9eb8-42c8-a262-e96c1457f686","Type":"ContainerStarted","Data":"b847215d00329a56ffa445c815cb09f06304f7917a8c399ea195f25ad640e745"} Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.506759 4824 generic.go:334] "Generic (PLEG): container finished" podID="8364adda-37b5-4a9d-8abd-3daf12470f20" containerID="01624aa0fea408599e8ca78243d73a26a5d305f3b193a4bc7fb78578ffb7fe1c" exitCode=0 Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.510839 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-htg7q" event={"ID":"8364adda-37b5-4a9d-8abd-3daf12470f20","Type":"ContainerDied","Data":"01624aa0fea408599e8ca78243d73a26a5d305f3b193a4bc7fb78578ffb7fe1c"} Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.510884 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-htg7q" event={"ID":"8364adda-37b5-4a9d-8abd-3daf12470f20","Type":"ContainerStarted","Data":"57ced44b4bd056825dfdc76b73c5ce663d44e6ce7c97b61af5207421cc94a24c"} Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.515015 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jsn9x\" (UniqueName: \"kubernetes.io/projected/2279132e-c274-424b-8c8b-80ae5bd7dfe5-kube-api-access-jsn9x\") pod \"downloads-7954f5f757-8g4qk\" (UID: \"2279132e-c274-424b-8c8b-80ae5bd7dfe5\") " pod="openshift-console/downloads-7954f5f757-8g4qk" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.524505 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-4dkd8" event={"ID":"d962d0be-438a-4d08-8113-525238fa2156","Type":"ContainerStarted","Data":"13eb81088f05a509df912f982b6e26892b6c0fee8b275279ea88e029d704d428"} Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.524711 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-4dkd8" event={"ID":"d962d0be-438a-4d08-8113-525238fa2156","Type":"ContainerStarted","Data":"90b18225419fa36b91ae49d12b220ac91e071e5928e03c8bbf47f6f6ddb46af0"} Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.525018 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-4dkd8" Jan 21 11:13:12 crc kubenswrapper[4824]: W0121 11:13:12.525166 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4b22a7aa_bcc0_4876_a374_a9943a6040bd.slice/crio-3d55b184fa201e03b48f73696a0f95c5e0b4f0887605e69e2620a010c15909bf WatchSource:0}: Error finding container 3d55b184fa201e03b48f73696a0f95c5e0b4f0887605e69e2620a010c15909bf: Status 404 returned error can't find the container with id 3d55b184fa201e03b48f73696a0f95c5e0b4f0887605e69e2620a010c15909bf Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.527089 4824 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-4dkd8 container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.6:8443/healthz\": dial tcp 10.217.0.6:8443: connect: connection refused" start-of-body= Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.527146 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-4dkd8" podUID="d962d0be-438a-4d08-8113-525238fa2156" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.6:8443/healthz\": dial tcp 10.217.0.6:8443: connect: connection refused" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.527286 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-fbq6l" event={"ID":"8d0d706a-5fe5-402d-9040-3c763e525554","Type":"ContainerStarted","Data":"57c4a967c5b9aabd63ab037e2cb9dc29b3931b78fab3f197b33da1b36c5c1370"} Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.527311 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-fbq6l" event={"ID":"8d0d706a-5fe5-402d-9040-3c763e525554","Type":"ContainerStarted","Data":"c63ddc91317112389bd667f708219c27c993346cd6fab5c172fde1b836cc4835"} Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.534874 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-7grz2" event={"ID":"fb8bdb2b-bdc7-4433-be10-296fc67dc1c3","Type":"ContainerStarted","Data":"57b5bf1ffd0cee5939e39d5fe2a424668f24f6cb5ff7c71259477e24f21e95e6"} Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.534900 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-7grz2" event={"ID":"fb8bdb2b-bdc7-4433-be10-296fc67dc1c3","Type":"ContainerStarted","Data":"a7f32826f8477e3b840ca1f13a4b314f384df8cd972a12fc948ec057aef32d6f"} Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.538007 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-mh6ds" event={"ID":"29d5a4fe-fd42-42ec-85b4-c79af6424296","Type":"ContainerStarted","Data":"f0b3399b1909d22a75a5c79456700bfad80a9ed4c2b2ec54d4b775df25ba84ed"} Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.548399 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j8xhg\" (UniqueName: \"kubernetes.io/projected/37d2c136-5241-41da-a630-3d70273f1670-kube-api-access-j8xhg\") pod \"openshift-controller-manager-operator-756b6f6bc6-cnr8w\" (UID: \"37d2c136-5241-41da-a630-3d70273f1670\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-cnr8w" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.550591 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 21 11:13:12 crc kubenswrapper[4824]: E0121 11:13:12.554023 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-21 11:13:13.054004091 +0000 UTC m=+135.347033383 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.564208 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h6sxp\" (UniqueName: \"kubernetes.io/projected/16be0b45-eab8-4d0a-823b-5d877031dba5-kube-api-access-h6sxp\") pod \"olm-operator-6b444d44fb-cq5sn\" (UID: \"16be0b45-eab8-4d0a-823b-5d877031dba5\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-cq5sn" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.570621 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/07ad02bb-f06a-4619-87d1-112c78642603-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-7fl6h\" (UID: \"07ad02bb-f06a-4619-87d1-112c78642603\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-7fl6h" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.575727 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-hzhhs" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.586318 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29483220-6r4fz"] Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.589880 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/5dab0fba-f35e-42bd-9a7e-913b70c2dd2b-bound-sa-token\") pod \"ingress-operator-5b745b69d9-n8lql\" (UID: \"5dab0fba-f35e-42bd-9a7e-913b70c2dd2b\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-n8lql" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.590664 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-f6c7b" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.602715 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2tjlq" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.606713 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qrlgj\" (UniqueName: \"kubernetes.io/projected/13490666-4efb-492b-be88-1c6fde3f378e-kube-api-access-qrlgj\") pod \"service-ca-operator-777779d784-c5xks\" (UID: \"13490666-4efb-492b-be88-1c6fde3f378e\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-c5xks" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.615160 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-7fl6h" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.615472 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-gjhp5" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.636187 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-md94v\" (UniqueName: \"kubernetes.io/projected/02c3b11b-dbc4-445d-94fa-95c5ccc03cb8-kube-api-access-md94v\") pod \"catalog-operator-68c6474976-j8dcx\" (UID: \"02c3b11b-dbc4-445d-94fa-95c5ccc03cb8\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-j8dcx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.640105 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-bkkzh"] Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.651305 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-cq5sn" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.652050 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bhtz6\" (UniqueName: \"kubernetes.io/projected/86b5dca2-989f-4b6b-983b-87763ce0d469-kube-api-access-bhtz6\") pod \"oauth-openshift-558db77b4-fz7gx\" (UID: \"86b5dca2-989f-4b6b-983b-87763ce0d469\") " pod="openshift-authentication/oauth-openshift-558db77b4-fz7gx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.658870 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-t2thx"] Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.661887 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxlxx\" (UID: \"94e71514-0af6-4254-8815-c933ade6c9da\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxlxx" Jan 21 11:13:12 crc kubenswrapper[4824]: E0121 11:13:12.662256 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-21 11:13:13.162243119 +0000 UTC m=+135.455272411 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxlxx" (UID: "94e71514-0af6-4254-8815-c933ade6c9da") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 11:13:12 crc kubenswrapper[4824]: W0121 11:13:12.664485 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod86dc1658_d302_449f_9ae3_d7bc5e795c57.slice/crio-41bb1d013c63d1613e3f9736a883f4a3862526187c159bfbeb51c381ef8623d0 WatchSource:0}: Error finding container 41bb1d013c63d1613e3f9736a883f4a3862526187c159bfbeb51c381ef8623d0: Status 404 returned error can't find the container with id 41bb1d013c63d1613e3f9736a883f4a3862526187c159bfbeb51c381ef8623d0 Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.666677 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fh969\" (UniqueName: \"kubernetes.io/projected/74172695-d17f-4a47-b013-9e025e534916-kube-api-access-fh969\") pod \"kube-storage-version-migrator-operator-b67b599dd-s7wn9\" (UID: \"74172695-d17f-4a47-b013-9e025e534916\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-s7wn9" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.671553 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-8g4qk" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.686581 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-j8dcx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.690387 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pmsvd\" (UniqueName: \"kubernetes.io/projected/3358782c-3435-4c68-9f9c-c1040a6ada32-kube-api-access-pmsvd\") pod \"service-ca-9c57cc56f-rdcbg\" (UID: \"3358782c-3435-4c68-9f9c-c1040a6ada32\") " pod="openshift-service-ca/service-ca-9c57cc56f-rdcbg" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.703933 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-cnr8w" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.707908 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q27ql\" (UniqueName: \"kubernetes.io/projected/5dab0fba-f35e-42bd-9a7e-913b70c2dd2b-kube-api-access-q27ql\") pod \"ingress-operator-5b745b69d9-n8lql\" (UID: \"5dab0fba-f35e-42bd-9a7e-913b70c2dd2b\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-n8lql" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.709329 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-n8lql" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.712277 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-27d2q"] Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.722513 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5lmnb" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.729315 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5ksj6\" (UniqueName: \"kubernetes.io/projected/51e27630-d5c1-4eed-8232-3698ab99fc19-kube-api-access-5ksj6\") pod \"route-controller-manager-6576b87f9c-qs8d6\" (UID: \"51e27630-d5c1-4eed-8232-3698ab99fc19\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qs8d6" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.747063 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-c5xks" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.749347 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xgr59\" (UniqueName: \"kubernetes.io/projected/024ed20d-cf91-4660-998b-3d06d3464607-kube-api-access-xgr59\") pod \"etcd-operator-b45778765-cpgg4\" (UID: \"024ed20d-cf91-4660-998b-3d06d3464607\") " pod="openshift-etcd-operator/etcd-operator-b45778765-cpgg4" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.755777 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-j6bst"] Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.755827 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-r65dm"] Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.759206 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-s7wn9" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.762917 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 21 11:13:12 crc kubenswrapper[4824]: E0121 11:13:12.763513 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-21 11:13:13.263496239 +0000 UTC m=+135.556525531 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 11:13:12 crc kubenswrapper[4824]: W0121 11:13:12.767981 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7352d717_b2a3_4cd2_9e35_3ae2fe2d7b34.slice/crio-a58b1e840c1fbaaf4323bfd0d7baf951ec2faa2f74b8bef14349f6474ca6b7d8 WatchSource:0}: Error finding container a58b1e840c1fbaaf4323bfd0d7baf951ec2faa2f74b8bef14349f6474ca6b7d8: Status 404 returned error can't find the container with id a58b1e840c1fbaaf4323bfd0d7baf951ec2faa2f74b8bef14349f6474ca6b7d8 Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.814635 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kxstx\" (UniqueName: \"kubernetes.io/projected/d93236c5-17ab-4cd0-88a3-f4b3479067e5-kube-api-access-kxstx\") pod \"machine-config-controller-84d6567774-9tldv\" (UID: \"d93236c5-17ab-4cd0-88a3-f4b3479067e5\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-9tldv" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.840506 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h5dpw\" (UniqueName: \"kubernetes.io/projected/faede743-203b-4bd1-b17a-776a4744bdf1-kube-api-access-h5dpw\") pod \"ingress-canary-b7tsx\" (UID: \"faede743-203b-4bd1-b17a-776a4744bdf1\") " pod="openshift-ingress-canary/ingress-canary-b7tsx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.840810 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6nqck\" (UniqueName: \"kubernetes.io/projected/c7fe95fc-b7d1-4ad1-8acf-e7ee4564babe-kube-api-access-6nqck\") pod \"dns-default-dlgmw\" (UID: \"c7fe95fc-b7d1-4ad1-8acf-e7ee4564babe\") " pod="openshift-dns/dns-default-dlgmw" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.859230 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nz5wq\" (UniqueName: \"kubernetes.io/projected/26739e4d-344c-4e25-bd0a-019f11456897-kube-api-access-nz5wq\") pod \"csi-hostpathplugin-jgvgl\" (UID: \"26739e4d-344c-4e25-bd0a-019f11456897\") " pod="hostpath-provisioner/csi-hostpathplugin-jgvgl" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.877697 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxlxx\" (UID: \"94e71514-0af6-4254-8815-c933ade6c9da\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxlxx" Jan 21 11:13:12 crc kubenswrapper[4824]: W0121 11:13:12.877823 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod27b93806_10a5_46ba_8a81_20b05fd990c1.slice/crio-b90145e68496b029877d20e05a0410ab89139f34372c1cd90d9a9eea4f9469d9 WatchSource:0}: Error finding container b90145e68496b029877d20e05a0410ab89139f34372c1cd90d9a9eea4f9469d9: Status 404 returned error can't find the container with id b90145e68496b029877d20e05a0410ab89139f34372c1cd90d9a9eea4f9469d9 Jan 21 11:13:12 crc kubenswrapper[4824]: E0121 11:13:12.878186 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-21 11:13:13.378171875 +0000 UTC m=+135.671201156 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxlxx" (UID: "94e71514-0af6-4254-8815-c933ade6c9da") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.885560 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-fz7gx" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.895633 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lk25w\" (UniqueName: \"kubernetes.io/projected/920ef22d-4a57-455c-a30f-c6d5f3a419e2-kube-api-access-lk25w\") pod \"machine-config-server-5s88g\" (UID: \"920ef22d-4a57-455c-a30f-c6d5f3a419e2\") " pod="openshift-machine-config-operator/machine-config-server-5s88g" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.917622 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s4hmd\" (UniqueName: \"kubernetes.io/projected/63c09cd8-12ad-40a1-9877-3fc1500acb7a-kube-api-access-s4hmd\") pod \"machine-config-operator-74547568cd-kqtr5\" (UID: \"63c09cd8-12ad-40a1-9877-3fc1500acb7a\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kqtr5" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.920315 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j48kx\" (UniqueName: \"kubernetes.io/projected/23d7bb25-5d2c-4530-86dd-25a956953a4f-kube-api-access-j48kx\") pod \"packageserver-d55dfcdfc-cslqn\" (UID: \"23d7bb25-5d2c-4530-86dd-25a956953a4f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-cslqn" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.945003 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-cpgg4" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.978344 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 21 11:13:12 crc kubenswrapper[4824]: E0121 11:13:12.978727 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-21 11:13:13.478713617 +0000 UTC m=+135.771742909 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.978823 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-rdcbg" Jan 21 11:13:12 crc kubenswrapper[4824]: I0121 11:13:12.997057 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qs8d6" Jan 21 11:13:13 crc kubenswrapper[4824]: I0121 11:13:13.045943 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-7fl6h"] Jan 21 11:13:13 crc kubenswrapper[4824]: I0121 11:13:13.066862 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-9tldv" Jan 21 11:13:13 crc kubenswrapper[4824]: I0121 11:13:13.072642 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kqtr5" Jan 21 11:13:13 crc kubenswrapper[4824]: I0121 11:13:13.083442 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-cslqn" Jan 21 11:13:13 crc kubenswrapper[4824]: I0121 11:13:13.083974 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxlxx\" (UID: \"94e71514-0af6-4254-8815-c933ade6c9da\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxlxx" Jan 21 11:13:13 crc kubenswrapper[4824]: E0121 11:13:13.084293 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-21 11:13:13.584278313 +0000 UTC m=+135.877307605 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxlxx" (UID: "94e71514-0af6-4254-8815-c933ade6c9da") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 11:13:13 crc kubenswrapper[4824]: I0121 11:13:13.095858 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-5s88g" Jan 21 11:13:13 crc kubenswrapper[4824]: I0121 11:13:13.096218 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-dlgmw" Jan 21 11:13:13 crc kubenswrapper[4824]: I0121 11:13:13.110037 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-jgvgl" Jan 21 11:13:13 crc kubenswrapper[4824]: I0121 11:13:13.116142 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-b7tsx" Jan 21 11:13:13 crc kubenswrapper[4824]: I0121 11:13:13.184877 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 21 11:13:13 crc kubenswrapper[4824]: I0121 11:13:13.184978 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-cq5sn"] Jan 21 11:13:13 crc kubenswrapper[4824]: E0121 11:13:13.185232 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-21 11:13:13.685218384 +0000 UTC m=+135.978247676 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 11:13:13 crc kubenswrapper[4824]: W0121 11:13:13.220028 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod16be0b45_eab8_4d0a_823b_5d877031dba5.slice/crio-5fa3fc155f4cf8db417ce9eee59a7f91effb645e0b1ba49d021da1f9ce053d7f WatchSource:0}: Error finding container 5fa3fc155f4cf8db417ce9eee59a7f91effb645e0b1ba49d021da1f9ce053d7f: Status 404 returned error can't find the container with id 5fa3fc155f4cf8db417ce9eee59a7f91effb645e0b1ba49d021da1f9ce053d7f Jan 21 11:13:13 crc kubenswrapper[4824]: I0121 11:13:13.288807 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxlxx\" (UID: \"94e71514-0af6-4254-8815-c933ade6c9da\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxlxx" Jan 21 11:13:13 crc kubenswrapper[4824]: E0121 11:13:13.289206 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-21 11:13:13.789182985 +0000 UTC m=+136.082212277 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxlxx" (UID: "94e71514-0af6-4254-8815-c933ade6c9da") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 11:13:13 crc kubenswrapper[4824]: I0121 11:13:13.344714 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-j8dcx"] Jan 21 11:13:13 crc kubenswrapper[4824]: I0121 11:13:13.391828 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 21 11:13:13 crc kubenswrapper[4824]: E0121 11:13:13.393170 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-21 11:13:13.893137317 +0000 UTC m=+136.186166609 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 11:13:13 crc kubenswrapper[4824]: I0121 11:13:13.403170 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxlxx\" (UID: \"94e71514-0af6-4254-8815-c933ade6c9da\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxlxx" Jan 21 11:13:13 crc kubenswrapper[4824]: E0121 11:13:13.403555 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-21 11:13:13.903542186 +0000 UTC m=+136.196571479 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxlxx" (UID: "94e71514-0af6-4254-8815-c933ade6c9da") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 11:13:13 crc kubenswrapper[4824]: I0121 11:13:13.444970 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-f6c7b"] Jan 21 11:13:13 crc kubenswrapper[4824]: I0121 11:13:13.455945 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2tjlq"] Jan 21 11:13:13 crc kubenswrapper[4824]: I0121 11:13:13.463111 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-gjhp5"] Jan 21 11:13:13 crc kubenswrapper[4824]: I0121 11:13:13.463146 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-cnr8w"] Jan 21 11:13:13 crc kubenswrapper[4824]: I0121 11:13:13.465721 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-c5xks"] Jan 21 11:13:13 crc kubenswrapper[4824]: I0121 11:13:13.506243 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 21 11:13:13 crc kubenswrapper[4824]: E0121 11:13:13.506490 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-21 11:13:14.006470431 +0000 UTC m=+136.299499723 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 11:13:13 crc kubenswrapper[4824]: I0121 11:13:13.506633 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxlxx\" (UID: \"94e71514-0af6-4254-8815-c933ade6c9da\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxlxx" Jan 21 11:13:13 crc kubenswrapper[4824]: E0121 11:13:13.506887 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-21 11:13:14.006875271 +0000 UTC m=+136.299904564 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxlxx" (UID: "94e71514-0af6-4254-8815-c933ade6c9da") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 11:13:13 crc kubenswrapper[4824]: I0121 11:13:13.533383 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-8g4qk"] Jan 21 11:13:13 crc kubenswrapper[4824]: I0121 11:13:13.564547 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-ct4lp" event={"ID":"6a0a5723-a022-41f3-9222-c62cac06397e","Type":"ContainerStarted","Data":"a1cf9f71f9e51a9e84b32e1eb48baecb69e1af53943f073eac248f4bbd6e3490"} Jan 21 11:13:13 crc kubenswrapper[4824]: I0121 11:13:13.564624 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-ct4lp" event={"ID":"6a0a5723-a022-41f3-9222-c62cac06397e","Type":"ContainerStarted","Data":"24154a5c62917e47b8c541d57c970b6a4e0616e7f66b0df654db39d08754f71e"} Jan 21 11:13:13 crc kubenswrapper[4824]: I0121 11:13:13.581279 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-j8dcx" event={"ID":"02c3b11b-dbc4-445d-94fa-95c5ccc03cb8","Type":"ContainerStarted","Data":"c0d9b7b7dc305ec1f4f68cddbad0ca766683d3701f7d30b811f77e3b1b850c5c"} Jan 21 11:13:13 crc kubenswrapper[4824]: I0121 11:13:13.604244 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-bkkzh" event={"ID":"86dc1658-d302-449f-9ae3-d7bc5e795c57","Type":"ContainerStarted","Data":"863b0c3992a3892d462af79ff6fb9a86a7cdd408feecc5165baae4f4d4acded6"} Jan 21 11:13:13 crc kubenswrapper[4824]: I0121 11:13:13.604296 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-bkkzh" event={"ID":"86dc1658-d302-449f-9ae3-d7bc5e795c57","Type":"ContainerStarted","Data":"41bb1d013c63d1613e3f9736a883f4a3862526187c159bfbeb51c381ef8623d0"} Jan 21 11:13:13 crc kubenswrapper[4824]: I0121 11:13:13.607378 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 21 11:13:13 crc kubenswrapper[4824]: E0121 11:13:13.607951 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-21 11:13:14.107934527 +0000 UTC m=+136.400963818 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 11:13:13 crc kubenswrapper[4824]: I0121 11:13:13.609863 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-j6bst" event={"ID":"27b93806-10a5-46ba-8a81-20b05fd990c1","Type":"ContainerStarted","Data":"b90145e68496b029877d20e05a0410ab89139f34372c1cd90d9a9eea4f9469d9"} Jan 21 11:13:13 crc kubenswrapper[4824]: I0121 11:13:13.617585 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7s7wx" event={"ID":"0475fd50-10a8-4195-8643-9408478e953c","Type":"ContainerStarted","Data":"eba6f1475ef962456ec8d4aca6659cffa1eee13b1d1aef2f9c235c7f6532e8c1"} Jan 21 11:13:13 crc kubenswrapper[4824]: I0121 11:13:13.629650 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-cq5sn" event={"ID":"16be0b45-eab8-4d0a-823b-5d877031dba5","Type":"ContainerStarted","Data":"5fa3fc155f4cf8db417ce9eee59a7f91effb645e0b1ba49d021da1f9ce053d7f"} Jan 21 11:13:13 crc kubenswrapper[4824]: I0121 11:13:13.689634 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29483220-6r4fz" event={"ID":"09bd0b9c-46fb-433f-bd30-014c69c90d9f","Type":"ContainerStarted","Data":"277011d8fed4b1333d53f5a4a084181f52dada1e6d85b31750d9bab9937b3678"} Jan 21 11:13:13 crc kubenswrapper[4824]: I0121 11:13:13.689676 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29483220-6r4fz" event={"ID":"09bd0b9c-46fb-433f-bd30-014c69c90d9f","Type":"ContainerStarted","Data":"68b9bb179c8bbe2beb17e913adc7df65d063a4b5a6cdcd18f32da44f14fdb4ae"} Jan 21 11:13:13 crc kubenswrapper[4824]: I0121 11:13:13.719646 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxlxx\" (UID: \"94e71514-0af6-4254-8815-c933ade6c9da\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxlxx" Jan 21 11:13:13 crc kubenswrapper[4824]: E0121 11:13:13.720573 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-21 11:13:14.220562928 +0000 UTC m=+136.513592220 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxlxx" (UID: "94e71514-0af6-4254-8815-c933ade6c9da") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 11:13:13 crc kubenswrapper[4824]: I0121 11:13:13.723622 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-7grz2" event={"ID":"fb8bdb2b-bdc7-4433-be10-296fc67dc1c3","Type":"ContainerStarted","Data":"13825c9a9e1496a36936513a845471c8cd2d12f3e5dd6f1ea597f62b7a5881e7"} Jan 21 11:13:13 crc kubenswrapper[4824]: I0121 11:13:13.778260 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-nwlwd" event={"ID":"c585ef2f-3e29-4b81-918f-11b075006ca0","Type":"ContainerStarted","Data":"3a9966ed610a4f72371cd05219170eb5e70c5b477a36cba97f1fa20640f0ccd7"} Jan 21 11:13:13 crc kubenswrapper[4824]: I0121 11:13:13.785138 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-bblkp" event={"ID":"4b22a7aa-bcc0-4876-a374-a9943a6040bd","Type":"ContainerStarted","Data":"60ddd7b0675ec58022e55c78cbe86c12e0d7acd9251135287d12b3098481d22d"} Jan 21 11:13:13 crc kubenswrapper[4824]: I0121 11:13:13.785175 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-bblkp" event={"ID":"4b22a7aa-bcc0-4876-a374-a9943a6040bd","Type":"ContainerStarted","Data":"3d55b184fa201e03b48f73696a0f95c5e0b4f0887605e69e2620a010c15909bf"} Jan 21 11:13:13 crc kubenswrapper[4824]: I0121 11:13:13.791584 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-t2thx" event={"ID":"d132606d-b1ae-4cb7-865e-e1065c2eb4a9","Type":"ContainerStarted","Data":"50860c3a46aa9f38581345e1681b00dcccbfdd69caedca48718ba3de06ce2862"} Jan 21 11:13:13 crc kubenswrapper[4824]: I0121 11:13:13.791630 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-t2thx" event={"ID":"d132606d-b1ae-4cb7-865e-e1065c2eb4a9","Type":"ContainerStarted","Data":"d15d5576fb1d7a7584b073ba478a594a4c953db85611b5679f8498b6ed394016"} Jan 21 11:13:13 crc kubenswrapper[4824]: I0121 11:13:13.803573 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-27d2q" event={"ID":"7352d717-b2a3-4cd2-9e35-3ae2fe2d7b34","Type":"ContainerStarted","Data":"a58b1e840c1fbaaf4323bfd0d7baf951ec2faa2f74b8bef14349f6474ca6b7d8"} Jan 21 11:13:13 crc kubenswrapper[4824]: I0121 11:13:13.805090 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-5s88g" event={"ID":"920ef22d-4a57-455c-a30f-c6d5f3a419e2","Type":"ContainerStarted","Data":"4c429d9e761441d396247bc1947df1a4fc0f76fe16e4927368fc3f1b799adee0"} Jan 21 11:13:13 crc kubenswrapper[4824]: I0121 11:13:13.816145 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-htg7q" event={"ID":"8364adda-37b5-4a9d-8abd-3daf12470f20","Type":"ContainerStarted","Data":"85188d995a9d3025fb63f5bb716e23bca5109219b40e8522015af40804bea5d3"} Jan 21 11:13:13 crc kubenswrapper[4824]: I0121 11:13:13.820636 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 21 11:13:13 crc kubenswrapper[4824]: E0121 11:13:13.821914 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-21 11:13:14.321900004 +0000 UTC m=+136.614929296 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 11:13:13 crc kubenswrapper[4824]: I0121 11:13:13.827399 4824 generic.go:334] "Generic (PLEG): container finished" podID="8d0d706a-5fe5-402d-9040-3c763e525554" containerID="57c4a967c5b9aabd63ab037e2cb9dc29b3931b78fab3f197b33da1b36c5c1370" exitCode=0 Jan 21 11:13:13 crc kubenswrapper[4824]: I0121 11:13:13.827667 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-fbq6l" event={"ID":"8d0d706a-5fe5-402d-9040-3c763e525554","Type":"ContainerDied","Data":"57c4a967c5b9aabd63ab037e2cb9dc29b3931b78fab3f197b33da1b36c5c1370"} Jan 21 11:13:13 crc kubenswrapper[4824]: I0121 11:13:13.832050 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-mh6ds" event={"ID":"29d5a4fe-fd42-42ec-85b4-c79af6424296","Type":"ContainerStarted","Data":"42f8e34d6093d1ac6aa986143cc096ebe737677ee356032a8cfa97ce107372d0"} Jan 21 11:13:13 crc kubenswrapper[4824]: I0121 11:13:13.833585 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-mh6ds" Jan 21 11:13:13 crc kubenswrapper[4824]: I0121 11:13:13.838470 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-r65dm" event={"ID":"c117483a-fe5c-423b-995e-9c2e7f59566a","Type":"ContainerStarted","Data":"05346edfb2dc22987a7065666d0c7ed1f8a5cfaee344800819a0a26c4db5b9ae"} Jan 21 11:13:13 crc kubenswrapper[4824]: I0121 11:13:13.843278 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-7fl6h" event={"ID":"07ad02bb-f06a-4619-87d1-112c78642603","Type":"ContainerStarted","Data":"7f2a98a80d971de8e219b03ef193886a6ae7e63f2383d5c3777335050c9f8efd"} Jan 21 11:13:13 crc kubenswrapper[4824]: I0121 11:13:13.882350 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-hzhhs" event={"ID":"7a5d2d02-347e-4799-a28f-f303d51758d0","Type":"ContainerStarted","Data":"33f8d434af3d7825f96a18549c7957930822a3b9862435f1df2b941c88dc50b6"} Jan 21 11:13:13 crc kubenswrapper[4824]: I0121 11:13:13.882386 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-hzhhs" event={"ID":"7a5d2d02-347e-4799-a28f-f303d51758d0","Type":"ContainerStarted","Data":"7208fa8eaa4127528fdc27fa8019ce379ca5e5017e76eebb7bbfc1571ac29f3d"} Jan 21 11:13:13 crc kubenswrapper[4824]: I0121 11:13:13.898574 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-4dkd8" Jan 21 11:13:13 crc kubenswrapper[4824]: I0121 11:13:13.924946 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxlxx\" (UID: \"94e71514-0af6-4254-8815-c933ade6c9da\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxlxx" Jan 21 11:13:13 crc kubenswrapper[4824]: E0121 11:13:13.929253 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-21 11:13:14.429238932 +0000 UTC m=+136.722268224 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxlxx" (UID: "94e71514-0af6-4254-8815-c933ade6c9da") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 11:13:14 crc kubenswrapper[4824]: I0121 11:13:14.030385 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 21 11:13:14 crc kubenswrapper[4824]: E0121 11:13:14.030760 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-21 11:13:14.530745507 +0000 UTC m=+136.823774798 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 11:13:14 crc kubenswrapper[4824]: I0121 11:13:14.032062 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxlxx\" (UID: \"94e71514-0af6-4254-8815-c933ade6c9da\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxlxx" Jan 21 11:13:14 crc kubenswrapper[4824]: E0121 11:13:14.052262 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-21 11:13:14.552244211 +0000 UTC m=+136.845273503 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxlxx" (UID: "94e71514-0af6-4254-8815-c933ade6c9da") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 11:13:14 crc kubenswrapper[4824]: I0121 11:13:14.094141 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-4dkd8" podStartSLOduration=120.094127845 podStartE2EDuration="2m0.094127845s" podCreationTimestamp="2026-01-21 11:11:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:13:14.094088532 +0000 UTC m=+136.387117824" watchObservedRunningTime="2026-01-21 11:13:14.094127845 +0000 UTC m=+136.387157138" Jan 21 11:13:14 crc kubenswrapper[4824]: I0121 11:13:14.134301 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 21 11:13:14 crc kubenswrapper[4824]: E0121 11:13:14.134730 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-21 11:13:14.634713565 +0000 UTC m=+136.927742857 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 11:13:14 crc kubenswrapper[4824]: I0121 11:13:14.219707 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-qx94f" podStartSLOduration=120.21969199 podStartE2EDuration="2m0.21969199s" podCreationTimestamp="2026-01-21 11:11:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:13:14.219005701 +0000 UTC m=+136.512034993" watchObservedRunningTime="2026-01-21 11:13:14.21969199 +0000 UTC m=+136.512721282" Jan 21 11:13:14 crc kubenswrapper[4824]: I0121 11:13:14.241718 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxlxx\" (UID: \"94e71514-0af6-4254-8815-c933ade6c9da\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxlxx" Jan 21 11:13:14 crc kubenswrapper[4824]: E0121 11:13:14.242245 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-21 11:13:14.74223236 +0000 UTC m=+137.035261652 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxlxx" (UID: "94e71514-0af6-4254-8815-c933ade6c9da") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 11:13:14 crc kubenswrapper[4824]: I0121 11:13:14.297063 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-n8lql"] Jan 21 11:13:14 crc kubenswrapper[4824]: I0121 11:13:14.341370 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-qs8d6"] Jan 21 11:13:14 crc kubenswrapper[4824]: I0121 11:13:14.342292 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 21 11:13:14 crc kubenswrapper[4824]: E0121 11:13:14.342512 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-21 11:13:14.84249544 +0000 UTC m=+137.135524731 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 11:13:14 crc kubenswrapper[4824]: I0121 11:13:14.342543 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxlxx\" (UID: \"94e71514-0af6-4254-8815-c933ade6c9da\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxlxx" Jan 21 11:13:14 crc kubenswrapper[4824]: E0121 11:13:14.342906 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-21 11:13:14.842898847 +0000 UTC m=+137.135928139 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxlxx" (UID: "94e71514-0af6-4254-8815-c933ade6c9da") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 11:13:14 crc kubenswrapper[4824]: I0121 11:13:14.362276 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-mh6ds" Jan 21 11:13:14 crc kubenswrapper[4824]: I0121 11:13:14.388205 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-fz7gx"] Jan 21 11:13:14 crc kubenswrapper[4824]: I0121 11:13:14.392173 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-9tldv"] Jan 21 11:13:14 crc kubenswrapper[4824]: I0121 11:13:14.417935 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5lmnb"] Jan 21 11:13:14 crc kubenswrapper[4824]: I0121 11:13:14.445507 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 21 11:13:14 crc kubenswrapper[4824]: E0121 11:13:14.446546 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-21 11:13:14.946527247 +0000 UTC m=+137.239556538 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 11:13:14 crc kubenswrapper[4824]: I0121 11:13:14.462042 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-bblkp" Jan 21 11:13:14 crc kubenswrapper[4824]: I0121 11:13:14.469150 4824 patch_prober.go:28] interesting pod/router-default-5444994796-bblkp container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 21 11:13:14 crc kubenswrapper[4824]: [-]has-synced failed: reason withheld Jan 21 11:13:14 crc kubenswrapper[4824]: [+]process-running ok Jan 21 11:13:14 crc kubenswrapper[4824]: healthz check failed Jan 21 11:13:14 crc kubenswrapper[4824]: I0121 11:13:14.469214 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-bblkp" podUID="4b22a7aa-bcc0-4876-a374-a9943a6040bd" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 21 11:13:14 crc kubenswrapper[4824]: W0121 11:13:14.478719 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod51e27630_d5c1_4eed_8232_3698ab99fc19.slice/crio-66caec7ee49f045aa9a142ffed3e9d86fa441cb0eec919f264b1bbebed641a83 WatchSource:0}: Error finding container 66caec7ee49f045aa9a142ffed3e9d86fa441cb0eec919f264b1bbebed641a83: Status 404 returned error can't find the container with id 66caec7ee49f045aa9a142ffed3e9d86fa441cb0eec919f264b1bbebed641a83 Jan 21 11:13:14 crc kubenswrapper[4824]: I0121 11:13:14.519643 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-s2n22" podStartSLOduration=120.519627486 podStartE2EDuration="2m0.519627486s" podCreationTimestamp="2026-01-21 11:11:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:13:14.519052917 +0000 UTC m=+136.812082209" watchObservedRunningTime="2026-01-21 11:13:14.519627486 +0000 UTC m=+136.812656777" Jan 21 11:13:14 crc kubenswrapper[4824]: I0121 11:13:14.547696 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxlxx\" (UID: \"94e71514-0af6-4254-8815-c933ade6c9da\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxlxx" Jan 21 11:13:14 crc kubenswrapper[4824]: E0121 11:13:14.548051 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-21 11:13:15.048039412 +0000 UTC m=+137.341068705 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxlxx" (UID: "94e71514-0af6-4254-8815-c933ade6c9da") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 11:13:14 crc kubenswrapper[4824]: I0121 11:13:14.618998 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-nwlwd" podStartSLOduration=120.618951882 podStartE2EDuration="2m0.618951882s" podCreationTimestamp="2026-01-21 11:11:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:13:14.618397692 +0000 UTC m=+136.911426983" watchObservedRunningTime="2026-01-21 11:13:14.618951882 +0000 UTC m=+136.911981174" Jan 21 11:13:14 crc kubenswrapper[4824]: I0121 11:13:14.629517 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-b7tsx"] Jan 21 11:13:14 crc kubenswrapper[4824]: I0121 11:13:14.642703 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-rdcbg"] Jan 21 11:13:14 crc kubenswrapper[4824]: I0121 11:13:14.648999 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 21 11:13:14 crc kubenswrapper[4824]: E0121 11:13:14.649213 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-21 11:13:15.149190049 +0000 UTC m=+137.442219341 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 11:13:14 crc kubenswrapper[4824]: I0121 11:13:14.649449 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxlxx\" (UID: \"94e71514-0af6-4254-8815-c933ade6c9da\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxlxx" Jan 21 11:13:14 crc kubenswrapper[4824]: E0121 11:13:14.649821 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-21 11:13:15.149806196 +0000 UTC m=+137.442835487 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxlxx" (UID: "94e71514-0af6-4254-8815-c933ade6c9da") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 11:13:14 crc kubenswrapper[4824]: I0121 11:13:14.655495 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-mh6ds" podStartSLOduration=120.655483347 podStartE2EDuration="2m0.655483347s" podCreationTimestamp="2026-01-21 11:11:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:13:14.652940402 +0000 UTC m=+136.945969693" watchObservedRunningTime="2026-01-21 11:13:14.655483347 +0000 UTC m=+136.948512639" Jan 21 11:13:14 crc kubenswrapper[4824]: I0121 11:13:14.703367 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-cslqn"] Jan 21 11:13:14 crc kubenswrapper[4824]: I0121 11:13:14.705037 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-cpgg4"] Jan 21 11:13:14 crc kubenswrapper[4824]: I0121 11:13:14.708736 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-dlgmw"] Jan 21 11:13:14 crc kubenswrapper[4824]: I0121 11:13:14.722024 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-s7wn9"] Jan 21 11:13:14 crc kubenswrapper[4824]: I0121 11:13:14.722798 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-bblkp" podStartSLOduration=119.72278645 podStartE2EDuration="1m59.72278645s" podCreationTimestamp="2026-01-21 11:11:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:13:14.70327106 +0000 UTC m=+136.996300352" watchObservedRunningTime="2026-01-21 11:13:14.72278645 +0000 UTC m=+137.015815741" Jan 21 11:13:14 crc kubenswrapper[4824]: I0121 11:13:14.750882 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 21 11:13:14 crc kubenswrapper[4824]: E0121 11:13:14.751050 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-21 11:13:15.251027164 +0000 UTC m=+137.544056456 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 11:13:14 crc kubenswrapper[4824]: I0121 11:13:14.751285 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxlxx\" (UID: \"94e71514-0af6-4254-8815-c933ade6c9da\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxlxx" Jan 21 11:13:14 crc kubenswrapper[4824]: E0121 11:13:14.751638 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-21 11:13:15.251623745 +0000 UTC m=+137.544653037 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxlxx" (UID: "94e71514-0af6-4254-8815-c933ade6c9da") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 11:13:14 crc kubenswrapper[4824]: I0121 11:13:14.805718 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-r65dm" podStartSLOduration=119.805700658 podStartE2EDuration="1m59.805700658s" podCreationTimestamp="2026-01-21 11:11:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:13:14.770350011 +0000 UTC m=+137.063379303" watchObservedRunningTime="2026-01-21 11:13:14.805700658 +0000 UTC m=+137.098729951" Jan 21 11:13:14 crc kubenswrapper[4824]: I0121 11:13:14.806130 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29483220-6r4fz" podStartSLOduration=119.806122791 podStartE2EDuration="1m59.806122791s" podCreationTimestamp="2026-01-21 11:11:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:13:14.804796621 +0000 UTC m=+137.097825913" watchObservedRunningTime="2026-01-21 11:13:14.806122791 +0000 UTC m=+137.099152083" Jan 21 11:13:14 crc kubenswrapper[4824]: W0121 11:13:14.813586 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod23d7bb25_5d2c_4530_86dd_25a956953a4f.slice/crio-49d3a818b492b8df859fbc6c1e8a1a805dbadec0d6a740e52d27b04d4158d925 WatchSource:0}: Error finding container 49d3a818b492b8df859fbc6c1e8a1a805dbadec0d6a740e52d27b04d4158d925: Status 404 returned error can't find the container with id 49d3a818b492b8df859fbc6c1e8a1a805dbadec0d6a740e52d27b04d4158d925 Jan 21 11:13:14 crc kubenswrapper[4824]: W0121 11:13:14.834984 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc7fe95fc_b7d1_4ad1_8acf_e7ee4564babe.slice/crio-a92e9a2d4247fb90b0a505fd7792042cdffd7c5469973ffe9021e3c2156f63eb WatchSource:0}: Error finding container a92e9a2d4247fb90b0a505fd7792042cdffd7c5469973ffe9021e3c2156f63eb: Status 404 returned error can't find the container with id a92e9a2d4247fb90b0a505fd7792042cdffd7c5469973ffe9021e3c2156f63eb Jan 21 11:13:14 crc kubenswrapper[4824]: I0121 11:13:14.853683 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 21 11:13:14 crc kubenswrapper[4824]: E0121 11:13:14.854116 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-21 11:13:15.354104269 +0000 UTC m=+137.647133561 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 11:13:14 crc kubenswrapper[4824]: W0121 11:13:14.872233 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod74172695_d17f_4a47_b013_9e025e534916.slice/crio-2bb9c7e7f018544ad416d59b3c538a5f29156538646e38a8327e89c2bd11cd88 WatchSource:0}: Error finding container 2bb9c7e7f018544ad416d59b3c538a5f29156538646e38a8327e89c2bd11cd88: Status 404 returned error can't find the container with id 2bb9c7e7f018544ad416d59b3c538a5f29156538646e38a8327e89c2bd11cd88 Jan 21 11:13:14 crc kubenswrapper[4824]: I0121 11:13:14.893553 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-j8dcx" event={"ID":"02c3b11b-dbc4-445d-94fa-95c5ccc03cb8","Type":"ContainerStarted","Data":"757f71987259387576ae95787f25d067b1c7a3e6c5dcd0206a1faa0653381151"} Jan 21 11:13:14 crc kubenswrapper[4824]: I0121 11:13:14.895707 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-j8dcx" Jan 21 11:13:14 crc kubenswrapper[4824]: I0121 11:13:14.915798 4824 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-j8dcx container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.20:8443/healthz\": dial tcp 10.217.0.20:8443: connect: connection refused" start-of-body= Jan 21 11:13:14 crc kubenswrapper[4824]: I0121 11:13:14.915836 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-j8dcx" podUID="02c3b11b-dbc4-445d-94fa-95c5ccc03cb8" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.20:8443/healthz\": dial tcp 10.217.0.20:8443: connect: connection refused" Jan 21 11:13:14 crc kubenswrapper[4824]: I0121 11:13:14.916448 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-s7wn9" event={"ID":"74172695-d17f-4a47-b013-9e025e534916","Type":"ContainerStarted","Data":"2bb9c7e7f018544ad416d59b3c538a5f29156538646e38a8327e89c2bd11cd88"} Jan 21 11:13:14 crc kubenswrapper[4824]: I0121 11:13:14.921625 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7s7wx" podStartSLOduration=119.921613037 podStartE2EDuration="1m59.921613037s" podCreationTimestamp="2026-01-21 11:11:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:13:14.854312058 +0000 UTC m=+137.147341351" watchObservedRunningTime="2026-01-21 11:13:14.921613037 +0000 UTC m=+137.214642328" Jan 21 11:13:14 crc kubenswrapper[4824]: I0121 11:13:14.925146 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-kqtr5"] Jan 21 11:13:14 crc kubenswrapper[4824]: I0121 11:13:14.934878 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-jgvgl"] Jan 21 11:13:14 crc kubenswrapper[4824]: I0121 11:13:14.947421 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-cq5sn" event={"ID":"16be0b45-eab8-4d0a-823b-5d877031dba5","Type":"ContainerStarted","Data":"dd377815656d0707d5d6e9d707e7afef4665367c1db4361ef545e1a8bf287d97"} Jan 21 11:13:14 crc kubenswrapper[4824]: I0121 11:13:14.948614 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-cq5sn" Jan 21 11:13:14 crc kubenswrapper[4824]: I0121 11:13:14.959092 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxlxx\" (UID: \"94e71514-0af6-4254-8815-c933ade6c9da\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxlxx" Jan 21 11:13:14 crc kubenswrapper[4824]: E0121 11:13:14.959365 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-21 11:13:15.459354183 +0000 UTC m=+137.752383475 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxlxx" (UID: "94e71514-0af6-4254-8815-c933ade6c9da") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 11:13:14 crc kubenswrapper[4824]: I0121 11:13:14.991206 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-t2thx" podStartSLOduration=119.991188828 podStartE2EDuration="1m59.991188828s" podCreationTimestamp="2026-01-21 11:11:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:13:14.991022154 +0000 UTC m=+137.284051446" watchObservedRunningTime="2026-01-21 11:13:14.991188828 +0000 UTC m=+137.284218120" Jan 21 11:13:15 crc kubenswrapper[4824]: I0121 11:13:14.995393 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-cq5sn" Jan 21 11:13:15 crc kubenswrapper[4824]: I0121 11:13:15.000728 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-rdcbg" event={"ID":"3358782c-3435-4c68-9f9c-c1040a6ada32","Type":"ContainerStarted","Data":"38a77399b318a83fe4ece3efc7fadb9b3a89f0203b4293698f30d603fa103891"} Jan 21 11:13:15 crc kubenswrapper[4824]: I0121 11:13:15.010594 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-cnr8w" event={"ID":"37d2c136-5241-41da-a630-3d70273f1670","Type":"ContainerStarted","Data":"7b51fb3348172ec85dd20f861b1f1bf51a4054b0bd408215322d12c361203133"} Jan 21 11:13:15 crc kubenswrapper[4824]: I0121 11:13:15.010627 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-cnr8w" event={"ID":"37d2c136-5241-41da-a630-3d70273f1670","Type":"ContainerStarted","Data":"923df253da37dd5b6251f3ca08df8620e565dc4afda0554844090400996c31f6"} Jan 21 11:13:15 crc kubenswrapper[4824]: I0121 11:13:15.066540 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 21 11:13:15 crc kubenswrapper[4824]: E0121 11:13:15.067622 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-21 11:13:15.567608651 +0000 UTC m=+137.860637942 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 11:13:15 crc kubenswrapper[4824]: I0121 11:13:15.093092 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-n8lql" event={"ID":"5dab0fba-f35e-42bd-9a7e-913b70c2dd2b","Type":"ContainerStarted","Data":"573d45c88ddfeee5cecbfabe535ab387ebbe06773fb6b02c5cc55e6a1d0fffa1"} Jan 21 11:13:15 crc kubenswrapper[4824]: I0121 11:13:15.101012 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-fz7gx" event={"ID":"86b5dca2-989f-4b6b-983b-87763ce0d469","Type":"ContainerStarted","Data":"046a7d0640fbf37bae3b158d9284d146391f1ce208f167ac80fea14196ea051f"} Jan 21 11:13:15 crc kubenswrapper[4824]: I0121 11:13:15.102887 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-7grz2" podStartSLOduration=120.102873276 podStartE2EDuration="2m0.102873276s" podCreationTimestamp="2026-01-21 11:11:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:13:15.090574559 +0000 UTC m=+137.383603851" watchObservedRunningTime="2026-01-21 11:13:15.102873276 +0000 UTC m=+137.395902567" Jan 21 11:13:15 crc kubenswrapper[4824]: I0121 11:13:15.132946 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-c5xks" event={"ID":"13490666-4efb-492b-be88-1c6fde3f378e","Type":"ContainerStarted","Data":"cf24e15effeb0a5a7bbfe0c5b7d94b4824adff54c21b171c42502c9ba19f9242"} Jan 21 11:13:15 crc kubenswrapper[4824]: I0121 11:13:15.133006 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-c5xks" event={"ID":"13490666-4efb-492b-be88-1c6fde3f378e","Type":"ContainerStarted","Data":"b1b1ae1f7226d453f6259d47fde178296c1e39b0d8660341bc42a84e5f5fc305"} Jan 21 11:13:15 crc kubenswrapper[4824]: I0121 11:13:15.150368 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-j8dcx" podStartSLOduration=120.150351498 podStartE2EDuration="2m0.150351498s" podCreationTimestamp="2026-01-21 11:11:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:13:15.132179361 +0000 UTC m=+137.425208653" watchObservedRunningTime="2026-01-21 11:13:15.150351498 +0000 UTC m=+137.443380790" Jan 21 11:13:15 crc kubenswrapper[4824]: I0121 11:13:15.150739 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-cnr8w" podStartSLOduration=121.150734577 podStartE2EDuration="2m1.150734577s" podCreationTimestamp="2026-01-21 11:11:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:13:15.149610967 +0000 UTC m=+137.442640259" watchObservedRunningTime="2026-01-21 11:13:15.150734577 +0000 UTC m=+137.443763870" Jan 21 11:13:15 crc kubenswrapper[4824]: I0121 11:13:15.161789 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-cpgg4" event={"ID":"024ed20d-cf91-4660-998b-3d06d3464607","Type":"ContainerStarted","Data":"907f0e9f1ba0abbd21d3b2ebb8e7d365907202e44d50409f58a962b67e9c0f82"} Jan 21 11:13:15 crc kubenswrapper[4824]: I0121 11:13:15.168323 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxlxx\" (UID: \"94e71514-0af6-4254-8815-c933ade6c9da\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxlxx" Jan 21 11:13:15 crc kubenswrapper[4824]: E0121 11:13:15.169320 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-21 11:13:15.669310381 +0000 UTC m=+137.962339674 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxlxx" (UID: "94e71514-0af6-4254-8815-c933ade6c9da") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 11:13:15 crc kubenswrapper[4824]: I0121 11:13:15.187873 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-cq5sn" podStartSLOduration=120.18786142 podStartE2EDuration="2m0.18786142s" podCreationTimestamp="2026-01-21 11:11:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:13:15.187273095 +0000 UTC m=+137.480302386" watchObservedRunningTime="2026-01-21 11:13:15.18786142 +0000 UTC m=+137.480890712" Jan 21 11:13:15 crc kubenswrapper[4824]: I0121 11:13:15.196330 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-5s88g" event={"ID":"920ef22d-4a57-455c-a30f-c6d5f3a419e2","Type":"ContainerStarted","Data":"3b506a482bf40c00e2d6a4eeadde103dc948a6bbad06449b4b5cc81ed27d731f"} Jan 21 11:13:15 crc kubenswrapper[4824]: I0121 11:13:15.244527 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-5s88g" podStartSLOduration=6.244510774 podStartE2EDuration="6.244510774s" podCreationTimestamp="2026-01-21 11:13:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:13:15.243365555 +0000 UTC m=+137.536394846" watchObservedRunningTime="2026-01-21 11:13:15.244510774 +0000 UTC m=+137.537540056" Jan 21 11:13:15 crc kubenswrapper[4824]: I0121 11:13:15.245450 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-c5xks" podStartSLOduration=120.245436534 podStartE2EDuration="2m0.245436534s" podCreationTimestamp="2026-01-21 11:11:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:13:15.222323818 +0000 UTC m=+137.515353111" watchObservedRunningTime="2026-01-21 11:13:15.245436534 +0000 UTC m=+137.538465826" Jan 21 11:13:15 crc kubenswrapper[4824]: I0121 11:13:15.272493 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 21 11:13:15 crc kubenswrapper[4824]: E0121 11:13:15.273492 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-21 11:13:15.773475198 +0000 UTC m=+138.066504490 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 11:13:15 crc kubenswrapper[4824]: I0121 11:13:15.286023 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qs8d6" event={"ID":"51e27630-d5c1-4eed-8232-3698ab99fc19","Type":"ContainerStarted","Data":"66caec7ee49f045aa9a142ffed3e9d86fa441cb0eec919f264b1bbebed641a83"} Jan 21 11:13:15 crc kubenswrapper[4824]: I0121 11:13:15.286194 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qs8d6" Jan 21 11:13:15 crc kubenswrapper[4824]: I0121 11:13:15.294275 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-j6bst" event={"ID":"27b93806-10a5-46ba-8a81-20b05fd990c1","Type":"ContainerStarted","Data":"66f680a3a461408efa0f9f7f82f5fd91d998ef92fe166391b9a76fcfcd379415"} Jan 21 11:13:15 crc kubenswrapper[4824]: I0121 11:13:15.294339 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-j6bst" event={"ID":"27b93806-10a5-46ba-8a81-20b05fd990c1","Type":"ContainerStarted","Data":"ac68dbd34db8c9a005fcd37771a0552c714428d27022925e4208aa9c867c7924"} Jan 21 11:13:15 crc kubenswrapper[4824]: I0121 11:13:15.303541 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qs8d6" podStartSLOduration=120.303529319 podStartE2EDuration="2m0.303529319s" podCreationTimestamp="2026-01-21 11:11:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:13:15.303186645 +0000 UTC m=+137.596215937" watchObservedRunningTime="2026-01-21 11:13:15.303529319 +0000 UTC m=+137.596558611" Jan 21 11:13:15 crc kubenswrapper[4824]: I0121 11:13:15.304324 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-dlgmw" event={"ID":"c7fe95fc-b7d1-4ad1-8acf-e7ee4564babe","Type":"ContainerStarted","Data":"a92e9a2d4247fb90b0a505fd7792042cdffd7c5469973ffe9021e3c2156f63eb"} Jan 21 11:13:15 crc kubenswrapper[4824]: I0121 11:13:15.309460 4824 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-qs8d6 container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.22:8443/healthz\": dial tcp 10.217.0.22:8443: connect: connection refused" start-of-body= Jan 21 11:13:15 crc kubenswrapper[4824]: I0121 11:13:15.309521 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qs8d6" podUID="51e27630-d5c1-4eed-8232-3698ab99fc19" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.22:8443/healthz\": dial tcp 10.217.0.22:8443: connect: connection refused" Jan 21 11:13:15 crc kubenswrapper[4824]: I0121 11:13:15.325766 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-8g4qk" event={"ID":"2279132e-c274-424b-8c8b-80ae5bd7dfe5","Type":"ContainerStarted","Data":"becc6485ca48ad42d37b3ff35bcbe4b1895daea3fb4e50341e5518aa753d7f77"} Jan 21 11:13:15 crc kubenswrapper[4824]: I0121 11:13:15.326669 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-8g4qk" Jan 21 11:13:15 crc kubenswrapper[4824]: I0121 11:13:15.328117 4824 patch_prober.go:28] interesting pod/downloads-7954f5f757-8g4qk container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.16:8080/\": dial tcp 10.217.0.16:8080: connect: connection refused" start-of-body= Jan 21 11:13:15 crc kubenswrapper[4824]: I0121 11:13:15.328157 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-8g4qk" podUID="2279132e-c274-424b-8c8b-80ae5bd7dfe5" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.16:8080/\": dial tcp 10.217.0.16:8080: connect: connection refused" Jan 21 11:13:15 crc kubenswrapper[4824]: I0121 11:13:15.338647 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-j6bst" podStartSLOduration=120.338634265 podStartE2EDuration="2m0.338634265s" podCreationTimestamp="2026-01-21 11:11:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:13:15.335307187 +0000 UTC m=+137.628336479" watchObservedRunningTime="2026-01-21 11:13:15.338634265 +0000 UTC m=+137.631663556" Jan 21 11:13:15 crc kubenswrapper[4824]: I0121 11:13:15.366014 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2tjlq" event={"ID":"b79b805a-7ab1-467f-90a5-ac7944cc564e","Type":"ContainerStarted","Data":"ff6b12f5197bc273b9b1f5eaa9fe8891d095a52df4dba219ea169c425cdbcc29"} Jan 21 11:13:15 crc kubenswrapper[4824]: I0121 11:13:15.374079 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxlxx\" (UID: \"94e71514-0af6-4254-8815-c933ade6c9da\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxlxx" Jan 21 11:13:15 crc kubenswrapper[4824]: I0121 11:13:15.374790 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-r65dm" event={"ID":"c117483a-fe5c-423b-995e-9c2e7f59566a","Type":"ContainerStarted","Data":"0afb8e9f166ddbd00968e6ab571144882d40df74bd23f93bf7a9e95410697ecc"} Jan 21 11:13:15 crc kubenswrapper[4824]: E0121 11:13:15.377300 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-21 11:13:15.877280781 +0000 UTC m=+138.170310073 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxlxx" (UID: "94e71514-0af6-4254-8815-c933ade6c9da") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 11:13:15 crc kubenswrapper[4824]: I0121 11:13:15.386389 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-8g4qk" podStartSLOduration=121.386370471 podStartE2EDuration="2m1.386370471s" podCreationTimestamp="2026-01-21 11:11:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:13:15.371765734 +0000 UTC m=+137.664795026" watchObservedRunningTime="2026-01-21 11:13:15.386370471 +0000 UTC m=+137.679399763" Jan 21 11:13:15 crc kubenswrapper[4824]: I0121 11:13:15.387033 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-7fl6h" event={"ID":"07ad02bb-f06a-4619-87d1-112c78642603","Type":"ContainerStarted","Data":"2d7963c8981edc983480fac716eac85cd99025b69e6ddc50c3242d01f5c53f13"} Jan 21 11:13:15 crc kubenswrapper[4824]: I0121 11:13:15.399943 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2tjlq" podStartSLOduration=120.399925456 podStartE2EDuration="2m0.399925456s" podCreationTimestamp="2026-01-21 11:11:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:13:15.399526557 +0000 UTC m=+137.692555849" watchObservedRunningTime="2026-01-21 11:13:15.399925456 +0000 UTC m=+137.692954749" Jan 21 11:13:15 crc kubenswrapper[4824]: I0121 11:13:15.402225 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-b7tsx" event={"ID":"faede743-203b-4bd1-b17a-776a4744bdf1","Type":"ContainerStarted","Data":"409842eba4ded972cdf6e0d8b0cc3dfc9e94886c2d221b1cd5479b50e36f942f"} Jan 21 11:13:15 crc kubenswrapper[4824]: I0121 11:13:15.446970 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-27d2q" event={"ID":"7352d717-b2a3-4cd2-9e35-3ae2fe2d7b34","Type":"ContainerStarted","Data":"5c55044ce64ff381aa81be5bb4c9b389c84a5b9b921435cd1eeedf44d2c77613"} Jan 21 11:13:15 crc kubenswrapper[4824]: I0121 11:13:15.447027 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-27d2q" event={"ID":"7352d717-b2a3-4cd2-9e35-3ae2fe2d7b34","Type":"ContainerStarted","Data":"061b951668ab926a35a6e5500308fd9f83880f1a4fec23f785997d53ab76d390"} Jan 21 11:13:15 crc kubenswrapper[4824]: I0121 11:13:15.466194 4824 patch_prober.go:28] interesting pod/router-default-5444994796-bblkp container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 21 11:13:15 crc kubenswrapper[4824]: [-]has-synced failed: reason withheld Jan 21 11:13:15 crc kubenswrapper[4824]: [+]process-running ok Jan 21 11:13:15 crc kubenswrapper[4824]: healthz check failed Jan 21 11:13:15 crc kubenswrapper[4824]: I0121 11:13:15.466265 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-bblkp" podUID="4b22a7aa-bcc0-4876-a374-a9943a6040bd" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 21 11:13:15 crc kubenswrapper[4824]: I0121 11:13:15.467283 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-7fl6h" podStartSLOduration=120.467264386 podStartE2EDuration="2m0.467264386s" podCreationTimestamp="2026-01-21 11:11:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:13:15.425931645 +0000 UTC m=+137.718960937" watchObservedRunningTime="2026-01-21 11:13:15.467264386 +0000 UTC m=+137.760293678" Jan 21 11:13:15 crc kubenswrapper[4824]: I0121 11:13:15.467685 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-27d2q" podStartSLOduration=120.467674637 podStartE2EDuration="2m0.467674637s" podCreationTimestamp="2026-01-21 11:11:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:13:15.466785627 +0000 UTC m=+137.759814919" watchObservedRunningTime="2026-01-21 11:13:15.467674637 +0000 UTC m=+137.760703928" Jan 21 11:13:15 crc kubenswrapper[4824]: I0121 11:13:15.478907 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 21 11:13:15 crc kubenswrapper[4824]: E0121 11:13:15.479295 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-21 11:13:15.979277747 +0000 UTC m=+138.272307039 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 11:13:15 crc kubenswrapper[4824]: I0121 11:13:15.479421 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-bkkzh" event={"ID":"86dc1658-d302-449f-9ae3-d7bc5e795c57","Type":"ContainerStarted","Data":"ec7097ec47b1ede21a87b07e962cafae7130cdbdb00e1277f2a250754e3a5cd3"} Jan 21 11:13:15 crc kubenswrapper[4824]: I0121 11:13:15.479848 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-bkkzh" Jan 21 11:13:15 crc kubenswrapper[4824]: I0121 11:13:15.480731 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxlxx\" (UID: \"94e71514-0af6-4254-8815-c933ade6c9da\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxlxx" Jan 21 11:13:15 crc kubenswrapper[4824]: E0121 11:13:15.482831 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-21 11:13:15.982811092 +0000 UTC m=+138.275840384 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxlxx" (UID: "94e71514-0af6-4254-8815-c933ade6c9da") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 11:13:15 crc kubenswrapper[4824]: I0121 11:13:15.488031 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-gjhp5" event={"ID":"14d99a1f-fca3-421a-8b47-ab4cda2f77fc","Type":"ContainerStarted","Data":"605262189eb74a464efb5cb3518becda7f59cb81cb5a3ce736fbc1a29ea45dc4"} Jan 21 11:13:15 crc kubenswrapper[4824]: I0121 11:13:15.488088 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-gjhp5" event={"ID":"14d99a1f-fca3-421a-8b47-ab4cda2f77fc","Type":"ContainerStarted","Data":"3217737690e95d2837429663404e46050468bcb19f6fbf887ccc71adb8493eef"} Jan 21 11:13:15 crc kubenswrapper[4824]: I0121 11:13:15.488891 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-gjhp5" Jan 21 11:13:15 crc kubenswrapper[4824]: I0121 11:13:15.505047 4824 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-gjhp5 container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.36:8080/healthz\": dial tcp 10.217.0.36:8080: connect: connection refused" start-of-body= Jan 21 11:13:15 crc kubenswrapper[4824]: I0121 11:13:15.505084 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-gjhp5" podUID="14d99a1f-fca3-421a-8b47-ab4cda2f77fc" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.36:8080/healthz\": dial tcp 10.217.0.36:8080: connect: connection refused" Jan 21 11:13:15 crc kubenswrapper[4824]: I0121 11:13:15.506762 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-bkkzh" podStartSLOduration=120.506742224 podStartE2EDuration="2m0.506742224s" podCreationTimestamp="2026-01-21 11:11:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:13:15.505342165 +0000 UTC m=+137.798371457" watchObservedRunningTime="2026-01-21 11:13:15.506742224 +0000 UTC m=+137.799771516" Jan 21 11:13:15 crc kubenswrapper[4824]: I0121 11:13:15.511669 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-cslqn" event={"ID":"23d7bb25-5d2c-4530-86dd-25a956953a4f","Type":"ContainerStarted","Data":"49d3a818b492b8df859fbc6c1e8a1a805dbadec0d6a740e52d27b04d4158d925"} Jan 21 11:13:15 crc kubenswrapper[4824]: I0121 11:13:15.542337 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-ct4lp" event={"ID":"6a0a5723-a022-41f3-9222-c62cac06397e","Type":"ContainerStarted","Data":"06728a7822f65072829b4ddbca0e331708111d1e264540df6b8ad4f8e4b65f68"} Jan 21 11:13:15 crc kubenswrapper[4824]: I0121 11:13:15.582297 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-9tldv" event={"ID":"d93236c5-17ab-4cd0-88a3-f4b3479067e5","Type":"ContainerStarted","Data":"69311d84e7163549b83ee0ed47ea75b896b482d3e2b029ab6cc5815195deb913"} Jan 21 11:13:15 crc kubenswrapper[4824]: I0121 11:13:15.583889 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 21 11:13:15 crc kubenswrapper[4824]: E0121 11:13:15.587118 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-21 11:13:16.087102397 +0000 UTC m=+138.380131688 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 11:13:15 crc kubenswrapper[4824]: I0121 11:13:15.589665 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-gjhp5" podStartSLOduration=120.589643949 podStartE2EDuration="2m0.589643949s" podCreationTimestamp="2026-01-21 11:11:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:13:15.535042209 +0000 UTC m=+137.828071502" watchObservedRunningTime="2026-01-21 11:13:15.589643949 +0000 UTC m=+137.882673242" Jan 21 11:13:15 crc kubenswrapper[4824]: I0121 11:13:15.589851 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-ct4lp" podStartSLOduration=121.589847152 podStartE2EDuration="2m1.589847152s" podCreationTimestamp="2026-01-21 11:11:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:13:15.589554391 +0000 UTC m=+137.882583693" watchObservedRunningTime="2026-01-21 11:13:15.589847152 +0000 UTC m=+137.882876443" Jan 21 11:13:15 crc kubenswrapper[4824]: I0121 11:13:15.619376 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-f6c7b" event={"ID":"ae7ce20e-9b65-4dd4-9e00-d96cf91803cb","Type":"ContainerStarted","Data":"04ca14e605aa7cca976a17de1d5b8d43161815ffcc069e03b99cab4fbbd6715e"} Jan 21 11:13:15 crc kubenswrapper[4824]: I0121 11:13:15.628260 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5lmnb" event={"ID":"c873221f-c31d-411a-bbf5-c59e45b473a5","Type":"ContainerStarted","Data":"09c5871e2b1da5897a16dc587de0a6ce894cfabce3a6d3002ba916a166a963d4"} Jan 21 11:13:15 crc kubenswrapper[4824]: I0121 11:13:15.650597 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5lmnb" podStartSLOduration=120.650579583 podStartE2EDuration="2m0.650579583s" podCreationTimestamp="2026-01-21 11:11:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:13:15.641065116 +0000 UTC m=+137.934094408" watchObservedRunningTime="2026-01-21 11:13:15.650579583 +0000 UTC m=+137.943608876" Jan 21 11:13:15 crc kubenswrapper[4824]: I0121 11:13:15.688276 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxlxx\" (UID: \"94e71514-0af6-4254-8815-c933ade6c9da\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxlxx" Jan 21 11:13:15 crc kubenswrapper[4824]: E0121 11:13:15.689314 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-21 11:13:16.189298797 +0000 UTC m=+138.482328088 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxlxx" (UID: "94e71514-0af6-4254-8815-c933ade6c9da") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 11:13:15 crc kubenswrapper[4824]: I0121 11:13:15.788995 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 21 11:13:15 crc kubenswrapper[4824]: E0121 11:13:15.789233 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-21 11:13:16.289207571 +0000 UTC m=+138.582236863 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 11:13:15 crc kubenswrapper[4824]: I0121 11:13:15.789659 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxlxx\" (UID: \"94e71514-0af6-4254-8815-c933ade6c9da\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxlxx" Jan 21 11:13:15 crc kubenswrapper[4824]: E0121 11:13:15.797715 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-21 11:13:16.297698877 +0000 UTC m=+138.590728170 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxlxx" (UID: "94e71514-0af6-4254-8815-c933ade6c9da") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 11:13:15 crc kubenswrapper[4824]: I0121 11:13:15.894455 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 21 11:13:15 crc kubenswrapper[4824]: E0121 11:13:15.894987 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-21 11:13:16.394936465 +0000 UTC m=+138.687965758 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 11:13:15 crc kubenswrapper[4824]: I0121 11:13:15.996905 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxlxx\" (UID: \"94e71514-0af6-4254-8815-c933ade6c9da\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxlxx" Jan 21 11:13:15 crc kubenswrapper[4824]: E0121 11:13:15.997358 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-21 11:13:16.497337841 +0000 UTC m=+138.790367132 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxlxx" (UID: "94e71514-0af6-4254-8815-c933ade6c9da") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 11:13:16 crc kubenswrapper[4824]: I0121 11:13:16.102395 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 21 11:13:16 crc kubenswrapper[4824]: E0121 11:13:16.102518 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-21 11:13:16.602499911 +0000 UTC m=+138.895529203 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 11:13:16 crc kubenswrapper[4824]: I0121 11:13:16.102952 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxlxx\" (UID: \"94e71514-0af6-4254-8815-c933ade6c9da\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxlxx" Jan 21 11:13:16 crc kubenswrapper[4824]: E0121 11:13:16.103219 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-21 11:13:16.603209604 +0000 UTC m=+138.896238896 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxlxx" (UID: "94e71514-0af6-4254-8815-c933ade6c9da") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 11:13:16 crc kubenswrapper[4824]: I0121 11:13:16.163572 4824 csr.go:261] certificate signing request csr-k74p7 is approved, waiting to be issued Jan 21 11:13:16 crc kubenswrapper[4824]: I0121 11:13:16.168747 4824 csr.go:257] certificate signing request csr-k74p7 is issued Jan 21 11:13:16 crc kubenswrapper[4824]: I0121 11:13:16.204716 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 21 11:13:16 crc kubenswrapper[4824]: E0121 11:13:16.204896 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-21 11:13:16.704868214 +0000 UTC m=+138.997897506 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 11:13:16 crc kubenswrapper[4824]: I0121 11:13:16.205005 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxlxx\" (UID: \"94e71514-0af6-4254-8815-c933ade6c9da\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxlxx" Jan 21 11:13:16 crc kubenswrapper[4824]: E0121 11:13:16.205343 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-21 11:13:16.705330822 +0000 UTC m=+138.998360115 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxlxx" (UID: "94e71514-0af6-4254-8815-c933ade6c9da") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 11:13:16 crc kubenswrapper[4824]: I0121 11:13:16.306506 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 21 11:13:16 crc kubenswrapper[4824]: E0121 11:13:16.306687 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-21 11:13:16.806664072 +0000 UTC m=+139.099693364 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 11:13:16 crc kubenswrapper[4824]: I0121 11:13:16.306950 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxlxx\" (UID: \"94e71514-0af6-4254-8815-c933ade6c9da\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxlxx" Jan 21 11:13:16 crc kubenswrapper[4824]: E0121 11:13:16.307240 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-21 11:13:16.807230727 +0000 UTC m=+139.100260018 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxlxx" (UID: "94e71514-0af6-4254-8815-c933ade6c9da") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 11:13:16 crc kubenswrapper[4824]: I0121 11:13:16.407856 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 21 11:13:16 crc kubenswrapper[4824]: E0121 11:13:16.408051 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-21 11:13:16.908018381 +0000 UTC m=+139.201047673 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 11:13:16 crc kubenswrapper[4824]: I0121 11:13:16.408233 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxlxx\" (UID: \"94e71514-0af6-4254-8815-c933ade6c9da\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxlxx" Jan 21 11:13:16 crc kubenswrapper[4824]: E0121 11:13:16.408665 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-21 11:13:16.908646581 +0000 UTC m=+139.201675872 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxlxx" (UID: "94e71514-0af6-4254-8815-c933ade6c9da") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 11:13:16 crc kubenswrapper[4824]: I0121 11:13:16.465279 4824 patch_prober.go:28] interesting pod/router-default-5444994796-bblkp container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 21 11:13:16 crc kubenswrapper[4824]: [-]has-synced failed: reason withheld Jan 21 11:13:16 crc kubenswrapper[4824]: [+]process-running ok Jan 21 11:13:16 crc kubenswrapper[4824]: healthz check failed Jan 21 11:13:16 crc kubenswrapper[4824]: I0121 11:13:16.465349 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-bblkp" podUID="4b22a7aa-bcc0-4876-a374-a9943a6040bd" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 21 11:13:16 crc kubenswrapper[4824]: I0121 11:13:16.509241 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 21 11:13:16 crc kubenswrapper[4824]: E0121 11:13:16.509413 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-21 11:13:17.009388549 +0000 UTC m=+139.302417842 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 11:13:16 crc kubenswrapper[4824]: I0121 11:13:16.509647 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxlxx\" (UID: \"94e71514-0af6-4254-8815-c933ade6c9da\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxlxx" Jan 21 11:13:16 crc kubenswrapper[4824]: E0121 11:13:16.509923 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-21 11:13:17.009911081 +0000 UTC m=+139.302940373 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxlxx" (UID: "94e71514-0af6-4254-8815-c933ade6c9da") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 11:13:16 crc kubenswrapper[4824]: I0121 11:13:16.611144 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 21 11:13:16 crc kubenswrapper[4824]: E0121 11:13:16.611340 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-21 11:13:17.111313721 +0000 UTC m=+139.404343013 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 11:13:16 crc kubenswrapper[4824]: I0121 11:13:16.611577 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxlxx\" (UID: \"94e71514-0af6-4254-8815-c933ade6c9da\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxlxx" Jan 21 11:13:16 crc kubenswrapper[4824]: E0121 11:13:16.611917 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-21 11:13:17.111903788 +0000 UTC m=+139.404933081 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxlxx" (UID: "94e71514-0af6-4254-8815-c933ade6c9da") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 11:13:16 crc kubenswrapper[4824]: I0121 11:13:16.631567 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-jgvgl" event={"ID":"26739e4d-344c-4e25-bd0a-019f11456897","Type":"ContainerStarted","Data":"2357fdef6edff16fe248b51b2573129538bac0ebcbe355fd4f0744b4760e074f"} Jan 21 11:13:16 crc kubenswrapper[4824]: I0121 11:13:16.631624 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-jgvgl" event={"ID":"26739e4d-344c-4e25-bd0a-019f11456897","Type":"ContainerStarted","Data":"191c1cbbbb507e726a9fc6a06f49f5bfd133346772551af833babba6bac384fc"} Jan 21 11:13:16 crc kubenswrapper[4824]: I0121 11:13:16.633052 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-cpgg4" event={"ID":"024ed20d-cf91-4660-998b-3d06d3464607","Type":"ContainerStarted","Data":"8d6da99e5afb06b10adc2e159c9dfb91014cadd7b29865889c55079801523753"} Jan 21 11:13:16 crc kubenswrapper[4824]: I0121 11:13:16.634187 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-cslqn" event={"ID":"23d7bb25-5d2c-4530-86dd-25a956953a4f","Type":"ContainerStarted","Data":"a18fe41bc9f1a62f11b63c70a3c09ab814ba529ea96b52baf9317658c9a34f69"} Jan 21 11:13:16 crc kubenswrapper[4824]: I0121 11:13:16.634574 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-cslqn" Jan 21 11:13:16 crc kubenswrapper[4824]: I0121 11:13:16.635523 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2tjlq" event={"ID":"b79b805a-7ab1-467f-90a5-ac7944cc564e","Type":"ContainerStarted","Data":"fbd69b8205e4cc5b6a7dc3738ab90446c2c9b31eda354ec6976608d7bb2336a3"} Jan 21 11:13:16 crc kubenswrapper[4824]: I0121 11:13:16.637923 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-f6c7b" event={"ID":"ae7ce20e-9b65-4dd4-9e00-d96cf91803cb","Type":"ContainerStarted","Data":"030ab318b65de947022b4f352c2e1b6768805b8d1dc79b53fdc3a85c031ea89d"} Jan 21 11:13:16 crc kubenswrapper[4824]: I0121 11:13:16.637969 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-f6c7b" event={"ID":"ae7ce20e-9b65-4dd4-9e00-d96cf91803cb","Type":"ContainerStarted","Data":"bd42c74cf5e6dfa746c5d497cdc62153be642693c9ec03faf754c61476ddd331"} Jan 21 11:13:16 crc kubenswrapper[4824]: I0121 11:13:16.639414 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-dlgmw" event={"ID":"c7fe95fc-b7d1-4ad1-8acf-e7ee4564babe","Type":"ContainerStarted","Data":"bdf1dfb3e18ae28b5dd5316bda516f1af9158072d9d605c478bfa2943dc87d2a"} Jan 21 11:13:16 crc kubenswrapper[4824]: I0121 11:13:16.639439 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-dlgmw" event={"ID":"c7fe95fc-b7d1-4ad1-8acf-e7ee4564babe","Type":"ContainerStarted","Data":"2c1919ef7a931735c653fb352fe2a897a5d074c0b408afd192aa4c432f141422"} Jan 21 11:13:16 crc kubenswrapper[4824]: I0121 11:13:16.639755 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-dlgmw" Jan 21 11:13:16 crc kubenswrapper[4824]: I0121 11:13:16.640728 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-s7wn9" event={"ID":"74172695-d17f-4a47-b013-9e025e534916","Type":"ContainerStarted","Data":"98f08cff82a80cfcccb794726fdd3ef6aa3878c820ee87ff8f0ce7a1689c9be4"} Jan 21 11:13:16 crc kubenswrapper[4824]: I0121 11:13:16.641909 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qs8d6" event={"ID":"51e27630-d5c1-4eed-8232-3698ab99fc19","Type":"ContainerStarted","Data":"e407f561a94b91929f1905dd35fb9fce0694555be001231861d30f794a254728"} Jan 21 11:13:16 crc kubenswrapper[4824]: I0121 11:13:16.643743 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-rdcbg" event={"ID":"3358782c-3435-4c68-9f9c-c1040a6ada32","Type":"ContainerStarted","Data":"33d6f04d21a2f84840ecc147d2a73698c0b7579e93456756c1f45bca7450a4d9"} Jan 21 11:13:16 crc kubenswrapper[4824]: I0121 11:13:16.644972 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-8g4qk" event={"ID":"2279132e-c274-424b-8c8b-80ae5bd7dfe5","Type":"ContainerStarted","Data":"e8a4c509ba9741577881459900939f844d9d1987bece7045fdf2ed555a8af9d8"} Jan 21 11:13:16 crc kubenswrapper[4824]: I0121 11:13:16.645560 4824 patch_prober.go:28] interesting pod/downloads-7954f5f757-8g4qk container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.16:8080/\": dial tcp 10.217.0.16:8080: connect: connection refused" start-of-body= Jan 21 11:13:16 crc kubenswrapper[4824]: I0121 11:13:16.645592 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-8g4qk" podUID="2279132e-c274-424b-8c8b-80ae5bd7dfe5" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.16:8080/\": dial tcp 10.217.0.16:8080: connect: connection refused" Jan 21 11:13:16 crc kubenswrapper[4824]: I0121 11:13:16.645748 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qs8d6" Jan 21 11:13:16 crc kubenswrapper[4824]: I0121 11:13:16.647415 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-htg7q" event={"ID":"8364adda-37b5-4a9d-8abd-3daf12470f20","Type":"ContainerStarted","Data":"e116aa985fb0389e3d80883c44d0f137bc3abbb6bd14ffacfad6761ccd102d60"} Jan 21 11:13:16 crc kubenswrapper[4824]: I0121 11:13:16.648543 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-fz7gx" event={"ID":"86b5dca2-989f-4b6b-983b-87763ce0d469","Type":"ContainerStarted","Data":"839b4db59e6f15755832bf687632140307e984c4933ab5994cda3f84fb56f80f"} Jan 21 11:13:16 crc kubenswrapper[4824]: I0121 11:13:16.648744 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-fz7gx" Jan 21 11:13:16 crc kubenswrapper[4824]: I0121 11:13:16.650383 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kqtr5" event={"ID":"63c09cd8-12ad-40a1-9877-3fc1500acb7a","Type":"ContainerStarted","Data":"2ae4c495da6ee06fbb44a8d0b9889c50c26e09b1bd4eed72119afd916722467d"} Jan 21 11:13:16 crc kubenswrapper[4824]: I0121 11:13:16.650423 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kqtr5" event={"ID":"63c09cd8-12ad-40a1-9877-3fc1500acb7a","Type":"ContainerStarted","Data":"fe6496e1452e7435e48ed695e321941578784919e86b59cbda2becfb36ff4c6b"} Jan 21 11:13:16 crc kubenswrapper[4824]: I0121 11:13:16.650437 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kqtr5" event={"ID":"63c09cd8-12ad-40a1-9877-3fc1500acb7a","Type":"ContainerStarted","Data":"f05b730bf98627ba801da84010b12c900b72c91b8de345b422941b9fbfff5a3f"} Jan 21 11:13:16 crc kubenswrapper[4824]: I0121 11:13:16.651607 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5lmnb" event={"ID":"c873221f-c31d-411a-bbf5-c59e45b473a5","Type":"ContainerStarted","Data":"9876f148fac8d256f8df6efbf5ac5f62b7c3b46dead9a743743d3d5ebb4abb69"} Jan 21 11:13:16 crc kubenswrapper[4824]: I0121 11:13:16.653599 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-9tldv" event={"ID":"d93236c5-17ab-4cd0-88a3-f4b3479067e5","Type":"ContainerStarted","Data":"766c63224fae3c18fae3adec0594b20bcd6d133f487a645ea28f65cfabad9132"} Jan 21 11:13:16 crc kubenswrapper[4824]: I0121 11:13:16.653672 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-9tldv" event={"ID":"d93236c5-17ab-4cd0-88a3-f4b3479067e5","Type":"ContainerStarted","Data":"951567f3adb934892728b522d94d22bc9bc79d7a05baeb1945237817c10327d5"} Jan 21 11:13:16 crc kubenswrapper[4824]: I0121 11:13:16.655463 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-n8lql" event={"ID":"5dab0fba-f35e-42bd-9a7e-913b70c2dd2b","Type":"ContainerStarted","Data":"03bc2ab81d22972fd9c8c1c1d652fdedcb038fe3b6031b95a9d65586e643fd12"} Jan 21 11:13:16 crc kubenswrapper[4824]: I0121 11:13:16.655492 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-n8lql" event={"ID":"5dab0fba-f35e-42bd-9a7e-913b70c2dd2b","Type":"ContainerStarted","Data":"9b0b5099e430097dfbaa3a46a392984fa7c9bbdf05ca3d28a430160a4137e4d8"} Jan 21 11:13:16 crc kubenswrapper[4824]: I0121 11:13:16.657032 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-b7tsx" event={"ID":"faede743-203b-4bd1-b17a-776a4744bdf1","Type":"ContainerStarted","Data":"934e6b741f99fc948601142b6c4dfe7be70cfbdae26baa682593e0cec9afd503"} Jan 21 11:13:16 crc kubenswrapper[4824]: I0121 11:13:16.662439 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-hzhhs" event={"ID":"7a5d2d02-347e-4799-a28f-f303d51758d0","Type":"ContainerStarted","Data":"eacddd49ef3bc5441741a259a2c6d10efa1a1a68c48c3d11e446b57eb54286ee"} Jan 21 11:13:16 crc kubenswrapper[4824]: I0121 11:13:16.687904 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-fbq6l" event={"ID":"8d0d706a-5fe5-402d-9040-3c763e525554","Type":"ContainerStarted","Data":"8d2b64b2f8dc9c4cadefa7a97077294d8931104ce4823d281735da5165e308e9"} Jan 21 11:13:16 crc kubenswrapper[4824]: I0121 11:13:16.693053 4824 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-gjhp5 container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.36:8080/healthz\": dial tcp 10.217.0.36:8080: connect: connection refused" start-of-body= Jan 21 11:13:16 crc kubenswrapper[4824]: I0121 11:13:16.693109 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-gjhp5" podUID="14d99a1f-fca3-421a-8b47-ab4cda2f77fc" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.36:8080/healthz\": dial tcp 10.217.0.36:8080: connect: connection refused" Jan 21 11:13:16 crc kubenswrapper[4824]: I0121 11:13:16.712512 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 21 11:13:16 crc kubenswrapper[4824]: E0121 11:13:16.712943 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-21 11:13:17.212926766 +0000 UTC m=+139.505956057 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 11:13:16 crc kubenswrapper[4824]: I0121 11:13:16.718598 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-j8dcx" Jan 21 11:13:16 crc kubenswrapper[4824]: I0121 11:13:16.732458 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-cpgg4" podStartSLOduration=121.732441063 podStartE2EDuration="2m1.732441063s" podCreationTimestamp="2026-01-21 11:11:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:13:16.676983576 +0000 UTC m=+138.970012868" watchObservedRunningTime="2026-01-21 11:13:16.732441063 +0000 UTC m=+139.025470355" Jan 21 11:13:16 crc kubenswrapper[4824]: I0121 11:13:16.733152 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-hzhhs" podStartSLOduration=122.733146156 podStartE2EDuration="2m2.733146156s" podCreationTimestamp="2026-01-21 11:11:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:13:16.731538638 +0000 UTC m=+139.024567919" watchObservedRunningTime="2026-01-21 11:13:16.733146156 +0000 UTC m=+139.026175448" Jan 21 11:13:16 crc kubenswrapper[4824]: I0121 11:13:16.770180 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-b7tsx" podStartSLOduration=7.770156099 podStartE2EDuration="7.770156099s" podCreationTimestamp="2026-01-21 11:13:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:13:16.765145711 +0000 UTC m=+139.058175002" watchObservedRunningTime="2026-01-21 11:13:16.770156099 +0000 UTC m=+139.063185392" Jan 21 11:13:16 crc kubenswrapper[4824]: I0121 11:13:16.781184 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-cslqn" podStartSLOduration=121.781153743 podStartE2EDuration="2m1.781153743s" podCreationTimestamp="2026-01-21 11:11:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:13:16.780949589 +0000 UTC m=+139.073978881" watchObservedRunningTime="2026-01-21 11:13:16.781153743 +0000 UTC m=+139.074183035" Jan 21 11:13:16 crc kubenswrapper[4824]: I0121 11:13:16.813596 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-f6c7b" podStartSLOduration=122.813579416 podStartE2EDuration="2m2.813579416s" podCreationTimestamp="2026-01-21 11:11:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:13:16.799375462 +0000 UTC m=+139.092404755" watchObservedRunningTime="2026-01-21 11:13:16.813579416 +0000 UTC m=+139.106608708" Jan 21 11:13:16 crc kubenswrapper[4824]: I0121 11:13:16.814750 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-fz7gx" podStartSLOduration=122.81474238 podStartE2EDuration="2m2.81474238s" podCreationTimestamp="2026-01-21 11:11:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:13:16.813641203 +0000 UTC m=+139.106670495" watchObservedRunningTime="2026-01-21 11:13:16.81474238 +0000 UTC m=+139.107771673" Jan 21 11:13:16 crc kubenswrapper[4824]: I0121 11:13:16.819043 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxlxx\" (UID: \"94e71514-0af6-4254-8815-c933ade6c9da\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxlxx" Jan 21 11:13:16 crc kubenswrapper[4824]: E0121 11:13:16.823057 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-21 11:13:17.3230437 +0000 UTC m=+139.616072993 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxlxx" (UID: "94e71514-0af6-4254-8815-c933ade6c9da") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 11:13:16 crc kubenswrapper[4824]: I0121 11:13:16.831392 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-rdcbg" podStartSLOduration=121.831375126 podStartE2EDuration="2m1.831375126s" podCreationTimestamp="2026-01-21 11:11:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:13:16.828254035 +0000 UTC m=+139.121283327" watchObservedRunningTime="2026-01-21 11:13:16.831375126 +0000 UTC m=+139.124404418" Jan 21 11:13:16 crc kubenswrapper[4824]: I0121 11:13:16.854041 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-htg7q" Jan 21 11:13:16 crc kubenswrapper[4824]: I0121 11:13:16.854421 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-htg7q" Jan 21 11:13:16 crc kubenswrapper[4824]: I0121 11:13:16.871021 4824 patch_prober.go:28] interesting pod/apiserver-76f77b778f-htg7q container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Jan 21 11:13:16 crc kubenswrapper[4824]: [+]log ok Jan 21 11:13:16 crc kubenswrapper[4824]: [+]etcd ok Jan 21 11:13:16 crc kubenswrapper[4824]: [+]poststarthook/start-apiserver-admission-initializer ok Jan 21 11:13:16 crc kubenswrapper[4824]: [+]poststarthook/generic-apiserver-start-informers ok Jan 21 11:13:16 crc kubenswrapper[4824]: [+]poststarthook/max-in-flight-filter ok Jan 21 11:13:16 crc kubenswrapper[4824]: [+]poststarthook/storage-object-count-tracker-hook ok Jan 21 11:13:16 crc kubenswrapper[4824]: [+]poststarthook/image.openshift.io-apiserver-caches ok Jan 21 11:13:16 crc kubenswrapper[4824]: [-]poststarthook/authorization.openshift.io-bootstrapclusterroles failed: reason withheld Jan 21 11:13:16 crc kubenswrapper[4824]: [+]poststarthook/authorization.openshift.io-ensurenodebootstrap-sa ok Jan 21 11:13:16 crc kubenswrapper[4824]: [+]poststarthook/project.openshift.io-projectcache ok Jan 21 11:13:16 crc kubenswrapper[4824]: [+]poststarthook/project.openshift.io-projectauthorizationcache ok Jan 21 11:13:16 crc kubenswrapper[4824]: [+]poststarthook/openshift.io-startinformers ok Jan 21 11:13:16 crc kubenswrapper[4824]: [+]poststarthook/openshift.io-restmapperupdater ok Jan 21 11:13:16 crc kubenswrapper[4824]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Jan 21 11:13:16 crc kubenswrapper[4824]: livez check failed Jan 21 11:13:16 crc kubenswrapper[4824]: I0121 11:13:16.871058 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-htg7q" podUID="8364adda-37b5-4a9d-8abd-3daf12470f20" containerName="openshift-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 21 11:13:16 crc kubenswrapper[4824]: I0121 11:13:16.891232 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-s7wn9" podStartSLOduration=121.891212789 podStartE2EDuration="2m1.891212789s" podCreationTimestamp="2026-01-21 11:11:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:13:16.860996194 +0000 UTC m=+139.154025485" watchObservedRunningTime="2026-01-21 11:13:16.891212789 +0000 UTC m=+139.184242071" Jan 21 11:13:16 crc kubenswrapper[4824]: I0121 11:13:16.892794 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kqtr5" podStartSLOduration=121.892783869 podStartE2EDuration="2m1.892783869s" podCreationTimestamp="2026-01-21 11:11:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:13:16.890288643 +0000 UTC m=+139.183317935" watchObservedRunningTime="2026-01-21 11:13:16.892783869 +0000 UTC m=+139.185813162" Jan 21 11:13:16 crc kubenswrapper[4824]: I0121 11:13:16.922561 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-fz7gx" Jan 21 11:13:16 crc kubenswrapper[4824]: I0121 11:13:16.923147 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 21 11:13:16 crc kubenswrapper[4824]: E0121 11:13:16.923512 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-21 11:13:17.423497538 +0000 UTC m=+139.716526831 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 11:13:16 crc kubenswrapper[4824]: I0121 11:13:16.952483 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-dlgmw" podStartSLOduration=7.952466801 podStartE2EDuration="7.952466801s" podCreationTimestamp="2026-01-21 11:13:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:13:16.952439069 +0000 UTC m=+139.245468361" watchObservedRunningTime="2026-01-21 11:13:16.952466801 +0000 UTC m=+139.245496093" Jan 21 11:13:16 crc kubenswrapper[4824]: I0121 11:13:16.952573 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-htg7q" podStartSLOduration=122.952568923 podStartE2EDuration="2m2.952568923s" podCreationTimestamp="2026-01-21 11:11:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:13:16.927718785 +0000 UTC m=+139.220748067" watchObservedRunningTime="2026-01-21 11:13:16.952568923 +0000 UTC m=+139.245598215" Jan 21 11:13:16 crc kubenswrapper[4824]: I0121 11:13:16.969683 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7s7wx" Jan 21 11:13:16 crc kubenswrapper[4824]: I0121 11:13:16.969736 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7s7wx" Jan 21 11:13:16 crc kubenswrapper[4824]: I0121 11:13:16.986496 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-7q9f2"] Jan 21 11:13:16 crc kubenswrapper[4824]: I0121 11:13:16.987440 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7s7wx" Jan 21 11:13:16 crc kubenswrapper[4824]: I0121 11:13:16.987601 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7q9f2" Jan 21 11:13:16 crc kubenswrapper[4824]: I0121 11:13:16.991279 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Jan 21 11:13:16 crc kubenswrapper[4824]: I0121 11:13:16.995299 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-n8lql" podStartSLOduration=121.995287777 podStartE2EDuration="2m1.995287777s" podCreationTimestamp="2026-01-21 11:11:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:13:16.994359353 +0000 UTC m=+139.287388645" watchObservedRunningTime="2026-01-21 11:13:16.995287777 +0000 UTC m=+139.288317069" Jan 21 11:13:17 crc kubenswrapper[4824]: I0121 11:13:17.014518 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-7q9f2"] Jan 21 11:13:17 crc kubenswrapper[4824]: I0121 11:13:17.024305 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxlxx\" (UID: \"94e71514-0af6-4254-8815-c933ade6c9da\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxlxx" Jan 21 11:13:17 crc kubenswrapper[4824]: E0121 11:13:17.024733 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-21 11:13:17.524722444 +0000 UTC m=+139.817751736 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxlxx" (UID: "94e71514-0af6-4254-8815-c933ade6c9da") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 11:13:17 crc kubenswrapper[4824]: I0121 11:13:17.050458 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-9tldv" podStartSLOduration=122.050442686 podStartE2EDuration="2m2.050442686s" podCreationTimestamp="2026-01-21 11:11:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:13:17.04928966 +0000 UTC m=+139.342318942" watchObservedRunningTime="2026-01-21 11:13:17.050442686 +0000 UTC m=+139.343471977" Jan 21 11:13:17 crc kubenswrapper[4824]: I0121 11:13:17.095847 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-fbq6l" podStartSLOduration=123.095831273 podStartE2EDuration="2m3.095831273s" podCreationTimestamp="2026-01-21 11:11:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:13:17.095016413 +0000 UTC m=+139.388045705" watchObservedRunningTime="2026-01-21 11:13:17.095831273 +0000 UTC m=+139.388860565" Jan 21 11:13:17 crc kubenswrapper[4824]: I0121 11:13:17.126312 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 21 11:13:17 crc kubenswrapper[4824]: I0121 11:13:17.126571 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1b040d3d-f4a0-451a-97de-08bea41c6dce-utilities\") pod \"community-operators-7q9f2\" (UID: \"1b040d3d-f4a0-451a-97de-08bea41c6dce\") " pod="openshift-marketplace/community-operators-7q9f2" Jan 21 11:13:17 crc kubenswrapper[4824]: I0121 11:13:17.126659 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mxt8j\" (UniqueName: \"kubernetes.io/projected/1b040d3d-f4a0-451a-97de-08bea41c6dce-kube-api-access-mxt8j\") pod \"community-operators-7q9f2\" (UID: \"1b040d3d-f4a0-451a-97de-08bea41c6dce\") " pod="openshift-marketplace/community-operators-7q9f2" Jan 21 11:13:17 crc kubenswrapper[4824]: E0121 11:13:17.127062 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-21 11:13:17.627036836 +0000 UTC m=+139.920066128 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 11:13:17 crc kubenswrapper[4824]: I0121 11:13:17.127129 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1b040d3d-f4a0-451a-97de-08bea41c6dce-catalog-content\") pod \"community-operators-7q9f2\" (UID: \"1b040d3d-f4a0-451a-97de-08bea41c6dce\") " pod="openshift-marketplace/community-operators-7q9f2" Jan 21 11:13:17 crc kubenswrapper[4824]: I0121 11:13:17.170741 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2027-01-21 11:08:16 +0000 UTC, rotation deadline is 2026-10-14 23:29:23.303812348 +0000 UTC Jan 21 11:13:17 crc kubenswrapper[4824]: I0121 11:13:17.170800 4824 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 6396h16m6.133015242s for next certificate rotation Jan 21 11:13:17 crc kubenswrapper[4824]: I0121 11:13:17.192700 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-b922x"] Jan 21 11:13:17 crc kubenswrapper[4824]: I0121 11:13:17.196047 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-b922x" Jan 21 11:13:17 crc kubenswrapper[4824]: I0121 11:13:17.198490 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Jan 21 11:13:17 crc kubenswrapper[4824]: I0121 11:13:17.205703 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-b922x"] Jan 21 11:13:17 crc kubenswrapper[4824]: I0121 11:13:17.228371 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1b040d3d-f4a0-451a-97de-08bea41c6dce-utilities\") pod \"community-operators-7q9f2\" (UID: \"1b040d3d-f4a0-451a-97de-08bea41c6dce\") " pod="openshift-marketplace/community-operators-7q9f2" Jan 21 11:13:17 crc kubenswrapper[4824]: I0121 11:13:17.228602 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1b040d3d-f4a0-451a-97de-08bea41c6dce-utilities\") pod \"community-operators-7q9f2\" (UID: \"1b040d3d-f4a0-451a-97de-08bea41c6dce\") " pod="openshift-marketplace/community-operators-7q9f2" Jan 21 11:13:17 crc kubenswrapper[4824]: I0121 11:13:17.228527 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mxt8j\" (UniqueName: \"kubernetes.io/projected/1b040d3d-f4a0-451a-97de-08bea41c6dce-kube-api-access-mxt8j\") pod \"community-operators-7q9f2\" (UID: \"1b040d3d-f4a0-451a-97de-08bea41c6dce\") " pod="openshift-marketplace/community-operators-7q9f2" Jan 21 11:13:17 crc kubenswrapper[4824]: I0121 11:13:17.228816 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxlxx\" (UID: \"94e71514-0af6-4254-8815-c933ade6c9da\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxlxx" Jan 21 11:13:17 crc kubenswrapper[4824]: I0121 11:13:17.228898 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1b040d3d-f4a0-451a-97de-08bea41c6dce-catalog-content\") pod \"community-operators-7q9f2\" (UID: \"1b040d3d-f4a0-451a-97de-08bea41c6dce\") " pod="openshift-marketplace/community-operators-7q9f2" Jan 21 11:13:17 crc kubenswrapper[4824]: E0121 11:13:17.229288 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-21 11:13:17.729275836 +0000 UTC m=+140.022305128 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxlxx" (UID: "94e71514-0af6-4254-8815-c933ade6c9da") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 11:13:17 crc kubenswrapper[4824]: I0121 11:13:17.229560 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1b040d3d-f4a0-451a-97de-08bea41c6dce-catalog-content\") pod \"community-operators-7q9f2\" (UID: \"1b040d3d-f4a0-451a-97de-08bea41c6dce\") " pod="openshift-marketplace/community-operators-7q9f2" Jan 21 11:13:17 crc kubenswrapper[4824]: I0121 11:13:17.232391 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-cslqn" Jan 21 11:13:17 crc kubenswrapper[4824]: I0121 11:13:17.250811 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mxt8j\" (UniqueName: \"kubernetes.io/projected/1b040d3d-f4a0-451a-97de-08bea41c6dce-kube-api-access-mxt8j\") pod \"community-operators-7q9f2\" (UID: \"1b040d3d-f4a0-451a-97de-08bea41c6dce\") " pod="openshift-marketplace/community-operators-7q9f2" Jan 21 11:13:17 crc kubenswrapper[4824]: I0121 11:13:17.320257 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7q9f2" Jan 21 11:13:17 crc kubenswrapper[4824]: I0121 11:13:17.331719 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 21 11:13:17 crc kubenswrapper[4824]: I0121 11:13:17.332014 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xjbt6\" (UniqueName: \"kubernetes.io/projected/3f11911f-e365-4151-b536-3988186b8757-kube-api-access-xjbt6\") pod \"certified-operators-b922x\" (UID: \"3f11911f-e365-4151-b536-3988186b8757\") " pod="openshift-marketplace/certified-operators-b922x" Jan 21 11:13:17 crc kubenswrapper[4824]: I0121 11:13:17.332070 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3f11911f-e365-4151-b536-3988186b8757-utilities\") pod \"certified-operators-b922x\" (UID: \"3f11911f-e365-4151-b536-3988186b8757\") " pod="openshift-marketplace/certified-operators-b922x" Jan 21 11:13:17 crc kubenswrapper[4824]: I0121 11:13:17.332099 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3f11911f-e365-4151-b536-3988186b8757-catalog-content\") pod \"certified-operators-b922x\" (UID: \"3f11911f-e365-4151-b536-3988186b8757\") " pod="openshift-marketplace/certified-operators-b922x" Jan 21 11:13:17 crc kubenswrapper[4824]: E0121 11:13:17.332214 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-21 11:13:17.832199001 +0000 UTC m=+140.125228293 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 11:13:17 crc kubenswrapper[4824]: I0121 11:13:17.374445 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-pckcl"] Jan 21 11:13:17 crc kubenswrapper[4824]: I0121 11:13:17.375474 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pckcl" Jan 21 11:13:17 crc kubenswrapper[4824]: I0121 11:13:17.404027 4824 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Jan 21 11:13:17 crc kubenswrapper[4824]: I0121 11:13:17.414284 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-pckcl"] Jan 21 11:13:17 crc kubenswrapper[4824]: I0121 11:13:17.433574 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxlxx\" (UID: \"94e71514-0af6-4254-8815-c933ade6c9da\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxlxx" Jan 21 11:13:17 crc kubenswrapper[4824]: I0121 11:13:17.433611 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xjbt6\" (UniqueName: \"kubernetes.io/projected/3f11911f-e365-4151-b536-3988186b8757-kube-api-access-xjbt6\") pod \"certified-operators-b922x\" (UID: \"3f11911f-e365-4151-b536-3988186b8757\") " pod="openshift-marketplace/certified-operators-b922x" Jan 21 11:13:17 crc kubenswrapper[4824]: I0121 11:13:17.433671 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3f11911f-e365-4151-b536-3988186b8757-utilities\") pod \"certified-operators-b922x\" (UID: \"3f11911f-e365-4151-b536-3988186b8757\") " pod="openshift-marketplace/certified-operators-b922x" Jan 21 11:13:17 crc kubenswrapper[4824]: I0121 11:13:17.433717 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3f11911f-e365-4151-b536-3988186b8757-catalog-content\") pod \"certified-operators-b922x\" (UID: \"3f11911f-e365-4151-b536-3988186b8757\") " pod="openshift-marketplace/certified-operators-b922x" Jan 21 11:13:17 crc kubenswrapper[4824]: I0121 11:13:17.434192 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3f11911f-e365-4151-b536-3988186b8757-catalog-content\") pod \"certified-operators-b922x\" (UID: \"3f11911f-e365-4151-b536-3988186b8757\") " pod="openshift-marketplace/certified-operators-b922x" Jan 21 11:13:17 crc kubenswrapper[4824]: E0121 11:13:17.434383 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-21 11:13:17.934364374 +0000 UTC m=+140.227393666 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxlxx" (UID: "94e71514-0af6-4254-8815-c933ade6c9da") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 11:13:17 crc kubenswrapper[4824]: I0121 11:13:17.434425 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3f11911f-e365-4151-b536-3988186b8757-utilities\") pod \"certified-operators-b922x\" (UID: \"3f11911f-e365-4151-b536-3988186b8757\") " pod="openshift-marketplace/certified-operators-b922x" Jan 21 11:13:17 crc kubenswrapper[4824]: I0121 11:13:17.455745 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xjbt6\" (UniqueName: \"kubernetes.io/projected/3f11911f-e365-4151-b536-3988186b8757-kube-api-access-xjbt6\") pod \"certified-operators-b922x\" (UID: \"3f11911f-e365-4151-b536-3988186b8757\") " pod="openshift-marketplace/certified-operators-b922x" Jan 21 11:13:17 crc kubenswrapper[4824]: I0121 11:13:17.467255 4824 patch_prober.go:28] interesting pod/router-default-5444994796-bblkp container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 21 11:13:17 crc kubenswrapper[4824]: [-]has-synced failed: reason withheld Jan 21 11:13:17 crc kubenswrapper[4824]: [+]process-running ok Jan 21 11:13:17 crc kubenswrapper[4824]: healthz check failed Jan 21 11:13:17 crc kubenswrapper[4824]: I0121 11:13:17.467315 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-bblkp" podUID="4b22a7aa-bcc0-4876-a374-a9943a6040bd" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 21 11:13:17 crc kubenswrapper[4824]: I0121 11:13:17.525010 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-b922x" Jan 21 11:13:17 crc kubenswrapper[4824]: I0121 11:13:17.534200 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-7q9f2"] Jan 21 11:13:17 crc kubenswrapper[4824]: I0121 11:13:17.534825 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 21 11:13:17 crc kubenswrapper[4824]: I0121 11:13:17.535121 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gd5wx\" (UniqueName: \"kubernetes.io/projected/0939eda4-fff7-447b-ad52-ddb427bf1272-kube-api-access-gd5wx\") pod \"community-operators-pckcl\" (UID: \"0939eda4-fff7-447b-ad52-ddb427bf1272\") " pod="openshift-marketplace/community-operators-pckcl" Jan 21 11:13:17 crc kubenswrapper[4824]: I0121 11:13:17.535204 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0939eda4-fff7-447b-ad52-ddb427bf1272-catalog-content\") pod \"community-operators-pckcl\" (UID: \"0939eda4-fff7-447b-ad52-ddb427bf1272\") " pod="openshift-marketplace/community-operators-pckcl" Jan 21 11:13:17 crc kubenswrapper[4824]: I0121 11:13:17.535244 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0939eda4-fff7-447b-ad52-ddb427bf1272-utilities\") pod \"community-operators-pckcl\" (UID: \"0939eda4-fff7-447b-ad52-ddb427bf1272\") " pod="openshift-marketplace/community-operators-pckcl" Jan 21 11:13:17 crc kubenswrapper[4824]: E0121 11:13:17.535354 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-21 11:13:18.03533927 +0000 UTC m=+140.328368562 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 11:13:17 crc kubenswrapper[4824]: W0121 11:13:17.548205 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1b040d3d_f4a0_451a_97de_08bea41c6dce.slice/crio-58f9378707e041408b55a43bb2ea6d9092a260cf0c9b8d81dc167499cf1c4702 WatchSource:0}: Error finding container 58f9378707e041408b55a43bb2ea6d9092a260cf0c9b8d81dc167499cf1c4702: Status 404 returned error can't find the container with id 58f9378707e041408b55a43bb2ea6d9092a260cf0c9b8d81dc167499cf1c4702 Jan 21 11:13:17 crc kubenswrapper[4824]: I0121 11:13:17.577876 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-s2hsr"] Jan 21 11:13:17 crc kubenswrapper[4824]: I0121 11:13:17.580454 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-s2hsr" Jan 21 11:13:17 crc kubenswrapper[4824]: I0121 11:13:17.581396 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-s2hsr"] Jan 21 11:13:17 crc kubenswrapper[4824]: I0121 11:13:17.636568 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gd5wx\" (UniqueName: \"kubernetes.io/projected/0939eda4-fff7-447b-ad52-ddb427bf1272-kube-api-access-gd5wx\") pod \"community-operators-pckcl\" (UID: \"0939eda4-fff7-447b-ad52-ddb427bf1272\") " pod="openshift-marketplace/community-operators-pckcl" Jan 21 11:13:17 crc kubenswrapper[4824]: I0121 11:13:17.636620 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxlxx\" (UID: \"94e71514-0af6-4254-8815-c933ade6c9da\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxlxx" Jan 21 11:13:17 crc kubenswrapper[4824]: I0121 11:13:17.636656 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0939eda4-fff7-447b-ad52-ddb427bf1272-catalog-content\") pod \"community-operators-pckcl\" (UID: \"0939eda4-fff7-447b-ad52-ddb427bf1272\") " pod="openshift-marketplace/community-operators-pckcl" Jan 21 11:13:17 crc kubenswrapper[4824]: I0121 11:13:17.636693 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0939eda4-fff7-447b-ad52-ddb427bf1272-utilities\") pod \"community-operators-pckcl\" (UID: \"0939eda4-fff7-447b-ad52-ddb427bf1272\") " pod="openshift-marketplace/community-operators-pckcl" Jan 21 11:13:17 crc kubenswrapper[4824]: I0121 11:13:17.637132 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0939eda4-fff7-447b-ad52-ddb427bf1272-utilities\") pod \"community-operators-pckcl\" (UID: \"0939eda4-fff7-447b-ad52-ddb427bf1272\") " pod="openshift-marketplace/community-operators-pckcl" Jan 21 11:13:17 crc kubenswrapper[4824]: E0121 11:13:17.637175 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-21 11:13:18.137145588 +0000 UTC m=+140.430174880 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxlxx" (UID: "94e71514-0af6-4254-8815-c933ade6c9da") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 11:13:17 crc kubenswrapper[4824]: I0121 11:13:17.637348 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0939eda4-fff7-447b-ad52-ddb427bf1272-catalog-content\") pod \"community-operators-pckcl\" (UID: \"0939eda4-fff7-447b-ad52-ddb427bf1272\") " pod="openshift-marketplace/community-operators-pckcl" Jan 21 11:13:17 crc kubenswrapper[4824]: I0121 11:13:17.655549 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gd5wx\" (UniqueName: \"kubernetes.io/projected/0939eda4-fff7-447b-ad52-ddb427bf1272-kube-api-access-gd5wx\") pod \"community-operators-pckcl\" (UID: \"0939eda4-fff7-447b-ad52-ddb427bf1272\") " pod="openshift-marketplace/community-operators-pckcl" Jan 21 11:13:17 crc kubenswrapper[4824]: I0121 11:13:17.692487 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pckcl" Jan 21 11:13:17 crc kubenswrapper[4824]: I0121 11:13:17.695646 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-jgvgl" event={"ID":"26739e4d-344c-4e25-bd0a-019f11456897","Type":"ContainerStarted","Data":"db772caf01df6719f648d34f4586836dd3d53aa61dba88808bffea38b508c47d"} Jan 21 11:13:17 crc kubenswrapper[4824]: I0121 11:13:17.695686 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-jgvgl" event={"ID":"26739e4d-344c-4e25-bd0a-019f11456897","Type":"ContainerStarted","Data":"b59554f789e277dc869c32d73511c732a6fce72010131678c5798f7d7d9489d4"} Jan 21 11:13:17 crc kubenswrapper[4824]: I0121 11:13:17.695697 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-jgvgl" event={"ID":"26739e4d-344c-4e25-bd0a-019f11456897","Type":"ContainerStarted","Data":"6585da42ef186bc5fcc0fb1c007fdf2eaf76ee83f09acbf799840ff171450fec"} Jan 21 11:13:17 crc kubenswrapper[4824]: I0121 11:13:17.698195 4824 generic.go:334] "Generic (PLEG): container finished" podID="1b040d3d-f4a0-451a-97de-08bea41c6dce" containerID="ca4b551b3e03faea4b176d4748d261efd694766575aea5c759f3fe832098d94c" exitCode=0 Jan 21 11:13:17 crc kubenswrapper[4824]: I0121 11:13:17.698578 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7q9f2" event={"ID":"1b040d3d-f4a0-451a-97de-08bea41c6dce","Type":"ContainerDied","Data":"ca4b551b3e03faea4b176d4748d261efd694766575aea5c759f3fe832098d94c"} Jan 21 11:13:17 crc kubenswrapper[4824]: I0121 11:13:17.698629 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7q9f2" event={"ID":"1b040d3d-f4a0-451a-97de-08bea41c6dce","Type":"ContainerStarted","Data":"58f9378707e041408b55a43bb2ea6d9092a260cf0c9b8d81dc167499cf1c4702"} Jan 21 11:13:17 crc kubenswrapper[4824]: I0121 11:13:17.699113 4824 patch_prober.go:28] interesting pod/downloads-7954f5f757-8g4qk container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.16:8080/\": dial tcp 10.217.0.16:8080: connect: connection refused" start-of-body= Jan 21 11:13:17 crc kubenswrapper[4824]: I0121 11:13:17.699140 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-8g4qk" podUID="2279132e-c274-424b-8c8b-80ae5bd7dfe5" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.16:8080/\": dial tcp 10.217.0.16:8080: connect: connection refused" Jan 21 11:13:17 crc kubenswrapper[4824]: I0121 11:13:17.699796 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-fbq6l" Jan 21 11:13:17 crc kubenswrapper[4824]: I0121 11:13:17.703227 4824 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 21 11:13:17 crc kubenswrapper[4824]: I0121 11:13:17.703351 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-gjhp5" Jan 21 11:13:17 crc kubenswrapper[4824]: I0121 11:13:17.705236 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-b922x"] Jan 21 11:13:17 crc kubenswrapper[4824]: I0121 11:13:17.705511 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7s7wx" Jan 21 11:13:17 crc kubenswrapper[4824]: I0121 11:13:17.714508 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-jgvgl" podStartSLOduration=8.712970503 podStartE2EDuration="8.712970503s" podCreationTimestamp="2026-01-21 11:13:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:13:17.711290378 +0000 UTC m=+140.004319670" watchObservedRunningTime="2026-01-21 11:13:17.712970503 +0000 UTC m=+140.005999795" Jan 21 11:13:17 crc kubenswrapper[4824]: W0121 11:13:17.718674 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3f11911f_e365_4151_b536_3988186b8757.slice/crio-8980ea76a1ff2b34ee51f0465162ae2c030d8714d1e864ebb767d3d7a1da8509 WatchSource:0}: Error finding container 8980ea76a1ff2b34ee51f0465162ae2c030d8714d1e864ebb767d3d7a1da8509: Status 404 returned error can't find the container with id 8980ea76a1ff2b34ee51f0465162ae2c030d8714d1e864ebb767d3d7a1da8509 Jan 21 11:13:17 crc kubenswrapper[4824]: I0121 11:13:17.739463 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 21 11:13:17 crc kubenswrapper[4824]: I0121 11:13:17.739679 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mtft2\" (UniqueName: \"kubernetes.io/projected/4205fbde-9c6c-40c8-8c8c-d34a607923e4-kube-api-access-mtft2\") pod \"certified-operators-s2hsr\" (UID: \"4205fbde-9c6c-40c8-8c8c-d34a607923e4\") " pod="openshift-marketplace/certified-operators-s2hsr" Jan 21 11:13:17 crc kubenswrapper[4824]: I0121 11:13:17.739713 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4205fbde-9c6c-40c8-8c8c-d34a607923e4-catalog-content\") pod \"certified-operators-s2hsr\" (UID: \"4205fbde-9c6c-40c8-8c8c-d34a607923e4\") " pod="openshift-marketplace/certified-operators-s2hsr" Jan 21 11:13:17 crc kubenswrapper[4824]: I0121 11:13:17.739749 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4205fbde-9c6c-40c8-8c8c-d34a607923e4-utilities\") pod \"certified-operators-s2hsr\" (UID: \"4205fbde-9c6c-40c8-8c8c-d34a607923e4\") " pod="openshift-marketplace/certified-operators-s2hsr" Jan 21 11:13:17 crc kubenswrapper[4824]: E0121 11:13:17.739932 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-21 11:13:18.239911587 +0000 UTC m=+140.532940879 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 11:13:17 crc kubenswrapper[4824]: I0121 11:13:17.841076 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxlxx\" (UID: \"94e71514-0af6-4254-8815-c933ade6c9da\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxlxx" Jan 21 11:13:17 crc kubenswrapper[4824]: I0121 11:13:17.841431 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mtft2\" (UniqueName: \"kubernetes.io/projected/4205fbde-9c6c-40c8-8c8c-d34a607923e4-kube-api-access-mtft2\") pod \"certified-operators-s2hsr\" (UID: \"4205fbde-9c6c-40c8-8c8c-d34a607923e4\") " pod="openshift-marketplace/certified-operators-s2hsr" Jan 21 11:13:17 crc kubenswrapper[4824]: I0121 11:13:17.841505 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4205fbde-9c6c-40c8-8c8c-d34a607923e4-catalog-content\") pod \"certified-operators-s2hsr\" (UID: \"4205fbde-9c6c-40c8-8c8c-d34a607923e4\") " pod="openshift-marketplace/certified-operators-s2hsr" Jan 21 11:13:17 crc kubenswrapper[4824]: I0121 11:13:17.841670 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4205fbde-9c6c-40c8-8c8c-d34a607923e4-utilities\") pod \"certified-operators-s2hsr\" (UID: \"4205fbde-9c6c-40c8-8c8c-d34a607923e4\") " pod="openshift-marketplace/certified-operators-s2hsr" Jan 21 11:13:17 crc kubenswrapper[4824]: E0121 11:13:17.842180 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-21 11:13:18.342163623 +0000 UTC m=+140.635192914 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxlxx" (UID: "94e71514-0af6-4254-8815-c933ade6c9da") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 11:13:17 crc kubenswrapper[4824]: I0121 11:13:17.844648 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4205fbde-9c6c-40c8-8c8c-d34a607923e4-catalog-content\") pod \"certified-operators-s2hsr\" (UID: \"4205fbde-9c6c-40c8-8c8c-d34a607923e4\") " pod="openshift-marketplace/certified-operators-s2hsr" Jan 21 11:13:17 crc kubenswrapper[4824]: I0121 11:13:17.845185 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4205fbde-9c6c-40c8-8c8c-d34a607923e4-utilities\") pod \"certified-operators-s2hsr\" (UID: \"4205fbde-9c6c-40c8-8c8c-d34a607923e4\") " pod="openshift-marketplace/certified-operators-s2hsr" Jan 21 11:13:17 crc kubenswrapper[4824]: I0121 11:13:17.872296 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mtft2\" (UniqueName: \"kubernetes.io/projected/4205fbde-9c6c-40c8-8c8c-d34a607923e4-kube-api-access-mtft2\") pod \"certified-operators-s2hsr\" (UID: \"4205fbde-9c6c-40c8-8c8c-d34a607923e4\") " pod="openshift-marketplace/certified-operators-s2hsr" Jan 21 11:13:17 crc kubenswrapper[4824]: I0121 11:13:17.902827 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-s2hsr" Jan 21 11:13:17 crc kubenswrapper[4824]: I0121 11:13:17.925800 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-pckcl"] Jan 21 11:13:17 crc kubenswrapper[4824]: I0121 11:13:17.944129 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 21 11:13:17 crc kubenswrapper[4824]: E0121 11:13:17.944532 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-21 11:13:18.444516638 +0000 UTC m=+140.737545929 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 11:13:18 crc kubenswrapper[4824]: I0121 11:13:18.045843 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxlxx\" (UID: \"94e71514-0af6-4254-8815-c933ade6c9da\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxlxx" Jan 21 11:13:18 crc kubenswrapper[4824]: E0121 11:13:18.046200 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-21 11:13:18.546185997 +0000 UTC m=+140.839215290 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxlxx" (UID: "94e71514-0af6-4254-8815-c933ade6c9da") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 11:13:18 crc kubenswrapper[4824]: I0121 11:13:18.127861 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-s2hsr"] Jan 21 11:13:18 crc kubenswrapper[4824]: W0121 11:13:18.145636 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4205fbde_9c6c_40c8_8c8c_d34a607923e4.slice/crio-f85af8c50c5e5798e08394dc9af278d2e32f05c7d0d8ddcab84b8549c975347a WatchSource:0}: Error finding container f85af8c50c5e5798e08394dc9af278d2e32f05c7d0d8ddcab84b8549c975347a: Status 404 returned error can't find the container with id f85af8c50c5e5798e08394dc9af278d2e32f05c7d0d8ddcab84b8549c975347a Jan 21 11:13:18 crc kubenswrapper[4824]: I0121 11:13:18.148077 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 21 11:13:18 crc kubenswrapper[4824]: E0121 11:13:18.148226 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-21 11:13:18.648210646 +0000 UTC m=+140.941239937 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 11:13:18 crc kubenswrapper[4824]: I0121 11:13:18.148371 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxlxx\" (UID: \"94e71514-0af6-4254-8815-c933ade6c9da\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxlxx" Jan 21 11:13:18 crc kubenswrapper[4824]: E0121 11:13:18.148613 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-21 11:13:18.648603693 +0000 UTC m=+140.941632995 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxlxx" (UID: "94e71514-0af6-4254-8815-c933ade6c9da") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 21 11:13:18 crc kubenswrapper[4824]: I0121 11:13:18.163242 4824 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2026-01-21T11:13:17.404276439Z","Handler":null,"Name":""} Jan 21 11:13:18 crc kubenswrapper[4824]: I0121 11:13:18.165351 4824 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Jan 21 11:13:18 crc kubenswrapper[4824]: I0121 11:13:18.165490 4824 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Jan 21 11:13:18 crc kubenswrapper[4824]: I0121 11:13:18.249894 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 21 11:13:18 crc kubenswrapper[4824]: I0121 11:13:18.253772 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Jan 21 11:13:18 crc kubenswrapper[4824]: I0121 11:13:18.351631 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxlxx\" (UID: \"94e71514-0af6-4254-8815-c933ade6c9da\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxlxx" Jan 21 11:13:18 crc kubenswrapper[4824]: I0121 11:13:18.364371 4824 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Jan 21 11:13:18 crc kubenswrapper[4824]: I0121 11:13:18.364412 4824 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxlxx\" (UID: \"94e71514-0af6-4254-8815-c933ade6c9da\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-wxlxx" Jan 21 11:13:18 crc kubenswrapper[4824]: I0121 11:13:18.382750 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxlxx\" (UID: \"94e71514-0af6-4254-8815-c933ade6c9da\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxlxx" Jan 21 11:13:18 crc kubenswrapper[4824]: I0121 11:13:18.463617 4824 patch_prober.go:28] interesting pod/router-default-5444994796-bblkp container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 21 11:13:18 crc kubenswrapper[4824]: [-]has-synced failed: reason withheld Jan 21 11:13:18 crc kubenswrapper[4824]: [+]process-running ok Jan 21 11:13:18 crc kubenswrapper[4824]: healthz check failed Jan 21 11:13:18 crc kubenswrapper[4824]: I0121 11:13:18.463681 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-bblkp" podUID="4b22a7aa-bcc0-4876-a374-a9943a6040bd" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 21 11:13:18 crc kubenswrapper[4824]: I0121 11:13:18.658632 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-wxlxx" Jan 21 11:13:18 crc kubenswrapper[4824]: I0121 11:13:18.708240 4824 generic.go:334] "Generic (PLEG): container finished" podID="0939eda4-fff7-447b-ad52-ddb427bf1272" containerID="f2ddd6ed1029c405c4138f6257aaa84375743b3602a5b630fdf5b86e73651795" exitCode=0 Jan 21 11:13:18 crc kubenswrapper[4824]: I0121 11:13:18.708316 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pckcl" event={"ID":"0939eda4-fff7-447b-ad52-ddb427bf1272","Type":"ContainerDied","Data":"f2ddd6ed1029c405c4138f6257aaa84375743b3602a5b630fdf5b86e73651795"} Jan 21 11:13:18 crc kubenswrapper[4824]: I0121 11:13:18.708345 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pckcl" event={"ID":"0939eda4-fff7-447b-ad52-ddb427bf1272","Type":"ContainerStarted","Data":"0fe9a2ce9dc5eb2c88168927f6206b6e2577521e0d1f3b71800ade62c141b6ff"} Jan 21 11:13:18 crc kubenswrapper[4824]: I0121 11:13:18.725607 4824 generic.go:334] "Generic (PLEG): container finished" podID="4205fbde-9c6c-40c8-8c8c-d34a607923e4" containerID="12e7f9c58f1247fa3830ac89333e154e7e3fd0757d08bc1ecbe6ca9ac5f27bd0" exitCode=0 Jan 21 11:13:18 crc kubenswrapper[4824]: I0121 11:13:18.725689 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-s2hsr" event={"ID":"4205fbde-9c6c-40c8-8c8c-d34a607923e4","Type":"ContainerDied","Data":"12e7f9c58f1247fa3830ac89333e154e7e3fd0757d08bc1ecbe6ca9ac5f27bd0"} Jan 21 11:13:18 crc kubenswrapper[4824]: I0121 11:13:18.725719 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-s2hsr" event={"ID":"4205fbde-9c6c-40c8-8c8c-d34a607923e4","Type":"ContainerStarted","Data":"f85af8c50c5e5798e08394dc9af278d2e32f05c7d0d8ddcab84b8549c975347a"} Jan 21 11:13:18 crc kubenswrapper[4824]: I0121 11:13:18.734552 4824 generic.go:334] "Generic (PLEG): container finished" podID="3f11911f-e365-4151-b536-3988186b8757" containerID="fe402f640205024f989b1231ac3c21b31c3d7e53bed03337b9aa7afe32b7848e" exitCode=0 Jan 21 11:13:18 crc kubenswrapper[4824]: I0121 11:13:18.734628 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-b922x" event={"ID":"3f11911f-e365-4151-b536-3988186b8757","Type":"ContainerDied","Data":"fe402f640205024f989b1231ac3c21b31c3d7e53bed03337b9aa7afe32b7848e"} Jan 21 11:13:18 crc kubenswrapper[4824]: I0121 11:13:18.734656 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-b922x" event={"ID":"3f11911f-e365-4151-b536-3988186b8757","Type":"ContainerStarted","Data":"8980ea76a1ff2b34ee51f0465162ae2c030d8714d1e864ebb767d3d7a1da8509"} Jan 21 11:13:18 crc kubenswrapper[4824]: I0121 11:13:18.736529 4824 generic.go:334] "Generic (PLEG): container finished" podID="09bd0b9c-46fb-433f-bd30-014c69c90d9f" containerID="277011d8fed4b1333d53f5a4a084181f52dada1e6d85b31750d9bab9937b3678" exitCode=0 Jan 21 11:13:18 crc kubenswrapper[4824]: I0121 11:13:18.736984 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29483220-6r4fz" event={"ID":"09bd0b9c-46fb-433f-bd30-014c69c90d9f","Type":"ContainerDied","Data":"277011d8fed4b1333d53f5a4a084181f52dada1e6d85b31750d9bab9937b3678"} Jan 21 11:13:18 crc kubenswrapper[4824]: I0121 11:13:18.749077 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-fbq6l" Jan 21 11:13:19 crc kubenswrapper[4824]: I0121 11:13:19.019573 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-wxlxx"] Jan 21 11:13:19 crc kubenswrapper[4824]: W0121 11:13:19.037008 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod94e71514_0af6_4254_8815_c933ade6c9da.slice/crio-7d8d535dae1a2ab198548905f9a9571895c25a2b895cfb327e93cc9833852054 WatchSource:0}: Error finding container 7d8d535dae1a2ab198548905f9a9571895c25a2b895cfb327e93cc9833852054: Status 404 returned error can't find the container with id 7d8d535dae1a2ab198548905f9a9571895c25a2b895cfb327e93cc9833852054 Jan 21 11:13:19 crc kubenswrapper[4824]: I0121 11:13:19.168906 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-4w5m6"] Jan 21 11:13:19 crc kubenswrapper[4824]: I0121 11:13:19.171386 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4w5m6" Jan 21 11:13:19 crc kubenswrapper[4824]: I0121 11:13:19.175064 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Jan 21 11:13:19 crc kubenswrapper[4824]: I0121 11:13:19.176626 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-4w5m6"] Jan 21 11:13:19 crc kubenswrapper[4824]: I0121 11:13:19.365365 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1bc8c652-fb4f-48b7-86ce-c45522dfd476-catalog-content\") pod \"redhat-marketplace-4w5m6\" (UID: \"1bc8c652-fb4f-48b7-86ce-c45522dfd476\") " pod="openshift-marketplace/redhat-marketplace-4w5m6" Jan 21 11:13:19 crc kubenswrapper[4824]: I0121 11:13:19.365441 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1bc8c652-fb4f-48b7-86ce-c45522dfd476-utilities\") pod \"redhat-marketplace-4w5m6\" (UID: \"1bc8c652-fb4f-48b7-86ce-c45522dfd476\") " pod="openshift-marketplace/redhat-marketplace-4w5m6" Jan 21 11:13:19 crc kubenswrapper[4824]: I0121 11:13:19.365499 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lchhw\" (UniqueName: \"kubernetes.io/projected/1bc8c652-fb4f-48b7-86ce-c45522dfd476-kube-api-access-lchhw\") pod \"redhat-marketplace-4w5m6\" (UID: \"1bc8c652-fb4f-48b7-86ce-c45522dfd476\") " pod="openshift-marketplace/redhat-marketplace-4w5m6" Jan 21 11:13:19 crc kubenswrapper[4824]: I0121 11:13:19.463544 4824 patch_prober.go:28] interesting pod/router-default-5444994796-bblkp container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 21 11:13:19 crc kubenswrapper[4824]: [-]has-synced failed: reason withheld Jan 21 11:13:19 crc kubenswrapper[4824]: [+]process-running ok Jan 21 11:13:19 crc kubenswrapper[4824]: healthz check failed Jan 21 11:13:19 crc kubenswrapper[4824]: I0121 11:13:19.463601 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-bblkp" podUID="4b22a7aa-bcc0-4876-a374-a9943a6040bd" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 21 11:13:19 crc kubenswrapper[4824]: I0121 11:13:19.474922 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1bc8c652-fb4f-48b7-86ce-c45522dfd476-utilities\") pod \"redhat-marketplace-4w5m6\" (UID: \"1bc8c652-fb4f-48b7-86ce-c45522dfd476\") " pod="openshift-marketplace/redhat-marketplace-4w5m6" Jan 21 11:13:19 crc kubenswrapper[4824]: I0121 11:13:19.475302 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lchhw\" (UniqueName: \"kubernetes.io/projected/1bc8c652-fb4f-48b7-86ce-c45522dfd476-kube-api-access-lchhw\") pod \"redhat-marketplace-4w5m6\" (UID: \"1bc8c652-fb4f-48b7-86ce-c45522dfd476\") " pod="openshift-marketplace/redhat-marketplace-4w5m6" Jan 21 11:13:19 crc kubenswrapper[4824]: I0121 11:13:19.475970 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1bc8c652-fb4f-48b7-86ce-c45522dfd476-utilities\") pod \"redhat-marketplace-4w5m6\" (UID: \"1bc8c652-fb4f-48b7-86ce-c45522dfd476\") " pod="openshift-marketplace/redhat-marketplace-4w5m6" Jan 21 11:13:19 crc kubenswrapper[4824]: I0121 11:13:19.475887 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1bc8c652-fb4f-48b7-86ce-c45522dfd476-catalog-content\") pod \"redhat-marketplace-4w5m6\" (UID: \"1bc8c652-fb4f-48b7-86ce-c45522dfd476\") " pod="openshift-marketplace/redhat-marketplace-4w5m6" Jan 21 11:13:19 crc kubenswrapper[4824]: I0121 11:13:19.476204 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1bc8c652-fb4f-48b7-86ce-c45522dfd476-catalog-content\") pod \"redhat-marketplace-4w5m6\" (UID: \"1bc8c652-fb4f-48b7-86ce-c45522dfd476\") " pod="openshift-marketplace/redhat-marketplace-4w5m6" Jan 21 11:13:19 crc kubenswrapper[4824]: I0121 11:13:19.515278 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lchhw\" (UniqueName: \"kubernetes.io/projected/1bc8c652-fb4f-48b7-86ce-c45522dfd476-kube-api-access-lchhw\") pod \"redhat-marketplace-4w5m6\" (UID: \"1bc8c652-fb4f-48b7-86ce-c45522dfd476\") " pod="openshift-marketplace/redhat-marketplace-4w5m6" Jan 21 11:13:19 crc kubenswrapper[4824]: I0121 11:13:19.573161 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-8dj94"] Jan 21 11:13:19 crc kubenswrapper[4824]: I0121 11:13:19.576621 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8dj94" Jan 21 11:13:19 crc kubenswrapper[4824]: I0121 11:13:19.584715 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-8dj94"] Jan 21 11:13:19 crc kubenswrapper[4824]: I0121 11:13:19.678738 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ff7a0886-d523-486a-9bf0-bae3456b85c2-catalog-content\") pod \"redhat-marketplace-8dj94\" (UID: \"ff7a0886-d523-486a-9bf0-bae3456b85c2\") " pod="openshift-marketplace/redhat-marketplace-8dj94" Jan 21 11:13:19 crc kubenswrapper[4824]: I0121 11:13:19.679050 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ff7a0886-d523-486a-9bf0-bae3456b85c2-utilities\") pod \"redhat-marketplace-8dj94\" (UID: \"ff7a0886-d523-486a-9bf0-bae3456b85c2\") " pod="openshift-marketplace/redhat-marketplace-8dj94" Jan 21 11:13:19 crc kubenswrapper[4824]: I0121 11:13:19.679241 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tskqq\" (UniqueName: \"kubernetes.io/projected/ff7a0886-d523-486a-9bf0-bae3456b85c2-kube-api-access-tskqq\") pod \"redhat-marketplace-8dj94\" (UID: \"ff7a0886-d523-486a-9bf0-bae3456b85c2\") " pod="openshift-marketplace/redhat-marketplace-8dj94" Jan 21 11:13:19 crc kubenswrapper[4824]: I0121 11:13:19.752400 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-wxlxx" event={"ID":"94e71514-0af6-4254-8815-c933ade6c9da","Type":"ContainerStarted","Data":"6c3387d6526e3b087ac4042dcb88df75f294f6b26a6635786027622ac70f54bf"} Jan 21 11:13:19 crc kubenswrapper[4824]: I0121 11:13:19.752444 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-wxlxx" event={"ID":"94e71514-0af6-4254-8815-c933ade6c9da","Type":"ContainerStarted","Data":"7d8d535dae1a2ab198548905f9a9571895c25a2b895cfb327e93cc9833852054"} Jan 21 11:13:19 crc kubenswrapper[4824]: I0121 11:13:19.752620 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-wxlxx" Jan 21 11:13:19 crc kubenswrapper[4824]: I0121 11:13:19.786208 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-wxlxx" podStartSLOduration=124.786187885 podStartE2EDuration="2m4.786187885s" podCreationTimestamp="2026-01-21 11:11:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:13:19.781783293 +0000 UTC m=+142.074812585" watchObservedRunningTime="2026-01-21 11:13:19.786187885 +0000 UTC m=+142.079217177" Jan 21 11:13:19 crc kubenswrapper[4824]: I0121 11:13:19.788937 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4w5m6" Jan 21 11:13:19 crc kubenswrapper[4824]: I0121 11:13:19.792071 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ff7a0886-d523-486a-9bf0-bae3456b85c2-catalog-content\") pod \"redhat-marketplace-8dj94\" (UID: \"ff7a0886-d523-486a-9bf0-bae3456b85c2\") " pod="openshift-marketplace/redhat-marketplace-8dj94" Jan 21 11:13:19 crc kubenswrapper[4824]: I0121 11:13:19.792189 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ff7a0886-d523-486a-9bf0-bae3456b85c2-utilities\") pod \"redhat-marketplace-8dj94\" (UID: \"ff7a0886-d523-486a-9bf0-bae3456b85c2\") " pod="openshift-marketplace/redhat-marketplace-8dj94" Jan 21 11:13:19 crc kubenswrapper[4824]: I0121 11:13:19.792404 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tskqq\" (UniqueName: \"kubernetes.io/projected/ff7a0886-d523-486a-9bf0-bae3456b85c2-kube-api-access-tskqq\") pod \"redhat-marketplace-8dj94\" (UID: \"ff7a0886-d523-486a-9bf0-bae3456b85c2\") " pod="openshift-marketplace/redhat-marketplace-8dj94" Jan 21 11:13:19 crc kubenswrapper[4824]: I0121 11:13:19.792707 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ff7a0886-d523-486a-9bf0-bae3456b85c2-catalog-content\") pod \"redhat-marketplace-8dj94\" (UID: \"ff7a0886-d523-486a-9bf0-bae3456b85c2\") " pod="openshift-marketplace/redhat-marketplace-8dj94" Jan 21 11:13:19 crc kubenswrapper[4824]: I0121 11:13:19.793335 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ff7a0886-d523-486a-9bf0-bae3456b85c2-utilities\") pod \"redhat-marketplace-8dj94\" (UID: \"ff7a0886-d523-486a-9bf0-bae3456b85c2\") " pod="openshift-marketplace/redhat-marketplace-8dj94" Jan 21 11:13:19 crc kubenswrapper[4824]: I0121 11:13:19.809251 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tskqq\" (UniqueName: \"kubernetes.io/projected/ff7a0886-d523-486a-9bf0-bae3456b85c2-kube-api-access-tskqq\") pod \"redhat-marketplace-8dj94\" (UID: \"ff7a0886-d523-486a-9bf0-bae3456b85c2\") " pod="openshift-marketplace/redhat-marketplace-8dj94" Jan 21 11:13:19 crc kubenswrapper[4824]: I0121 11:13:19.894580 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8dj94" Jan 21 11:13:20 crc kubenswrapper[4824]: I0121 11:13:20.081361 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Jan 21 11:13:20 crc kubenswrapper[4824]: I0121 11:13:20.082258 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-4w5m6"] Jan 21 11:13:20 crc kubenswrapper[4824]: I0121 11:13:20.090120 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29483220-6r4fz" Jan 21 11:13:20 crc kubenswrapper[4824]: W0121 11:13:20.093176 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1bc8c652_fb4f_48b7_86ce_c45522dfd476.slice/crio-1717bf3078cfbe6fee3800d069a35c471757cb03a1eddb1d8ff1e20167f89c9e WatchSource:0}: Error finding container 1717bf3078cfbe6fee3800d069a35c471757cb03a1eddb1d8ff1e20167f89c9e: Status 404 returned error can't find the container with id 1717bf3078cfbe6fee3800d069a35c471757cb03a1eddb1d8ff1e20167f89c9e Jan 21 11:13:20 crc kubenswrapper[4824]: I0121 11:13:20.142740 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Jan 21 11:13:20 crc kubenswrapper[4824]: E0121 11:13:20.143051 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="09bd0b9c-46fb-433f-bd30-014c69c90d9f" containerName="collect-profiles" Jan 21 11:13:20 crc kubenswrapper[4824]: I0121 11:13:20.143065 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="09bd0b9c-46fb-433f-bd30-014c69c90d9f" containerName="collect-profiles" Jan 21 11:13:20 crc kubenswrapper[4824]: I0121 11:13:20.143185 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="09bd0b9c-46fb-433f-bd30-014c69c90d9f" containerName="collect-profiles" Jan 21 11:13:20 crc kubenswrapper[4824]: I0121 11:13:20.143570 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 21 11:13:20 crc kubenswrapper[4824]: I0121 11:13:20.145269 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Jan 21 11:13:20 crc kubenswrapper[4824]: I0121 11:13:20.146301 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Jan 21 11:13:20 crc kubenswrapper[4824]: I0121 11:13:20.146400 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Jan 21 11:13:20 crc kubenswrapper[4824]: I0121 11:13:20.168665 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-hqs52"] Jan 21 11:13:20 crc kubenswrapper[4824]: I0121 11:13:20.169681 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hqs52" Jan 21 11:13:20 crc kubenswrapper[4824]: I0121 11:13:20.171704 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Jan 21 11:13:20 crc kubenswrapper[4824]: I0121 11:13:20.177276 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-hqs52"] Jan 21 11:13:20 crc kubenswrapper[4824]: I0121 11:13:20.199361 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/09bd0b9c-46fb-433f-bd30-014c69c90d9f-config-volume\") pod \"09bd0b9c-46fb-433f-bd30-014c69c90d9f\" (UID: \"09bd0b9c-46fb-433f-bd30-014c69c90d9f\") " Jan 21 11:13:20 crc kubenswrapper[4824]: I0121 11:13:20.199393 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/09bd0b9c-46fb-433f-bd30-014c69c90d9f-secret-volume\") pod \"09bd0b9c-46fb-433f-bd30-014c69c90d9f\" (UID: \"09bd0b9c-46fb-433f-bd30-014c69c90d9f\") " Jan 21 11:13:20 crc kubenswrapper[4824]: I0121 11:13:20.200260 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nqqww\" (UniqueName: \"kubernetes.io/projected/09bd0b9c-46fb-433f-bd30-014c69c90d9f-kube-api-access-nqqww\") pod \"09bd0b9c-46fb-433f-bd30-014c69c90d9f\" (UID: \"09bd0b9c-46fb-433f-bd30-014c69c90d9f\") " Jan 21 11:13:20 crc kubenswrapper[4824]: I0121 11:13:20.200245 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09bd0b9c-46fb-433f-bd30-014c69c90d9f-config-volume" (OuterVolumeSpecName: "config-volume") pod "09bd0b9c-46fb-433f-bd30-014c69c90d9f" (UID: "09bd0b9c-46fb-433f-bd30-014c69c90d9f"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:13:20 crc kubenswrapper[4824]: I0121 11:13:20.205999 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09bd0b9c-46fb-433f-bd30-014c69c90d9f-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "09bd0b9c-46fb-433f-bd30-014c69c90d9f" (UID: "09bd0b9c-46fb-433f-bd30-014c69c90d9f"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:13:20 crc kubenswrapper[4824]: I0121 11:13:20.206131 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09bd0b9c-46fb-433f-bd30-014c69c90d9f-kube-api-access-nqqww" (OuterVolumeSpecName: "kube-api-access-nqqww") pod "09bd0b9c-46fb-433f-bd30-014c69c90d9f" (UID: "09bd0b9c-46fb-433f-bd30-014c69c90d9f"). InnerVolumeSpecName "kube-api-access-nqqww". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:13:20 crc kubenswrapper[4824]: I0121 11:13:20.302182 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c3b56f2d-96c0-4597-9235-147fc4802d19-utilities\") pod \"redhat-operators-hqs52\" (UID: \"c3b56f2d-96c0-4597-9235-147fc4802d19\") " pod="openshift-marketplace/redhat-operators-hqs52" Jan 21 11:13:20 crc kubenswrapper[4824]: I0121 11:13:20.302512 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/5b5068e5-8cc3-4ed1-811c-835a08eefaa0-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"5b5068e5-8cc3-4ed1-811c-835a08eefaa0\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 21 11:13:20 crc kubenswrapper[4824]: I0121 11:13:20.302701 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c3b56f2d-96c0-4597-9235-147fc4802d19-catalog-content\") pod \"redhat-operators-hqs52\" (UID: \"c3b56f2d-96c0-4597-9235-147fc4802d19\") " pod="openshift-marketplace/redhat-operators-hqs52" Jan 21 11:13:20 crc kubenswrapper[4824]: I0121 11:13:20.302805 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8656b\" (UniqueName: \"kubernetes.io/projected/c3b56f2d-96c0-4597-9235-147fc4802d19-kube-api-access-8656b\") pod \"redhat-operators-hqs52\" (UID: \"c3b56f2d-96c0-4597-9235-147fc4802d19\") " pod="openshift-marketplace/redhat-operators-hqs52" Jan 21 11:13:20 crc kubenswrapper[4824]: I0121 11:13:20.302842 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5b5068e5-8cc3-4ed1-811c-835a08eefaa0-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"5b5068e5-8cc3-4ed1-811c-835a08eefaa0\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 21 11:13:20 crc kubenswrapper[4824]: I0121 11:13:20.303011 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nqqww\" (UniqueName: \"kubernetes.io/projected/09bd0b9c-46fb-433f-bd30-014c69c90d9f-kube-api-access-nqqww\") on node \"crc\" DevicePath \"\"" Jan 21 11:13:20 crc kubenswrapper[4824]: I0121 11:13:20.303040 4824 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/09bd0b9c-46fb-433f-bd30-014c69c90d9f-config-volume\") on node \"crc\" DevicePath \"\"" Jan 21 11:13:20 crc kubenswrapper[4824]: I0121 11:13:20.303052 4824 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/09bd0b9c-46fb-433f-bd30-014c69c90d9f-secret-volume\") on node \"crc\" DevicePath \"\"" Jan 21 11:13:20 crc kubenswrapper[4824]: I0121 11:13:20.374020 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-8dj94"] Jan 21 11:13:20 crc kubenswrapper[4824]: W0121 11:13:20.393748 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podff7a0886_d523_486a_9bf0_bae3456b85c2.slice/crio-1cc0a8fad4ae405d30baab5682d5e3fd0a25f77c99ddd8f27c737f65b8432c15 WatchSource:0}: Error finding container 1cc0a8fad4ae405d30baab5682d5e3fd0a25f77c99ddd8f27c737f65b8432c15: Status 404 returned error can't find the container with id 1cc0a8fad4ae405d30baab5682d5e3fd0a25f77c99ddd8f27c737f65b8432c15 Jan 21 11:13:20 crc kubenswrapper[4824]: I0121 11:13:20.403877 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8656b\" (UniqueName: \"kubernetes.io/projected/c3b56f2d-96c0-4597-9235-147fc4802d19-kube-api-access-8656b\") pod \"redhat-operators-hqs52\" (UID: \"c3b56f2d-96c0-4597-9235-147fc4802d19\") " pod="openshift-marketplace/redhat-operators-hqs52" Jan 21 11:13:20 crc kubenswrapper[4824]: I0121 11:13:20.403942 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5b5068e5-8cc3-4ed1-811c-835a08eefaa0-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"5b5068e5-8cc3-4ed1-811c-835a08eefaa0\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 21 11:13:20 crc kubenswrapper[4824]: I0121 11:13:20.403983 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c3b56f2d-96c0-4597-9235-147fc4802d19-utilities\") pod \"redhat-operators-hqs52\" (UID: \"c3b56f2d-96c0-4597-9235-147fc4802d19\") " pod="openshift-marketplace/redhat-operators-hqs52" Jan 21 11:13:20 crc kubenswrapper[4824]: I0121 11:13:20.404003 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/5b5068e5-8cc3-4ed1-811c-835a08eefaa0-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"5b5068e5-8cc3-4ed1-811c-835a08eefaa0\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 21 11:13:20 crc kubenswrapper[4824]: I0121 11:13:20.404070 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c3b56f2d-96c0-4597-9235-147fc4802d19-catalog-content\") pod \"redhat-operators-hqs52\" (UID: \"c3b56f2d-96c0-4597-9235-147fc4802d19\") " pod="openshift-marketplace/redhat-operators-hqs52" Jan 21 11:13:20 crc kubenswrapper[4824]: I0121 11:13:20.404447 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/5b5068e5-8cc3-4ed1-811c-835a08eefaa0-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"5b5068e5-8cc3-4ed1-811c-835a08eefaa0\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 21 11:13:20 crc kubenswrapper[4824]: I0121 11:13:20.405327 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c3b56f2d-96c0-4597-9235-147fc4802d19-catalog-content\") pod \"redhat-operators-hqs52\" (UID: \"c3b56f2d-96c0-4597-9235-147fc4802d19\") " pod="openshift-marketplace/redhat-operators-hqs52" Jan 21 11:13:20 crc kubenswrapper[4824]: I0121 11:13:20.406635 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c3b56f2d-96c0-4597-9235-147fc4802d19-utilities\") pod \"redhat-operators-hqs52\" (UID: \"c3b56f2d-96c0-4597-9235-147fc4802d19\") " pod="openshift-marketplace/redhat-operators-hqs52" Jan 21 11:13:20 crc kubenswrapper[4824]: I0121 11:13:20.417855 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8656b\" (UniqueName: \"kubernetes.io/projected/c3b56f2d-96c0-4597-9235-147fc4802d19-kube-api-access-8656b\") pod \"redhat-operators-hqs52\" (UID: \"c3b56f2d-96c0-4597-9235-147fc4802d19\") " pod="openshift-marketplace/redhat-operators-hqs52" Jan 21 11:13:20 crc kubenswrapper[4824]: I0121 11:13:20.420539 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5b5068e5-8cc3-4ed1-811c-835a08eefaa0-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"5b5068e5-8cc3-4ed1-811c-835a08eefaa0\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 21 11:13:20 crc kubenswrapper[4824]: I0121 11:13:20.462160 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 21 11:13:20 crc kubenswrapper[4824]: I0121 11:13:20.463673 4824 patch_prober.go:28] interesting pod/router-default-5444994796-bblkp container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 21 11:13:20 crc kubenswrapper[4824]: [-]has-synced failed: reason withheld Jan 21 11:13:20 crc kubenswrapper[4824]: [+]process-running ok Jan 21 11:13:20 crc kubenswrapper[4824]: healthz check failed Jan 21 11:13:20 crc kubenswrapper[4824]: I0121 11:13:20.463724 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-bblkp" podUID="4b22a7aa-bcc0-4876-a374-a9943a6040bd" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 21 11:13:20 crc kubenswrapper[4824]: I0121 11:13:20.495435 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hqs52" Jan 21 11:13:20 crc kubenswrapper[4824]: I0121 11:13:20.572554 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-2qhk6"] Jan 21 11:13:20 crc kubenswrapper[4824]: I0121 11:13:20.577786 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2qhk6" Jan 21 11:13:20 crc kubenswrapper[4824]: I0121 11:13:20.584608 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-2qhk6"] Jan 21 11:13:20 crc kubenswrapper[4824]: I0121 11:13:20.611595 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/30363b0b-09e8-4fb6-88e4-4e941d833ec8-utilities\") pod \"redhat-operators-2qhk6\" (UID: \"30363b0b-09e8-4fb6-88e4-4e941d833ec8\") " pod="openshift-marketplace/redhat-operators-2qhk6" Jan 21 11:13:20 crc kubenswrapper[4824]: I0121 11:13:20.611678 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xcjz2\" (UniqueName: \"kubernetes.io/projected/30363b0b-09e8-4fb6-88e4-4e941d833ec8-kube-api-access-xcjz2\") pod \"redhat-operators-2qhk6\" (UID: \"30363b0b-09e8-4fb6-88e4-4e941d833ec8\") " pod="openshift-marketplace/redhat-operators-2qhk6" Jan 21 11:13:20 crc kubenswrapper[4824]: I0121 11:13:20.613849 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/30363b0b-09e8-4fb6-88e4-4e941d833ec8-catalog-content\") pod \"redhat-operators-2qhk6\" (UID: \"30363b0b-09e8-4fb6-88e4-4e941d833ec8\") " pod="openshift-marketplace/redhat-operators-2qhk6" Jan 21 11:13:20 crc kubenswrapper[4824]: I0121 11:13:20.671716 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Jan 21 11:13:20 crc kubenswrapper[4824]: I0121 11:13:20.715519 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xcjz2\" (UniqueName: \"kubernetes.io/projected/30363b0b-09e8-4fb6-88e4-4e941d833ec8-kube-api-access-xcjz2\") pod \"redhat-operators-2qhk6\" (UID: \"30363b0b-09e8-4fb6-88e4-4e941d833ec8\") " pod="openshift-marketplace/redhat-operators-2qhk6" Jan 21 11:13:20 crc kubenswrapper[4824]: I0121 11:13:20.715598 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/30363b0b-09e8-4fb6-88e4-4e941d833ec8-catalog-content\") pod \"redhat-operators-2qhk6\" (UID: \"30363b0b-09e8-4fb6-88e4-4e941d833ec8\") " pod="openshift-marketplace/redhat-operators-2qhk6" Jan 21 11:13:20 crc kubenswrapper[4824]: I0121 11:13:20.715661 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/30363b0b-09e8-4fb6-88e4-4e941d833ec8-utilities\") pod \"redhat-operators-2qhk6\" (UID: \"30363b0b-09e8-4fb6-88e4-4e941d833ec8\") " pod="openshift-marketplace/redhat-operators-2qhk6" Jan 21 11:13:20 crc kubenswrapper[4824]: I0121 11:13:20.716499 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/30363b0b-09e8-4fb6-88e4-4e941d833ec8-utilities\") pod \"redhat-operators-2qhk6\" (UID: \"30363b0b-09e8-4fb6-88e4-4e941d833ec8\") " pod="openshift-marketplace/redhat-operators-2qhk6" Jan 21 11:13:20 crc kubenswrapper[4824]: I0121 11:13:20.716616 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/30363b0b-09e8-4fb6-88e4-4e941d833ec8-catalog-content\") pod \"redhat-operators-2qhk6\" (UID: \"30363b0b-09e8-4fb6-88e4-4e941d833ec8\") " pod="openshift-marketplace/redhat-operators-2qhk6" Jan 21 11:13:20 crc kubenswrapper[4824]: I0121 11:13:20.716994 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-hqs52"] Jan 21 11:13:20 crc kubenswrapper[4824]: I0121 11:13:20.731847 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xcjz2\" (UniqueName: \"kubernetes.io/projected/30363b0b-09e8-4fb6-88e4-4e941d833ec8-kube-api-access-xcjz2\") pod \"redhat-operators-2qhk6\" (UID: \"30363b0b-09e8-4fb6-88e4-4e941d833ec8\") " pod="openshift-marketplace/redhat-operators-2qhk6" Jan 21 11:13:20 crc kubenswrapper[4824]: I0121 11:13:20.760842 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29483220-6r4fz" Jan 21 11:13:20 crc kubenswrapper[4824]: I0121 11:13:20.760846 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29483220-6r4fz" event={"ID":"09bd0b9c-46fb-433f-bd30-014c69c90d9f","Type":"ContainerDied","Data":"68b9bb179c8bbe2beb17e913adc7df65d063a4b5a6cdcd18f32da44f14fdb4ae"} Jan 21 11:13:20 crc kubenswrapper[4824]: I0121 11:13:20.760879 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="68b9bb179c8bbe2beb17e913adc7df65d063a4b5a6cdcd18f32da44f14fdb4ae" Jan 21 11:13:20 crc kubenswrapper[4824]: I0121 11:13:20.767805 4824 generic.go:334] "Generic (PLEG): container finished" podID="ff7a0886-d523-486a-9bf0-bae3456b85c2" containerID="a9c629df6b9dc7cf2e177356c8434961e0a88d27574a92b65159b29366efbef3" exitCode=0 Jan 21 11:13:20 crc kubenswrapper[4824]: I0121 11:13:20.767908 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8dj94" event={"ID":"ff7a0886-d523-486a-9bf0-bae3456b85c2","Type":"ContainerDied","Data":"a9c629df6b9dc7cf2e177356c8434961e0a88d27574a92b65159b29366efbef3"} Jan 21 11:13:20 crc kubenswrapper[4824]: I0121 11:13:20.767973 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8dj94" event={"ID":"ff7a0886-d523-486a-9bf0-bae3456b85c2","Type":"ContainerStarted","Data":"1cc0a8fad4ae405d30baab5682d5e3fd0a25f77c99ddd8f27c737f65b8432c15"} Jan 21 11:13:20 crc kubenswrapper[4824]: I0121 11:13:20.775408 4824 generic.go:334] "Generic (PLEG): container finished" podID="1bc8c652-fb4f-48b7-86ce-c45522dfd476" containerID="d5b0c30b7ab307dc38d2c9ff03ad628bf5226a1bb36f64f546ca1dedef2c81fc" exitCode=0 Jan 21 11:13:20 crc kubenswrapper[4824]: I0121 11:13:20.775450 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4w5m6" event={"ID":"1bc8c652-fb4f-48b7-86ce-c45522dfd476","Type":"ContainerDied","Data":"d5b0c30b7ab307dc38d2c9ff03ad628bf5226a1bb36f64f546ca1dedef2c81fc"} Jan 21 11:13:20 crc kubenswrapper[4824]: I0121 11:13:20.775475 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4w5m6" event={"ID":"1bc8c652-fb4f-48b7-86ce-c45522dfd476","Type":"ContainerStarted","Data":"1717bf3078cfbe6fee3800d069a35c471757cb03a1eddb1d8ff1e20167f89c9e"} Jan 21 11:13:20 crc kubenswrapper[4824]: I0121 11:13:20.897484 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2qhk6" Jan 21 11:13:21 crc kubenswrapper[4824]: I0121 11:13:21.463672 4824 patch_prober.go:28] interesting pod/router-default-5444994796-bblkp container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 21 11:13:21 crc kubenswrapper[4824]: [-]has-synced failed: reason withheld Jan 21 11:13:21 crc kubenswrapper[4824]: [+]process-running ok Jan 21 11:13:21 crc kubenswrapper[4824]: healthz check failed Jan 21 11:13:21 crc kubenswrapper[4824]: I0121 11:13:21.463724 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-bblkp" podUID="4b22a7aa-bcc0-4876-a374-a9943a6040bd" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 21 11:13:21 crc kubenswrapper[4824]: I0121 11:13:21.859685 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-htg7q" Jan 21 11:13:21 crc kubenswrapper[4824]: I0121 11:13:21.863461 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-htg7q" Jan 21 11:13:21 crc kubenswrapper[4824]: I0121 11:13:21.963047 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-nwlwd" Jan 21 11:13:21 crc kubenswrapper[4824]: I0121 11:13:21.963093 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-nwlwd" Jan 21 11:13:21 crc kubenswrapper[4824]: I0121 11:13:21.964515 4824 patch_prober.go:28] interesting pod/console-f9d7485db-nwlwd container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.14:8443/health\": dial tcp 10.217.0.14:8443: connect: connection refused" start-of-body= Jan 21 11:13:21 crc kubenswrapper[4824]: I0121 11:13:21.964559 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-nwlwd" podUID="c585ef2f-3e29-4b81-918f-11b075006ca0" containerName="console" probeResult="failure" output="Get \"https://10.217.0.14:8443/health\": dial tcp 10.217.0.14:8443: connect: connection refused" Jan 21 11:13:22 crc kubenswrapper[4824]: I0121 11:13:22.411828 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Jan 21 11:13:22 crc kubenswrapper[4824]: I0121 11:13:22.412743 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 21 11:13:22 crc kubenswrapper[4824]: I0121 11:13:22.414933 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Jan 21 11:13:22 crc kubenswrapper[4824]: I0121 11:13:22.416025 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Jan 21 11:13:22 crc kubenswrapper[4824]: I0121 11:13:22.421228 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Jan 21 11:13:22 crc kubenswrapper[4824]: I0121 11:13:22.460765 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-bblkp" Jan 21 11:13:22 crc kubenswrapper[4824]: I0121 11:13:22.462768 4824 patch_prober.go:28] interesting pod/router-default-5444994796-bblkp container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 21 11:13:22 crc kubenswrapper[4824]: [-]has-synced failed: reason withheld Jan 21 11:13:22 crc kubenswrapper[4824]: [+]process-running ok Jan 21 11:13:22 crc kubenswrapper[4824]: healthz check failed Jan 21 11:13:22 crc kubenswrapper[4824]: I0121 11:13:22.462819 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-bblkp" podUID="4b22a7aa-bcc0-4876-a374-a9943a6040bd" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 21 11:13:22 crc kubenswrapper[4824]: I0121 11:13:22.549902 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/5a6cefc5-7c75-40ca-b14e-d0f7615b7cec-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"5a6cefc5-7c75-40ca-b14e-d0f7615b7cec\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 21 11:13:22 crc kubenswrapper[4824]: I0121 11:13:22.550055 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5a6cefc5-7c75-40ca-b14e-d0f7615b7cec-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"5a6cefc5-7c75-40ca-b14e-d0f7615b7cec\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 21 11:13:22 crc kubenswrapper[4824]: I0121 11:13:22.651386 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/5a6cefc5-7c75-40ca-b14e-d0f7615b7cec-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"5a6cefc5-7c75-40ca-b14e-d0f7615b7cec\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 21 11:13:22 crc kubenswrapper[4824]: I0121 11:13:22.651465 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5a6cefc5-7c75-40ca-b14e-d0f7615b7cec-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"5a6cefc5-7c75-40ca-b14e-d0f7615b7cec\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 21 11:13:22 crc kubenswrapper[4824]: I0121 11:13:22.651495 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/5a6cefc5-7c75-40ca-b14e-d0f7615b7cec-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"5a6cefc5-7c75-40ca-b14e-d0f7615b7cec\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 21 11:13:22 crc kubenswrapper[4824]: I0121 11:13:22.670306 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5a6cefc5-7c75-40ca-b14e-d0f7615b7cec-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"5a6cefc5-7c75-40ca-b14e-d0f7615b7cec\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 21 11:13:22 crc kubenswrapper[4824]: I0121 11:13:22.683671 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-8g4qk" Jan 21 11:13:22 crc kubenswrapper[4824]: I0121 11:13:22.729169 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 21 11:13:22 crc kubenswrapper[4824]: I0121 11:13:22.958219 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 11:13:22 crc kubenswrapper[4824]: I0121 11:13:22.958266 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 11:13:22 crc kubenswrapper[4824]: I0121 11:13:22.959293 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 11:13:22 crc kubenswrapper[4824]: I0121 11:13:22.973123 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 11:13:23 crc kubenswrapper[4824]: I0121 11:13:23.060322 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 11:13:23 crc kubenswrapper[4824]: I0121 11:13:23.060396 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 11:13:23 crc kubenswrapper[4824]: I0121 11:13:23.063033 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 11:13:23 crc kubenswrapper[4824]: I0121 11:13:23.066792 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 11:13:23 crc kubenswrapper[4824]: I0121 11:13:23.263089 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 21 11:13:23 crc kubenswrapper[4824]: I0121 11:13:23.358408 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 21 11:13:23 crc kubenswrapper[4824]: I0121 11:13:23.362237 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 11:13:23 crc kubenswrapper[4824]: I0121 11:13:23.462898 4824 patch_prober.go:28] interesting pod/router-default-5444994796-bblkp container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 21 11:13:23 crc kubenswrapper[4824]: [-]has-synced failed: reason withheld Jan 21 11:13:23 crc kubenswrapper[4824]: [+]process-running ok Jan 21 11:13:23 crc kubenswrapper[4824]: healthz check failed Jan 21 11:13:23 crc kubenswrapper[4824]: I0121 11:13:23.462968 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-bblkp" podUID="4b22a7aa-bcc0-4876-a374-a9943a6040bd" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 21 11:13:23 crc kubenswrapper[4824]: W0121 11:13:23.915586 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc3b56f2d_96c0_4597_9235_147fc4802d19.slice/crio-331b281d41eb6e081cb5f5178cf5cdf5ccdc6fe5086d467942c51189e3b27cd1 WatchSource:0}: Error finding container 331b281d41eb6e081cb5f5178cf5cdf5ccdc6fe5086d467942c51189e3b27cd1: Status 404 returned error can't find the container with id 331b281d41eb6e081cb5f5178cf5cdf5ccdc6fe5086d467942c51189e3b27cd1 Jan 21 11:13:23 crc kubenswrapper[4824]: W0121 11:13:23.917622 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod5b5068e5_8cc3_4ed1_811c_835a08eefaa0.slice/crio-7f857d19b011bee3eeae96a4b8c4c2a37ede3f41ff8641212d7e9d61e18c7440 WatchSource:0}: Error finding container 7f857d19b011bee3eeae96a4b8c4c2a37ede3f41ff8641212d7e9d61e18c7440: Status 404 returned error can't find the container with id 7f857d19b011bee3eeae96a4b8c4c2a37ede3f41ff8641212d7e9d61e18c7440 Jan 21 11:13:24 crc kubenswrapper[4824]: I0121 11:13:24.468271 4824 patch_prober.go:28] interesting pod/router-default-5444994796-bblkp container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 21 11:13:24 crc kubenswrapper[4824]: [-]has-synced failed: reason withheld Jan 21 11:13:24 crc kubenswrapper[4824]: [+]process-running ok Jan 21 11:13:24 crc kubenswrapper[4824]: healthz check failed Jan 21 11:13:24 crc kubenswrapper[4824]: I0121 11:13:24.468321 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-bblkp" podUID="4b22a7aa-bcc0-4876-a374-a9943a6040bd" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 21 11:13:24 crc kubenswrapper[4824]: W0121 11:13:24.777644 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5fe485a1_e14f_4c09_b5b9_f252bc42b7e8.slice/crio-2ab8a4a5ecc0fa65f52a10055dc79fca26c29a2c4809d5a9e1a0b673c37f378d WatchSource:0}: Error finding container 2ab8a4a5ecc0fa65f52a10055dc79fca26c29a2c4809d5a9e1a0b673c37f378d: Status 404 returned error can't find the container with id 2ab8a4a5ecc0fa65f52a10055dc79fca26c29a2c4809d5a9e1a0b673c37f378d Jan 21 11:13:24 crc kubenswrapper[4824]: I0121 11:13:24.800192 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"2ab8a4a5ecc0fa65f52a10055dc79fca26c29a2c4809d5a9e1a0b673c37f378d"} Jan 21 11:13:24 crc kubenswrapper[4824]: I0121 11:13:24.808263 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"5b5068e5-8cc3-4ed1-811c-835a08eefaa0","Type":"ContainerStarted","Data":"7f857d19b011bee3eeae96a4b8c4c2a37ede3f41ff8641212d7e9d61e18c7440"} Jan 21 11:13:24 crc kubenswrapper[4824]: I0121 11:13:24.810907 4824 generic.go:334] "Generic (PLEG): container finished" podID="c3b56f2d-96c0-4597-9235-147fc4802d19" containerID="550d4757a2a4f775989081e05280aba8d60bf329b7f947d0664b2247319cf4a8" exitCode=0 Jan 21 11:13:24 crc kubenswrapper[4824]: I0121 11:13:24.811000 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hqs52" event={"ID":"c3b56f2d-96c0-4597-9235-147fc4802d19","Type":"ContainerDied","Data":"550d4757a2a4f775989081e05280aba8d60bf329b7f947d0664b2247319cf4a8"} Jan 21 11:13:24 crc kubenswrapper[4824]: I0121 11:13:24.811045 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hqs52" event={"ID":"c3b56f2d-96c0-4597-9235-147fc4802d19","Type":"ContainerStarted","Data":"331b281d41eb6e081cb5f5178cf5cdf5ccdc6fe5086d467942c51189e3b27cd1"} Jan 21 11:13:24 crc kubenswrapper[4824]: I0121 11:13:24.818978 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"00d0630de43ad5fd4cd1bf06c27ca7ff6e3a12d85248fbd78a804e3acd633ef7"} Jan 21 11:13:24 crc kubenswrapper[4824]: I0121 11:13:24.820171 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Jan 21 11:13:24 crc kubenswrapper[4824]: I0121 11:13:24.915620 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-2qhk6"] Jan 21 11:13:24 crc kubenswrapper[4824]: W0121 11:13:24.922239 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod30363b0b_09e8_4fb6_88e4_4e941d833ec8.slice/crio-e23807ae2071e0cd34acdbc48558665e201952b8472f9a00085254fc560b0ace WatchSource:0}: Error finding container e23807ae2071e0cd34acdbc48558665e201952b8472f9a00085254fc560b0ace: Status 404 returned error can't find the container with id e23807ae2071e0cd34acdbc48558665e201952b8472f9a00085254fc560b0ace Jan 21 11:13:25 crc kubenswrapper[4824]: I0121 11:13:25.099044 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-dlgmw" Jan 21 11:13:25 crc kubenswrapper[4824]: I0121 11:13:25.463356 4824 patch_prober.go:28] interesting pod/router-default-5444994796-bblkp container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 21 11:13:25 crc kubenswrapper[4824]: [-]has-synced failed: reason withheld Jan 21 11:13:25 crc kubenswrapper[4824]: [+]process-running ok Jan 21 11:13:25 crc kubenswrapper[4824]: healthz check failed Jan 21 11:13:25 crc kubenswrapper[4824]: I0121 11:13:25.463575 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-bblkp" podUID="4b22a7aa-bcc0-4876-a374-a9943a6040bd" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 21 11:13:25 crc kubenswrapper[4824]: I0121 11:13:25.829236 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"a640fdf54e10f37059bf9750e724548215efaf4e93fd9602296529be005be06f"} Jan 21 11:13:25 crc kubenswrapper[4824]: I0121 11:13:25.839051 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"02e04011427f72355c3b733fee5ade53165d5785d1fa946743d9fba748deca4e"} Jan 21 11:13:25 crc kubenswrapper[4824]: I0121 11:13:25.849261 4824 generic.go:334] "Generic (PLEG): container finished" podID="5b5068e5-8cc3-4ed1-811c-835a08eefaa0" containerID="ef6e69b85db227503d6edb2ea4085936ac71a7ca73f8fe10106c4a95a99e875d" exitCode=0 Jan 21 11:13:25 crc kubenswrapper[4824]: I0121 11:13:25.849472 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"5b5068e5-8cc3-4ed1-811c-835a08eefaa0","Type":"ContainerDied","Data":"ef6e69b85db227503d6edb2ea4085936ac71a7ca73f8fe10106c4a95a99e875d"} Jan 21 11:13:25 crc kubenswrapper[4824]: I0121 11:13:25.859645 4824 generic.go:334] "Generic (PLEG): container finished" podID="5a6cefc5-7c75-40ca-b14e-d0f7615b7cec" containerID="7a44ff0047a3d569ff3920eabb699033dff3d6825b4197844557de3ddee34532" exitCode=0 Jan 21 11:13:25 crc kubenswrapper[4824]: I0121 11:13:25.860069 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"5a6cefc5-7c75-40ca-b14e-d0f7615b7cec","Type":"ContainerDied","Data":"7a44ff0047a3d569ff3920eabb699033dff3d6825b4197844557de3ddee34532"} Jan 21 11:13:25 crc kubenswrapper[4824]: I0121 11:13:25.860113 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"5a6cefc5-7c75-40ca-b14e-d0f7615b7cec","Type":"ContainerStarted","Data":"e40b90f30056636be3e795d67f23ef53655a16ac6e0047bd3710d44978191f9c"} Jan 21 11:13:25 crc kubenswrapper[4824]: I0121 11:13:25.867385 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"9e75f6e09ca96c00083080187868f0959b24c9bec3aa61af4b5463f9160f2778"} Jan 21 11:13:25 crc kubenswrapper[4824]: I0121 11:13:25.867411 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"67a4de3580d12468ade2e60b1a645d0031099e3f491957f9a01eb4b06f057b99"} Jan 21 11:13:25 crc kubenswrapper[4824]: I0121 11:13:25.867587 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 11:13:25 crc kubenswrapper[4824]: I0121 11:13:25.868664 4824 generic.go:334] "Generic (PLEG): container finished" podID="30363b0b-09e8-4fb6-88e4-4e941d833ec8" containerID="75d2b0ba70b6aa340dd2f71ac90a2f4234cfa6ec869faff5f999306922a53203" exitCode=0 Jan 21 11:13:25 crc kubenswrapper[4824]: I0121 11:13:25.868692 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2qhk6" event={"ID":"30363b0b-09e8-4fb6-88e4-4e941d833ec8","Type":"ContainerDied","Data":"75d2b0ba70b6aa340dd2f71ac90a2f4234cfa6ec869faff5f999306922a53203"} Jan 21 11:13:25 crc kubenswrapper[4824]: I0121 11:13:25.869215 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2qhk6" event={"ID":"30363b0b-09e8-4fb6-88e4-4e941d833ec8","Type":"ContainerStarted","Data":"e23807ae2071e0cd34acdbc48558665e201952b8472f9a00085254fc560b0ace"} Jan 21 11:13:26 crc kubenswrapper[4824]: I0121 11:13:26.462679 4824 patch_prober.go:28] interesting pod/router-default-5444994796-bblkp container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 21 11:13:26 crc kubenswrapper[4824]: [-]has-synced failed: reason withheld Jan 21 11:13:26 crc kubenswrapper[4824]: [+]process-running ok Jan 21 11:13:26 crc kubenswrapper[4824]: healthz check failed Jan 21 11:13:26 crc kubenswrapper[4824]: I0121 11:13:26.462928 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-bblkp" podUID="4b22a7aa-bcc0-4876-a374-a9943a6040bd" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 21 11:13:27 crc kubenswrapper[4824]: I0121 11:13:27.462914 4824 patch_prober.go:28] interesting pod/router-default-5444994796-bblkp container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 21 11:13:27 crc kubenswrapper[4824]: [-]has-synced failed: reason withheld Jan 21 11:13:27 crc kubenswrapper[4824]: [+]process-running ok Jan 21 11:13:27 crc kubenswrapper[4824]: healthz check failed Jan 21 11:13:27 crc kubenswrapper[4824]: I0121 11:13:27.462990 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-bblkp" podUID="4b22a7aa-bcc0-4876-a374-a9943a6040bd" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 21 11:13:28 crc kubenswrapper[4824]: I0121 11:13:28.462744 4824 patch_prober.go:28] interesting pod/router-default-5444994796-bblkp container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 21 11:13:28 crc kubenswrapper[4824]: [-]has-synced failed: reason withheld Jan 21 11:13:28 crc kubenswrapper[4824]: [+]process-running ok Jan 21 11:13:28 crc kubenswrapper[4824]: healthz check failed Jan 21 11:13:28 crc kubenswrapper[4824]: I0121 11:13:28.462981 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-bblkp" podUID="4b22a7aa-bcc0-4876-a374-a9943a6040bd" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 21 11:13:28 crc kubenswrapper[4824]: I0121 11:13:28.750716 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" Jan 21 11:13:29 crc kubenswrapper[4824]: I0121 11:13:29.462732 4824 patch_prober.go:28] interesting pod/router-default-5444994796-bblkp container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 21 11:13:29 crc kubenswrapper[4824]: [-]has-synced failed: reason withheld Jan 21 11:13:29 crc kubenswrapper[4824]: [+]process-running ok Jan 21 11:13:29 crc kubenswrapper[4824]: healthz check failed Jan 21 11:13:29 crc kubenswrapper[4824]: I0121 11:13:29.462788 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-bblkp" podUID="4b22a7aa-bcc0-4876-a374-a9943a6040bd" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 21 11:13:29 crc kubenswrapper[4824]: I0121 11:13:29.790141 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 21 11:13:29 crc kubenswrapper[4824]: I0121 11:13:29.854802 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/5b5068e5-8cc3-4ed1-811c-835a08eefaa0-kubelet-dir\") pod \"5b5068e5-8cc3-4ed1-811c-835a08eefaa0\" (UID: \"5b5068e5-8cc3-4ed1-811c-835a08eefaa0\") " Jan 21 11:13:29 crc kubenswrapper[4824]: I0121 11:13:29.854849 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5b5068e5-8cc3-4ed1-811c-835a08eefaa0-kube-api-access\") pod \"5b5068e5-8cc3-4ed1-811c-835a08eefaa0\" (UID: \"5b5068e5-8cc3-4ed1-811c-835a08eefaa0\") " Jan 21 11:13:29 crc kubenswrapper[4824]: I0121 11:13:29.854989 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5b5068e5-8cc3-4ed1-811c-835a08eefaa0-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "5b5068e5-8cc3-4ed1-811c-835a08eefaa0" (UID: "5b5068e5-8cc3-4ed1-811c-835a08eefaa0"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 21 11:13:29 crc kubenswrapper[4824]: I0121 11:13:29.855162 4824 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/5b5068e5-8cc3-4ed1-811c-835a08eefaa0-kubelet-dir\") on node \"crc\" DevicePath \"\"" Jan 21 11:13:29 crc kubenswrapper[4824]: I0121 11:13:29.859761 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b5068e5-8cc3-4ed1-811c-835a08eefaa0-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "5b5068e5-8cc3-4ed1-811c-835a08eefaa0" (UID: "5b5068e5-8cc3-4ed1-811c-835a08eefaa0"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:13:29 crc kubenswrapper[4824]: I0121 11:13:29.913159 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"5b5068e5-8cc3-4ed1-811c-835a08eefaa0","Type":"ContainerDied","Data":"7f857d19b011bee3eeae96a4b8c4c2a37ede3f41ff8641212d7e9d61e18c7440"} Jan 21 11:13:29 crc kubenswrapper[4824]: I0121 11:13:29.913193 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7f857d19b011bee3eeae96a4b8c4c2a37ede3f41ff8641212d7e9d61e18c7440" Jan 21 11:13:29 crc kubenswrapper[4824]: I0121 11:13:29.913240 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 21 11:13:29 crc kubenswrapper[4824]: I0121 11:13:29.956708 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5b5068e5-8cc3-4ed1-811c-835a08eefaa0-kube-api-access\") on node \"crc\" DevicePath \"\"" Jan 21 11:13:30 crc kubenswrapper[4824]: I0121 11:13:30.462435 4824 patch_prober.go:28] interesting pod/router-default-5444994796-bblkp container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 21 11:13:30 crc kubenswrapper[4824]: [-]has-synced failed: reason withheld Jan 21 11:13:30 crc kubenswrapper[4824]: [+]process-running ok Jan 21 11:13:30 crc kubenswrapper[4824]: healthz check failed Jan 21 11:13:30 crc kubenswrapper[4824]: I0121 11:13:30.462492 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-bblkp" podUID="4b22a7aa-bcc0-4876-a374-a9943a6040bd" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 21 11:13:31 crc kubenswrapper[4824]: I0121 11:13:31.050031 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 21 11:13:31 crc kubenswrapper[4824]: I0121 11:13:31.068561 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/5a6cefc5-7c75-40ca-b14e-d0f7615b7cec-kubelet-dir\") pod \"5a6cefc5-7c75-40ca-b14e-d0f7615b7cec\" (UID: \"5a6cefc5-7c75-40ca-b14e-d0f7615b7cec\") " Jan 21 11:13:31 crc kubenswrapper[4824]: I0121 11:13:31.068637 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5a6cefc5-7c75-40ca-b14e-d0f7615b7cec-kube-api-access\") pod \"5a6cefc5-7c75-40ca-b14e-d0f7615b7cec\" (UID: \"5a6cefc5-7c75-40ca-b14e-d0f7615b7cec\") " Jan 21 11:13:31 crc kubenswrapper[4824]: I0121 11:13:31.069240 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5a6cefc5-7c75-40ca-b14e-d0f7615b7cec-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "5a6cefc5-7c75-40ca-b14e-d0f7615b7cec" (UID: "5a6cefc5-7c75-40ca-b14e-d0f7615b7cec"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 21 11:13:31 crc kubenswrapper[4824]: I0121 11:13:31.072828 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5a6cefc5-7c75-40ca-b14e-d0f7615b7cec-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "5a6cefc5-7c75-40ca-b14e-d0f7615b7cec" (UID: "5a6cefc5-7c75-40ca-b14e-d0f7615b7cec"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:13:31 crc kubenswrapper[4824]: I0121 11:13:31.170124 4824 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/5a6cefc5-7c75-40ca-b14e-d0f7615b7cec-kubelet-dir\") on node \"crc\" DevicePath \"\"" Jan 21 11:13:31 crc kubenswrapper[4824]: I0121 11:13:31.170646 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5a6cefc5-7c75-40ca-b14e-d0f7615b7cec-kube-api-access\") on node \"crc\" DevicePath \"\"" Jan 21 11:13:31 crc kubenswrapper[4824]: I0121 11:13:31.462833 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-bblkp" Jan 21 11:13:31 crc kubenswrapper[4824]: I0121 11:13:31.465386 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-bblkp" Jan 21 11:13:31 crc kubenswrapper[4824]: I0121 11:13:31.923473 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 21 11:13:31 crc kubenswrapper[4824]: I0121 11:13:31.928869 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"5a6cefc5-7c75-40ca-b14e-d0f7615b7cec","Type":"ContainerDied","Data":"e40b90f30056636be3e795d67f23ef53655a16ac6e0047bd3710d44978191f9c"} Jan 21 11:13:31 crc kubenswrapper[4824]: I0121 11:13:31.928930 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e40b90f30056636be3e795d67f23ef53655a16ac6e0047bd3710d44978191f9c" Jan 21 11:13:31 crc kubenswrapper[4824]: I0121 11:13:31.963934 4824 patch_prober.go:28] interesting pod/console-f9d7485db-nwlwd container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.14:8443/health\": dial tcp 10.217.0.14:8443: connect: connection refused" start-of-body= Jan 21 11:13:31 crc kubenswrapper[4824]: I0121 11:13:31.963986 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-nwlwd" podUID="c585ef2f-3e29-4b81-918f-11b075006ca0" containerName="console" probeResult="failure" output="Get \"https://10.217.0.14:8443/health\": dial tcp 10.217.0.14:8443: connect: connection refused" Jan 21 11:13:35 crc kubenswrapper[4824]: I0121 11:13:35.943010 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8dj94" event={"ID":"ff7a0886-d523-486a-9bf0-bae3456b85c2","Type":"ContainerStarted","Data":"294b6d02b13eca2d11ff429f1d4ad8e41b5d9b3b5d8e4c71c522e3fcaf16fe06"} Jan 21 11:13:35 crc kubenswrapper[4824]: I0121 11:13:35.944982 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-s2hsr" event={"ID":"4205fbde-9c6c-40c8-8c8c-d34a607923e4","Type":"ContainerStarted","Data":"5328aceb9709614055840390079c1f17562775578d1745599737202d75d9d543"} Jan 21 11:13:36 crc kubenswrapper[4824]: I0121 11:13:36.239355 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8f9bd12b-22ee-44ed-a8b9-c100d043f691-metrics-certs\") pod \"network-metrics-daemon-9rmnw\" (UID: \"8f9bd12b-22ee-44ed-a8b9-c100d043f691\") " pod="openshift-multus/network-metrics-daemon-9rmnw" Jan 21 11:13:36 crc kubenswrapper[4824]: I0121 11:13:36.244334 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8f9bd12b-22ee-44ed-a8b9-c100d043f691-metrics-certs\") pod \"network-metrics-daemon-9rmnw\" (UID: \"8f9bd12b-22ee-44ed-a8b9-c100d043f691\") " pod="openshift-multus/network-metrics-daemon-9rmnw" Jan 21 11:13:36 crc kubenswrapper[4824]: I0121 11:13:36.458417 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9rmnw" Jan 21 11:13:36 crc kubenswrapper[4824]: I0121 11:13:36.793574 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-9rmnw"] Jan 21 11:13:36 crc kubenswrapper[4824]: W0121 11:13:36.797271 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8f9bd12b_22ee_44ed_a8b9_c100d043f691.slice/crio-7ca13b06e8160aab5b17fb64bf1b61c7509021a58c16e7661855aeaab33e7d91 WatchSource:0}: Error finding container 7ca13b06e8160aab5b17fb64bf1b61c7509021a58c16e7661855aeaab33e7d91: Status 404 returned error can't find the container with id 7ca13b06e8160aab5b17fb64bf1b61c7509021a58c16e7661855aeaab33e7d91 Jan 21 11:13:36 crc kubenswrapper[4824]: I0121 11:13:36.950641 4824 generic.go:334] "Generic (PLEG): container finished" podID="0939eda4-fff7-447b-ad52-ddb427bf1272" containerID="0cb76d614fd313ac58f734627ceef6ce03d41309478e9627230b2a1b511a93ea" exitCode=0 Jan 21 11:13:36 crc kubenswrapper[4824]: I0121 11:13:36.950736 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pckcl" event={"ID":"0939eda4-fff7-447b-ad52-ddb427bf1272","Type":"ContainerDied","Data":"0cb76d614fd313ac58f734627ceef6ce03d41309478e9627230b2a1b511a93ea"} Jan 21 11:13:36 crc kubenswrapper[4824]: I0121 11:13:36.952462 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-9rmnw" event={"ID":"8f9bd12b-22ee-44ed-a8b9-c100d043f691","Type":"ContainerStarted","Data":"90841fa6243d7a074f0ac613f01a7d1e84fde5829aa9cb007204f2ee1f35fe14"} Jan 21 11:13:36 crc kubenswrapper[4824]: I0121 11:13:36.952498 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-9rmnw" event={"ID":"8f9bd12b-22ee-44ed-a8b9-c100d043f691","Type":"ContainerStarted","Data":"7ca13b06e8160aab5b17fb64bf1b61c7509021a58c16e7661855aeaab33e7d91"} Jan 21 11:13:36 crc kubenswrapper[4824]: I0121 11:13:36.954777 4824 generic.go:334] "Generic (PLEG): container finished" podID="1bc8c652-fb4f-48b7-86ce-c45522dfd476" containerID="122ae0144dee544f8ce79c06f524513fd679f1334f79370a0f9e8b920f8d7ee4" exitCode=0 Jan 21 11:13:36 crc kubenswrapper[4824]: I0121 11:13:36.954842 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4w5m6" event={"ID":"1bc8c652-fb4f-48b7-86ce-c45522dfd476","Type":"ContainerDied","Data":"122ae0144dee544f8ce79c06f524513fd679f1334f79370a0f9e8b920f8d7ee4"} Jan 21 11:13:36 crc kubenswrapper[4824]: I0121 11:13:36.963833 4824 generic.go:334] "Generic (PLEG): container finished" podID="30363b0b-09e8-4fb6-88e4-4e941d833ec8" containerID="e964ec9dc2f3dad14f23b05ef15becda9a72c91eb5ec8bd728789277a97681be" exitCode=0 Jan 21 11:13:36 crc kubenswrapper[4824]: I0121 11:13:36.963917 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2qhk6" event={"ID":"30363b0b-09e8-4fb6-88e4-4e941d833ec8","Type":"ContainerDied","Data":"e964ec9dc2f3dad14f23b05ef15becda9a72c91eb5ec8bd728789277a97681be"} Jan 21 11:13:36 crc kubenswrapper[4824]: I0121 11:13:36.968455 4824 generic.go:334] "Generic (PLEG): container finished" podID="1b040d3d-f4a0-451a-97de-08bea41c6dce" containerID="9f778c1f1c308a32a3e2af072eab8265d05e159ea1a783308d30e563350f290a" exitCode=0 Jan 21 11:13:36 crc kubenswrapper[4824]: I0121 11:13:36.968532 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7q9f2" event={"ID":"1b040d3d-f4a0-451a-97de-08bea41c6dce","Type":"ContainerDied","Data":"9f778c1f1c308a32a3e2af072eab8265d05e159ea1a783308d30e563350f290a"} Jan 21 11:13:36 crc kubenswrapper[4824]: I0121 11:13:36.977001 4824 generic.go:334] "Generic (PLEG): container finished" podID="ff7a0886-d523-486a-9bf0-bae3456b85c2" containerID="294b6d02b13eca2d11ff429f1d4ad8e41b5d9b3b5d8e4c71c522e3fcaf16fe06" exitCode=0 Jan 21 11:13:36 crc kubenswrapper[4824]: I0121 11:13:36.977542 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8dj94" event={"ID":"ff7a0886-d523-486a-9bf0-bae3456b85c2","Type":"ContainerDied","Data":"294b6d02b13eca2d11ff429f1d4ad8e41b5d9b3b5d8e4c71c522e3fcaf16fe06"} Jan 21 11:13:36 crc kubenswrapper[4824]: I0121 11:13:36.982678 4824 generic.go:334] "Generic (PLEG): container finished" podID="3f11911f-e365-4151-b536-3988186b8757" containerID="bc5314fd2d078e4f04072ac4dda373754c451434b3d4dce44df1b40cbeecd1ba" exitCode=0 Jan 21 11:13:36 crc kubenswrapper[4824]: I0121 11:13:36.982733 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-b922x" event={"ID":"3f11911f-e365-4151-b536-3988186b8757","Type":"ContainerDied","Data":"bc5314fd2d078e4f04072ac4dda373754c451434b3d4dce44df1b40cbeecd1ba"} Jan 21 11:13:36 crc kubenswrapper[4824]: I0121 11:13:36.985525 4824 generic.go:334] "Generic (PLEG): container finished" podID="4205fbde-9c6c-40c8-8c8c-d34a607923e4" containerID="5328aceb9709614055840390079c1f17562775578d1745599737202d75d9d543" exitCode=0 Jan 21 11:13:36 crc kubenswrapper[4824]: I0121 11:13:36.985572 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-s2hsr" event={"ID":"4205fbde-9c6c-40c8-8c8c-d34a607923e4","Type":"ContainerDied","Data":"5328aceb9709614055840390079c1f17562775578d1745599737202d75d9d543"} Jan 21 11:13:36 crc kubenswrapper[4824]: I0121 11:13:36.988417 4824 generic.go:334] "Generic (PLEG): container finished" podID="c3b56f2d-96c0-4597-9235-147fc4802d19" containerID="2f3db218d18e31256264bc1967baec36484f60d00de32760d6c6503fe4e7b9fc" exitCode=0 Jan 21 11:13:36 crc kubenswrapper[4824]: I0121 11:13:36.988449 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hqs52" event={"ID":"c3b56f2d-96c0-4597-9235-147fc4802d19","Type":"ContainerDied","Data":"2f3db218d18e31256264bc1967baec36484f60d00de32760d6c6503fe4e7b9fc"} Jan 21 11:13:37 crc kubenswrapper[4824]: I0121 11:13:37.996799 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4w5m6" event={"ID":"1bc8c652-fb4f-48b7-86ce-c45522dfd476","Type":"ContainerStarted","Data":"fa2ad6f2a50b9350a988947c908be8adb082c0d1e40507b63b02b300f1699aae"} Jan 21 11:13:38 crc kubenswrapper[4824]: I0121 11:13:38.000005 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2qhk6" event={"ID":"30363b0b-09e8-4fb6-88e4-4e941d833ec8","Type":"ContainerStarted","Data":"812b638a67ec88d3b10766fb23536e55485cc458e7ca4b815bcac4f9331b7868"} Jan 21 11:13:38 crc kubenswrapper[4824]: I0121 11:13:38.007879 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-b922x" event={"ID":"3f11911f-e365-4151-b536-3988186b8757","Type":"ContainerStarted","Data":"f4386ff52380b3c795405ecbb9387d4d9b5fca2cffa267eb0ddb32961b0a7f10"} Jan 21 11:13:38 crc kubenswrapper[4824]: I0121 11:13:38.009917 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7q9f2" event={"ID":"1b040d3d-f4a0-451a-97de-08bea41c6dce","Type":"ContainerStarted","Data":"728ea008f822360aaa1493d68beaf88c96b2cdf76a7eeb252c13acb6bf91c9d5"} Jan 21 11:13:38 crc kubenswrapper[4824]: I0121 11:13:38.018181 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-s2hsr" event={"ID":"4205fbde-9c6c-40c8-8c8c-d34a607923e4","Type":"ContainerStarted","Data":"ca78012db1bd2e41838feee41b14658b90ea2761895e81ba4e7d350e19aec395"} Jan 21 11:13:38 crc kubenswrapper[4824]: I0121 11:13:38.020804 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pckcl" event={"ID":"0939eda4-fff7-447b-ad52-ddb427bf1272","Type":"ContainerStarted","Data":"89b6b86330b8be957af5668a743a12da8e677d657f18983640cca9f171c269d7"} Jan 21 11:13:38 crc kubenswrapper[4824]: I0121 11:13:38.022451 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-9rmnw" event={"ID":"8f9bd12b-22ee-44ed-a8b9-c100d043f691","Type":"ContainerStarted","Data":"8232a034b266977b397d20f7cdf89da9cd11f33af94ea1374cea91d5d9736037"} Jan 21 11:13:38 crc kubenswrapper[4824]: I0121 11:13:38.024407 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hqs52" event={"ID":"c3b56f2d-96c0-4597-9235-147fc4802d19","Type":"ContainerStarted","Data":"b1068da5b4766f19c34535e8e58a22325fb8a2d31ce06ee547c973539c887a47"} Jan 21 11:13:38 crc kubenswrapper[4824]: I0121 11:13:38.033531 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-b922x" podStartSLOduration=2.191925607 podStartE2EDuration="21.033521717s" podCreationTimestamp="2026-01-21 11:13:17 +0000 UTC" firstStartedPulling="2026-01-21 11:13:18.73587321 +0000 UTC m=+141.028902502" lastFinishedPulling="2026-01-21 11:13:37.57746932 +0000 UTC m=+159.870498612" observedRunningTime="2026-01-21 11:13:38.031985582 +0000 UTC m=+160.325014874" watchObservedRunningTime="2026-01-21 11:13:38.033521717 +0000 UTC m=+160.326551009" Jan 21 11:13:38 crc kubenswrapper[4824]: I0121 11:13:38.034130 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-4w5m6" podStartSLOduration=5.39873712 podStartE2EDuration="19.03412513s" podCreationTimestamp="2026-01-21 11:13:19 +0000 UTC" firstStartedPulling="2026-01-21 11:13:23.904887702 +0000 UTC m=+146.197916994" lastFinishedPulling="2026-01-21 11:13:37.540275712 +0000 UTC m=+159.833305004" observedRunningTime="2026-01-21 11:13:38.016888951 +0000 UTC m=+160.309918243" watchObservedRunningTime="2026-01-21 11:13:38.03412513 +0000 UTC m=+160.327154422" Jan 21 11:13:38 crc kubenswrapper[4824]: I0121 11:13:38.048906 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-2qhk6" podStartSLOduration=6.245616317 podStartE2EDuration="18.048892321s" podCreationTimestamp="2026-01-21 11:13:20 +0000 UTC" firstStartedPulling="2026-01-21 11:13:25.870187028 +0000 UTC m=+148.163216320" lastFinishedPulling="2026-01-21 11:13:37.673463032 +0000 UTC m=+159.966492324" observedRunningTime="2026-01-21 11:13:38.046673826 +0000 UTC m=+160.339703118" watchObservedRunningTime="2026-01-21 11:13:38.048892321 +0000 UTC m=+160.341921613" Jan 21 11:13:38 crc kubenswrapper[4824]: I0121 11:13:38.064211 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-7q9f2" podStartSLOduration=2.232781756 podStartE2EDuration="22.064192926s" podCreationTimestamp="2026-01-21 11:13:16 +0000 UTC" firstStartedPulling="2026-01-21 11:13:17.702937212 +0000 UTC m=+139.995966504" lastFinishedPulling="2026-01-21 11:13:37.534348381 +0000 UTC m=+159.827377674" observedRunningTime="2026-01-21 11:13:38.062946625 +0000 UTC m=+160.355975918" watchObservedRunningTime="2026-01-21 11:13:38.064192926 +0000 UTC m=+160.357222218" Jan 21 11:13:38 crc kubenswrapper[4824]: I0121 11:13:38.102187 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-pckcl" podStartSLOduration=2.284079136 podStartE2EDuration="21.10217267s" podCreationTimestamp="2026-01-21 11:13:17 +0000 UTC" firstStartedPulling="2026-01-21 11:13:18.710219794 +0000 UTC m=+141.003249086" lastFinishedPulling="2026-01-21 11:13:37.528313328 +0000 UTC m=+159.821342620" observedRunningTime="2026-01-21 11:13:38.101510446 +0000 UTC m=+160.394539738" watchObservedRunningTime="2026-01-21 11:13:38.10217267 +0000 UTC m=+160.395201962" Jan 21 11:13:38 crc kubenswrapper[4824]: I0121 11:13:38.102564 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-9rmnw" podStartSLOduration=144.102559767 podStartE2EDuration="2m24.102559767s" podCreationTimestamp="2026-01-21 11:11:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:13:38.085539363 +0000 UTC m=+160.378568656" watchObservedRunningTime="2026-01-21 11:13:38.102559767 +0000 UTC m=+160.395589059" Jan 21 11:13:38 crc kubenswrapper[4824]: I0121 11:13:38.118160 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-hqs52" podStartSLOduration=5.235102315 podStartE2EDuration="18.118145125s" podCreationTimestamp="2026-01-21 11:13:20 +0000 UTC" firstStartedPulling="2026-01-21 11:13:24.819519941 +0000 UTC m=+147.112549233" lastFinishedPulling="2026-01-21 11:13:37.702562751 +0000 UTC m=+159.995592043" observedRunningTime="2026-01-21 11:13:38.114780748 +0000 UTC m=+160.407810039" watchObservedRunningTime="2026-01-21 11:13:38.118145125 +0000 UTC m=+160.411174418" Jan 21 11:13:38 crc kubenswrapper[4824]: I0121 11:13:38.137792 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-s2hsr" podStartSLOduration=2.254256471 podStartE2EDuration="21.137741697s" podCreationTimestamp="2026-01-21 11:13:17 +0000 UTC" firstStartedPulling="2026-01-21 11:13:18.72741732 +0000 UTC m=+141.020446612" lastFinishedPulling="2026-01-21 11:13:37.610902546 +0000 UTC m=+159.903931838" observedRunningTime="2026-01-21 11:13:38.13755691 +0000 UTC m=+160.430586202" watchObservedRunningTime="2026-01-21 11:13:38.137741697 +0000 UTC m=+160.430770989" Jan 21 11:13:38 crc kubenswrapper[4824]: I0121 11:13:38.664346 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-wxlxx" Jan 21 11:13:39 crc kubenswrapper[4824]: I0121 11:13:39.046702 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8dj94" event={"ID":"ff7a0886-d523-486a-9bf0-bae3456b85c2","Type":"ContainerStarted","Data":"963e8246205e0fbf3da3eb5eadb52a32e5194169b632c986a341b4e6c7b913dd"} Jan 21 11:13:39 crc kubenswrapper[4824]: I0121 11:13:39.062824 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-8dj94" podStartSLOduration=6.064740655 podStartE2EDuration="20.062811307s" podCreationTimestamp="2026-01-21 11:13:19 +0000 UTC" firstStartedPulling="2026-01-21 11:13:23.90538185 +0000 UTC m=+146.198411142" lastFinishedPulling="2026-01-21 11:13:37.903452502 +0000 UTC m=+160.196481794" observedRunningTime="2026-01-21 11:13:39.061787946 +0000 UTC m=+161.354817238" watchObservedRunningTime="2026-01-21 11:13:39.062811307 +0000 UTC m=+161.355840599" Jan 21 11:13:39 crc kubenswrapper[4824]: I0121 11:13:39.789575 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-4w5m6" Jan 21 11:13:39 crc kubenswrapper[4824]: I0121 11:13:39.789622 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-4w5m6" Jan 21 11:13:39 crc kubenswrapper[4824]: I0121 11:13:39.859098 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-4w5m6" Jan 21 11:13:39 crc kubenswrapper[4824]: I0121 11:13:39.895181 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-8dj94" Jan 21 11:13:39 crc kubenswrapper[4824]: I0121 11:13:39.895225 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-8dj94" Jan 21 11:13:39 crc kubenswrapper[4824]: I0121 11:13:39.922491 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-8dj94" Jan 21 11:13:40 crc kubenswrapper[4824]: I0121 11:13:40.495766 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-hqs52" Jan 21 11:13:40 crc kubenswrapper[4824]: I0121 11:13:40.496233 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-hqs52" Jan 21 11:13:40 crc kubenswrapper[4824]: I0121 11:13:40.898144 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-2qhk6" Jan 21 11:13:40 crc kubenswrapper[4824]: I0121 11:13:40.898188 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-2qhk6" Jan 21 11:13:41 crc kubenswrapper[4824]: I0121 11:13:41.522458 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-hqs52" podUID="c3b56f2d-96c0-4597-9235-147fc4802d19" containerName="registry-server" probeResult="failure" output=< Jan 21 11:13:41 crc kubenswrapper[4824]: timeout: failed to connect service ":50051" within 1s Jan 21 11:13:41 crc kubenswrapper[4824]: > Jan 21 11:13:41 crc kubenswrapper[4824]: I0121 11:13:41.923566 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-2qhk6" podUID="30363b0b-09e8-4fb6-88e4-4e941d833ec8" containerName="registry-server" probeResult="failure" output=< Jan 21 11:13:41 crc kubenswrapper[4824]: timeout: failed to connect service ":50051" within 1s Jan 21 11:13:41 crc kubenswrapper[4824]: > Jan 21 11:13:41 crc kubenswrapper[4824]: I0121 11:13:41.968742 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-nwlwd" Jan 21 11:13:41 crc kubenswrapper[4824]: I0121 11:13:41.972070 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-nwlwd" Jan 21 11:13:46 crc kubenswrapper[4824]: I0121 11:13:46.065537 4824 patch_prober.go:28] interesting pod/machine-config-daemon-zqjg9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 21 11:13:46 crc kubenswrapper[4824]: I0121 11:13:46.065794 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 21 11:13:47 crc kubenswrapper[4824]: I0121 11:13:47.321543 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-7q9f2" Jan 21 11:13:47 crc kubenswrapper[4824]: I0121 11:13:47.322082 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-7q9f2" Jan 21 11:13:47 crc kubenswrapper[4824]: I0121 11:13:47.352502 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-7q9f2" Jan 21 11:13:47 crc kubenswrapper[4824]: I0121 11:13:47.525514 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-b922x" Jan 21 11:13:47 crc kubenswrapper[4824]: I0121 11:13:47.525745 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-b922x" Jan 21 11:13:47 crc kubenswrapper[4824]: I0121 11:13:47.553045 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-b922x" Jan 21 11:13:47 crc kubenswrapper[4824]: I0121 11:13:47.693565 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-pckcl" Jan 21 11:13:47 crc kubenswrapper[4824]: I0121 11:13:47.693792 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-pckcl" Jan 21 11:13:47 crc kubenswrapper[4824]: I0121 11:13:47.720050 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-pckcl" Jan 21 11:13:47 crc kubenswrapper[4824]: I0121 11:13:47.904235 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-s2hsr" Jan 21 11:13:47 crc kubenswrapper[4824]: I0121 11:13:47.904427 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-s2hsr" Jan 21 11:13:47 crc kubenswrapper[4824]: I0121 11:13:47.931122 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-s2hsr" Jan 21 11:13:48 crc kubenswrapper[4824]: I0121 11:13:48.107713 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-7q9f2" Jan 21 11:13:48 crc kubenswrapper[4824]: I0121 11:13:48.108143 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-s2hsr" Jan 21 11:13:48 crc kubenswrapper[4824]: I0121 11:13:48.109053 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-pckcl" Jan 21 11:13:48 crc kubenswrapper[4824]: I0121 11:13:48.109185 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-b922x" Jan 21 11:13:49 crc kubenswrapper[4824]: I0121 11:13:49.375113 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-s2hsr"] Jan 21 11:13:49 crc kubenswrapper[4824]: I0121 11:13:49.816095 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-4w5m6" Jan 21 11:13:49 crc kubenswrapper[4824]: I0121 11:13:49.921485 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-8dj94" Jan 21 11:13:49 crc kubenswrapper[4824]: I0121 11:13:49.978224 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-pckcl"] Jan 21 11:13:50 crc kubenswrapper[4824]: I0121 11:13:50.092539 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-s2hsr" podUID="4205fbde-9c6c-40c8-8c8c-d34a607923e4" containerName="registry-server" containerID="cri-o://ca78012db1bd2e41838feee41b14658b90ea2761895e81ba4e7d350e19aec395" gracePeriod=2 Jan 21 11:13:50 crc kubenswrapper[4824]: I0121 11:13:50.450833 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-s2hsr" Jan 21 11:13:50 crc kubenswrapper[4824]: I0121 11:13:50.523193 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-hqs52" Jan 21 11:13:50 crc kubenswrapper[4824]: I0121 11:13:50.528033 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4205fbde-9c6c-40c8-8c8c-d34a607923e4-catalog-content\") pod \"4205fbde-9c6c-40c8-8c8c-d34a607923e4\" (UID: \"4205fbde-9c6c-40c8-8c8c-d34a607923e4\") " Jan 21 11:13:50 crc kubenswrapper[4824]: I0121 11:13:50.528073 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4205fbde-9c6c-40c8-8c8c-d34a607923e4-utilities\") pod \"4205fbde-9c6c-40c8-8c8c-d34a607923e4\" (UID: \"4205fbde-9c6c-40c8-8c8c-d34a607923e4\") " Jan 21 11:13:50 crc kubenswrapper[4824]: I0121 11:13:50.528093 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mtft2\" (UniqueName: \"kubernetes.io/projected/4205fbde-9c6c-40c8-8c8c-d34a607923e4-kube-api-access-mtft2\") pod \"4205fbde-9c6c-40c8-8c8c-d34a607923e4\" (UID: \"4205fbde-9c6c-40c8-8c8c-d34a607923e4\") " Jan 21 11:13:50 crc kubenswrapper[4824]: I0121 11:13:50.529357 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4205fbde-9c6c-40c8-8c8c-d34a607923e4-utilities" (OuterVolumeSpecName: "utilities") pod "4205fbde-9c6c-40c8-8c8c-d34a607923e4" (UID: "4205fbde-9c6c-40c8-8c8c-d34a607923e4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:13:50 crc kubenswrapper[4824]: I0121 11:13:50.532787 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4205fbde-9c6c-40c8-8c8c-d34a607923e4-kube-api-access-mtft2" (OuterVolumeSpecName: "kube-api-access-mtft2") pod "4205fbde-9c6c-40c8-8c8c-d34a607923e4" (UID: "4205fbde-9c6c-40c8-8c8c-d34a607923e4"). InnerVolumeSpecName "kube-api-access-mtft2". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:13:50 crc kubenswrapper[4824]: I0121 11:13:50.550779 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-hqs52" Jan 21 11:13:50 crc kubenswrapper[4824]: I0121 11:13:50.575149 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4205fbde-9c6c-40c8-8c8c-d34a607923e4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4205fbde-9c6c-40c8-8c8c-d34a607923e4" (UID: "4205fbde-9c6c-40c8-8c8c-d34a607923e4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:13:50 crc kubenswrapper[4824]: I0121 11:13:50.629231 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4205fbde-9c6c-40c8-8c8c-d34a607923e4-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 21 11:13:50 crc kubenswrapper[4824]: I0121 11:13:50.629262 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4205fbde-9c6c-40c8-8c8c-d34a607923e4-utilities\") on node \"crc\" DevicePath \"\"" Jan 21 11:13:50 crc kubenswrapper[4824]: I0121 11:13:50.629274 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mtft2\" (UniqueName: \"kubernetes.io/projected/4205fbde-9c6c-40c8-8c8c-d34a607923e4-kube-api-access-mtft2\") on node \"crc\" DevicePath \"\"" Jan 21 11:13:50 crc kubenswrapper[4824]: I0121 11:13:50.924674 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-2qhk6" Jan 21 11:13:50 crc kubenswrapper[4824]: I0121 11:13:50.954488 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-2qhk6" Jan 21 11:13:51 crc kubenswrapper[4824]: I0121 11:13:51.098037 4824 generic.go:334] "Generic (PLEG): container finished" podID="4205fbde-9c6c-40c8-8c8c-d34a607923e4" containerID="ca78012db1bd2e41838feee41b14658b90ea2761895e81ba4e7d350e19aec395" exitCode=0 Jan 21 11:13:51 crc kubenswrapper[4824]: I0121 11:13:51.098618 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-s2hsr" Jan 21 11:13:51 crc kubenswrapper[4824]: I0121 11:13:51.105876 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-s2hsr" event={"ID":"4205fbde-9c6c-40c8-8c8c-d34a607923e4","Type":"ContainerDied","Data":"ca78012db1bd2e41838feee41b14658b90ea2761895e81ba4e7d350e19aec395"} Jan 21 11:13:51 crc kubenswrapper[4824]: I0121 11:13:51.105921 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-s2hsr" event={"ID":"4205fbde-9c6c-40c8-8c8c-d34a607923e4","Type":"ContainerDied","Data":"f85af8c50c5e5798e08394dc9af278d2e32f05c7d0d8ddcab84b8549c975347a"} Jan 21 11:13:51 crc kubenswrapper[4824]: I0121 11:13:51.105986 4824 scope.go:117] "RemoveContainer" containerID="ca78012db1bd2e41838feee41b14658b90ea2761895e81ba4e7d350e19aec395" Jan 21 11:13:51 crc kubenswrapper[4824]: I0121 11:13:51.106091 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-pckcl" podUID="0939eda4-fff7-447b-ad52-ddb427bf1272" containerName="registry-server" containerID="cri-o://89b6b86330b8be957af5668a743a12da8e677d657f18983640cca9f171c269d7" gracePeriod=2 Jan 21 11:13:51 crc kubenswrapper[4824]: I0121 11:13:51.119214 4824 scope.go:117] "RemoveContainer" containerID="5328aceb9709614055840390079c1f17562775578d1745599737202d75d9d543" Jan 21 11:13:51 crc kubenswrapper[4824]: I0121 11:13:51.125805 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-s2hsr"] Jan 21 11:13:51 crc kubenswrapper[4824]: I0121 11:13:51.127800 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-s2hsr"] Jan 21 11:13:51 crc kubenswrapper[4824]: I0121 11:13:51.159842 4824 scope.go:117] "RemoveContainer" containerID="12e7f9c58f1247fa3830ac89333e154e7e3fd0757d08bc1ecbe6ca9ac5f27bd0" Jan 21 11:13:51 crc kubenswrapper[4824]: I0121 11:13:51.208760 4824 scope.go:117] "RemoveContainer" containerID="ca78012db1bd2e41838feee41b14658b90ea2761895e81ba4e7d350e19aec395" Jan 21 11:13:51 crc kubenswrapper[4824]: E0121 11:13:51.209139 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ca78012db1bd2e41838feee41b14658b90ea2761895e81ba4e7d350e19aec395\": container with ID starting with ca78012db1bd2e41838feee41b14658b90ea2761895e81ba4e7d350e19aec395 not found: ID does not exist" containerID="ca78012db1bd2e41838feee41b14658b90ea2761895e81ba4e7d350e19aec395" Jan 21 11:13:51 crc kubenswrapper[4824]: I0121 11:13:51.209186 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ca78012db1bd2e41838feee41b14658b90ea2761895e81ba4e7d350e19aec395"} err="failed to get container status \"ca78012db1bd2e41838feee41b14658b90ea2761895e81ba4e7d350e19aec395\": rpc error: code = NotFound desc = could not find container \"ca78012db1bd2e41838feee41b14658b90ea2761895e81ba4e7d350e19aec395\": container with ID starting with ca78012db1bd2e41838feee41b14658b90ea2761895e81ba4e7d350e19aec395 not found: ID does not exist" Jan 21 11:13:51 crc kubenswrapper[4824]: I0121 11:13:51.209229 4824 scope.go:117] "RemoveContainer" containerID="5328aceb9709614055840390079c1f17562775578d1745599737202d75d9d543" Jan 21 11:13:51 crc kubenswrapper[4824]: E0121 11:13:51.209534 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5328aceb9709614055840390079c1f17562775578d1745599737202d75d9d543\": container with ID starting with 5328aceb9709614055840390079c1f17562775578d1745599737202d75d9d543 not found: ID does not exist" containerID="5328aceb9709614055840390079c1f17562775578d1745599737202d75d9d543" Jan 21 11:13:51 crc kubenswrapper[4824]: I0121 11:13:51.209560 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5328aceb9709614055840390079c1f17562775578d1745599737202d75d9d543"} err="failed to get container status \"5328aceb9709614055840390079c1f17562775578d1745599737202d75d9d543\": rpc error: code = NotFound desc = could not find container \"5328aceb9709614055840390079c1f17562775578d1745599737202d75d9d543\": container with ID starting with 5328aceb9709614055840390079c1f17562775578d1745599737202d75d9d543 not found: ID does not exist" Jan 21 11:13:51 crc kubenswrapper[4824]: I0121 11:13:51.209587 4824 scope.go:117] "RemoveContainer" containerID="12e7f9c58f1247fa3830ac89333e154e7e3fd0757d08bc1ecbe6ca9ac5f27bd0" Jan 21 11:13:51 crc kubenswrapper[4824]: E0121 11:13:51.209794 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"12e7f9c58f1247fa3830ac89333e154e7e3fd0757d08bc1ecbe6ca9ac5f27bd0\": container with ID starting with 12e7f9c58f1247fa3830ac89333e154e7e3fd0757d08bc1ecbe6ca9ac5f27bd0 not found: ID does not exist" containerID="12e7f9c58f1247fa3830ac89333e154e7e3fd0757d08bc1ecbe6ca9ac5f27bd0" Jan 21 11:13:51 crc kubenswrapper[4824]: I0121 11:13:51.209817 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"12e7f9c58f1247fa3830ac89333e154e7e3fd0757d08bc1ecbe6ca9ac5f27bd0"} err="failed to get container status \"12e7f9c58f1247fa3830ac89333e154e7e3fd0757d08bc1ecbe6ca9ac5f27bd0\": rpc error: code = NotFound desc = could not find container \"12e7f9c58f1247fa3830ac89333e154e7e3fd0757d08bc1ecbe6ca9ac5f27bd0\": container with ID starting with 12e7f9c58f1247fa3830ac89333e154e7e3fd0757d08bc1ecbe6ca9ac5f27bd0 not found: ID does not exist" Jan 21 11:13:51 crc kubenswrapper[4824]: I0121 11:13:51.457751 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pckcl" Jan 21 11:13:51 crc kubenswrapper[4824]: I0121 11:13:51.543162 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0939eda4-fff7-447b-ad52-ddb427bf1272-utilities\") pod \"0939eda4-fff7-447b-ad52-ddb427bf1272\" (UID: \"0939eda4-fff7-447b-ad52-ddb427bf1272\") " Jan 21 11:13:51 crc kubenswrapper[4824]: I0121 11:13:51.543404 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gd5wx\" (UniqueName: \"kubernetes.io/projected/0939eda4-fff7-447b-ad52-ddb427bf1272-kube-api-access-gd5wx\") pod \"0939eda4-fff7-447b-ad52-ddb427bf1272\" (UID: \"0939eda4-fff7-447b-ad52-ddb427bf1272\") " Jan 21 11:13:51 crc kubenswrapper[4824]: I0121 11:13:51.543552 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0939eda4-fff7-447b-ad52-ddb427bf1272-catalog-content\") pod \"0939eda4-fff7-447b-ad52-ddb427bf1272\" (UID: \"0939eda4-fff7-447b-ad52-ddb427bf1272\") " Jan 21 11:13:51 crc kubenswrapper[4824]: I0121 11:13:51.543839 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0939eda4-fff7-447b-ad52-ddb427bf1272-utilities" (OuterVolumeSpecName: "utilities") pod "0939eda4-fff7-447b-ad52-ddb427bf1272" (UID: "0939eda4-fff7-447b-ad52-ddb427bf1272"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:13:51 crc kubenswrapper[4824]: I0121 11:13:51.546246 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0939eda4-fff7-447b-ad52-ddb427bf1272-kube-api-access-gd5wx" (OuterVolumeSpecName: "kube-api-access-gd5wx") pod "0939eda4-fff7-447b-ad52-ddb427bf1272" (UID: "0939eda4-fff7-447b-ad52-ddb427bf1272"). InnerVolumeSpecName "kube-api-access-gd5wx". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:13:51 crc kubenswrapper[4824]: I0121 11:13:51.580998 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0939eda4-fff7-447b-ad52-ddb427bf1272-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0939eda4-fff7-447b-ad52-ddb427bf1272" (UID: "0939eda4-fff7-447b-ad52-ddb427bf1272"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:13:51 crc kubenswrapper[4824]: I0121 11:13:51.649312 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0939eda4-fff7-447b-ad52-ddb427bf1272-utilities\") on node \"crc\" DevicePath \"\"" Jan 21 11:13:51 crc kubenswrapper[4824]: I0121 11:13:51.649344 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gd5wx\" (UniqueName: \"kubernetes.io/projected/0939eda4-fff7-447b-ad52-ddb427bf1272-kube-api-access-gd5wx\") on node \"crc\" DevicePath \"\"" Jan 21 11:13:51 crc kubenswrapper[4824]: I0121 11:13:51.649355 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0939eda4-fff7-447b-ad52-ddb427bf1272-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 21 11:13:52 crc kubenswrapper[4824]: I0121 11:13:52.055095 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4205fbde-9c6c-40c8-8c8c-d34a607923e4" path="/var/lib/kubelet/pods/4205fbde-9c6c-40c8-8c8c-d34a607923e4/volumes" Jan 21 11:13:52 crc kubenswrapper[4824]: I0121 11:13:52.103652 4824 generic.go:334] "Generic (PLEG): container finished" podID="0939eda4-fff7-447b-ad52-ddb427bf1272" containerID="89b6b86330b8be957af5668a743a12da8e677d657f18983640cca9f171c269d7" exitCode=0 Jan 21 11:13:52 crc kubenswrapper[4824]: I0121 11:13:52.103685 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pckcl" event={"ID":"0939eda4-fff7-447b-ad52-ddb427bf1272","Type":"ContainerDied","Data":"89b6b86330b8be957af5668a743a12da8e677d657f18983640cca9f171c269d7"} Jan 21 11:13:52 crc kubenswrapper[4824]: I0121 11:13:52.103708 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pckcl" event={"ID":"0939eda4-fff7-447b-ad52-ddb427bf1272","Type":"ContainerDied","Data":"0fe9a2ce9dc5eb2c88168927f6206b6e2577521e0d1f3b71800ade62c141b6ff"} Jan 21 11:13:52 crc kubenswrapper[4824]: I0121 11:13:52.103716 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pckcl" Jan 21 11:13:52 crc kubenswrapper[4824]: I0121 11:13:52.103724 4824 scope.go:117] "RemoveContainer" containerID="89b6b86330b8be957af5668a743a12da8e677d657f18983640cca9f171c269d7" Jan 21 11:13:52 crc kubenswrapper[4824]: I0121 11:13:52.117384 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-pckcl"] Jan 21 11:13:52 crc kubenswrapper[4824]: I0121 11:13:52.117412 4824 scope.go:117] "RemoveContainer" containerID="0cb76d614fd313ac58f734627ceef6ce03d41309478e9627230b2a1b511a93ea" Jan 21 11:13:52 crc kubenswrapper[4824]: I0121 11:13:52.118198 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-pckcl"] Jan 21 11:13:52 crc kubenswrapper[4824]: I0121 11:13:52.128032 4824 scope.go:117] "RemoveContainer" containerID="f2ddd6ed1029c405c4138f6257aaa84375743b3602a5b630fdf5b86e73651795" Jan 21 11:13:52 crc kubenswrapper[4824]: I0121 11:13:52.136661 4824 scope.go:117] "RemoveContainer" containerID="89b6b86330b8be957af5668a743a12da8e677d657f18983640cca9f171c269d7" Jan 21 11:13:52 crc kubenswrapper[4824]: E0121 11:13:52.136981 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"89b6b86330b8be957af5668a743a12da8e677d657f18983640cca9f171c269d7\": container with ID starting with 89b6b86330b8be957af5668a743a12da8e677d657f18983640cca9f171c269d7 not found: ID does not exist" containerID="89b6b86330b8be957af5668a743a12da8e677d657f18983640cca9f171c269d7" Jan 21 11:13:52 crc kubenswrapper[4824]: I0121 11:13:52.137072 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"89b6b86330b8be957af5668a743a12da8e677d657f18983640cca9f171c269d7"} err="failed to get container status \"89b6b86330b8be957af5668a743a12da8e677d657f18983640cca9f171c269d7\": rpc error: code = NotFound desc = could not find container \"89b6b86330b8be957af5668a743a12da8e677d657f18983640cca9f171c269d7\": container with ID starting with 89b6b86330b8be957af5668a743a12da8e677d657f18983640cca9f171c269d7 not found: ID does not exist" Jan 21 11:13:52 crc kubenswrapper[4824]: I0121 11:13:52.137142 4824 scope.go:117] "RemoveContainer" containerID="0cb76d614fd313ac58f734627ceef6ce03d41309478e9627230b2a1b511a93ea" Jan 21 11:13:52 crc kubenswrapper[4824]: E0121 11:13:52.137422 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0cb76d614fd313ac58f734627ceef6ce03d41309478e9627230b2a1b511a93ea\": container with ID starting with 0cb76d614fd313ac58f734627ceef6ce03d41309478e9627230b2a1b511a93ea not found: ID does not exist" containerID="0cb76d614fd313ac58f734627ceef6ce03d41309478e9627230b2a1b511a93ea" Jan 21 11:13:52 crc kubenswrapper[4824]: I0121 11:13:52.137501 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0cb76d614fd313ac58f734627ceef6ce03d41309478e9627230b2a1b511a93ea"} err="failed to get container status \"0cb76d614fd313ac58f734627ceef6ce03d41309478e9627230b2a1b511a93ea\": rpc error: code = NotFound desc = could not find container \"0cb76d614fd313ac58f734627ceef6ce03d41309478e9627230b2a1b511a93ea\": container with ID starting with 0cb76d614fd313ac58f734627ceef6ce03d41309478e9627230b2a1b511a93ea not found: ID does not exist" Jan 21 11:13:52 crc kubenswrapper[4824]: I0121 11:13:52.137561 4824 scope.go:117] "RemoveContainer" containerID="f2ddd6ed1029c405c4138f6257aaa84375743b3602a5b630fdf5b86e73651795" Jan 21 11:13:52 crc kubenswrapper[4824]: E0121 11:13:52.137840 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f2ddd6ed1029c405c4138f6257aaa84375743b3602a5b630fdf5b86e73651795\": container with ID starting with f2ddd6ed1029c405c4138f6257aaa84375743b3602a5b630fdf5b86e73651795 not found: ID does not exist" containerID="f2ddd6ed1029c405c4138f6257aaa84375743b3602a5b630fdf5b86e73651795" Jan 21 11:13:52 crc kubenswrapper[4824]: I0121 11:13:52.137868 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f2ddd6ed1029c405c4138f6257aaa84375743b3602a5b630fdf5b86e73651795"} err="failed to get container status \"f2ddd6ed1029c405c4138f6257aaa84375743b3602a5b630fdf5b86e73651795\": rpc error: code = NotFound desc = could not find container \"f2ddd6ed1029c405c4138f6257aaa84375743b3602a5b630fdf5b86e73651795\": container with ID starting with f2ddd6ed1029c405c4138f6257aaa84375743b3602a5b630fdf5b86e73651795 not found: ID does not exist" Jan 21 11:13:52 crc kubenswrapper[4824]: I0121 11:13:52.376395 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-8dj94"] Jan 21 11:13:52 crc kubenswrapper[4824]: I0121 11:13:52.376790 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-8dj94" podUID="ff7a0886-d523-486a-9bf0-bae3456b85c2" containerName="registry-server" containerID="cri-o://963e8246205e0fbf3da3eb5eadb52a32e5194169b632c986a341b4e6c7b913dd" gracePeriod=2 Jan 21 11:13:52 crc kubenswrapper[4824]: I0121 11:13:52.421331 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-bkkzh" Jan 21 11:13:52 crc kubenswrapper[4824]: I0121 11:13:52.728693 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8dj94" Jan 21 11:13:52 crc kubenswrapper[4824]: I0121 11:13:52.862555 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tskqq\" (UniqueName: \"kubernetes.io/projected/ff7a0886-d523-486a-9bf0-bae3456b85c2-kube-api-access-tskqq\") pod \"ff7a0886-d523-486a-9bf0-bae3456b85c2\" (UID: \"ff7a0886-d523-486a-9bf0-bae3456b85c2\") " Jan 21 11:13:52 crc kubenswrapper[4824]: I0121 11:13:52.862679 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ff7a0886-d523-486a-9bf0-bae3456b85c2-utilities\") pod \"ff7a0886-d523-486a-9bf0-bae3456b85c2\" (UID: \"ff7a0886-d523-486a-9bf0-bae3456b85c2\") " Jan 21 11:13:52 crc kubenswrapper[4824]: I0121 11:13:52.862717 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ff7a0886-d523-486a-9bf0-bae3456b85c2-catalog-content\") pod \"ff7a0886-d523-486a-9bf0-bae3456b85c2\" (UID: \"ff7a0886-d523-486a-9bf0-bae3456b85c2\") " Jan 21 11:13:52 crc kubenswrapper[4824]: I0121 11:13:52.868347 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ff7a0886-d523-486a-9bf0-bae3456b85c2-kube-api-access-tskqq" (OuterVolumeSpecName: "kube-api-access-tskqq") pod "ff7a0886-d523-486a-9bf0-bae3456b85c2" (UID: "ff7a0886-d523-486a-9bf0-bae3456b85c2"). InnerVolumeSpecName "kube-api-access-tskqq". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:13:52 crc kubenswrapper[4824]: I0121 11:13:52.877979 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ff7a0886-d523-486a-9bf0-bae3456b85c2-utilities" (OuterVolumeSpecName: "utilities") pod "ff7a0886-d523-486a-9bf0-bae3456b85c2" (UID: "ff7a0886-d523-486a-9bf0-bae3456b85c2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:13:52 crc kubenswrapper[4824]: I0121 11:13:52.879550 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ff7a0886-d523-486a-9bf0-bae3456b85c2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ff7a0886-d523-486a-9bf0-bae3456b85c2" (UID: "ff7a0886-d523-486a-9bf0-bae3456b85c2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:13:52 crc kubenswrapper[4824]: I0121 11:13:52.963638 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ff7a0886-d523-486a-9bf0-bae3456b85c2-utilities\") on node \"crc\" DevicePath \"\"" Jan 21 11:13:52 crc kubenswrapper[4824]: I0121 11:13:52.963661 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ff7a0886-d523-486a-9bf0-bae3456b85c2-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 21 11:13:52 crc kubenswrapper[4824]: I0121 11:13:52.963673 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tskqq\" (UniqueName: \"kubernetes.io/projected/ff7a0886-d523-486a-9bf0-bae3456b85c2-kube-api-access-tskqq\") on node \"crc\" DevicePath \"\"" Jan 21 11:13:53 crc kubenswrapper[4824]: I0121 11:13:53.108839 4824 generic.go:334] "Generic (PLEG): container finished" podID="ff7a0886-d523-486a-9bf0-bae3456b85c2" containerID="963e8246205e0fbf3da3eb5eadb52a32e5194169b632c986a341b4e6c7b913dd" exitCode=0 Jan 21 11:13:53 crc kubenswrapper[4824]: I0121 11:13:53.108898 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8dj94" Jan 21 11:13:53 crc kubenswrapper[4824]: I0121 11:13:53.108905 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8dj94" event={"ID":"ff7a0886-d523-486a-9bf0-bae3456b85c2","Type":"ContainerDied","Data":"963e8246205e0fbf3da3eb5eadb52a32e5194169b632c986a341b4e6c7b913dd"} Jan 21 11:13:53 crc kubenswrapper[4824]: I0121 11:13:53.108984 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8dj94" event={"ID":"ff7a0886-d523-486a-9bf0-bae3456b85c2","Type":"ContainerDied","Data":"1cc0a8fad4ae405d30baab5682d5e3fd0a25f77c99ddd8f27c737f65b8432c15"} Jan 21 11:13:53 crc kubenswrapper[4824]: I0121 11:13:53.109012 4824 scope.go:117] "RemoveContainer" containerID="963e8246205e0fbf3da3eb5eadb52a32e5194169b632c986a341b4e6c7b913dd" Jan 21 11:13:53 crc kubenswrapper[4824]: I0121 11:13:53.120551 4824 scope.go:117] "RemoveContainer" containerID="294b6d02b13eca2d11ff429f1d4ad8e41b5d9b3b5d8e4c71c522e3fcaf16fe06" Jan 21 11:13:53 crc kubenswrapper[4824]: I0121 11:13:53.130005 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-8dj94"] Jan 21 11:13:53 crc kubenswrapper[4824]: I0121 11:13:53.132231 4824 scope.go:117] "RemoveContainer" containerID="a9c629df6b9dc7cf2e177356c8434961e0a88d27574a92b65159b29366efbef3" Jan 21 11:13:53 crc kubenswrapper[4824]: I0121 11:13:53.132316 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-8dj94"] Jan 21 11:13:53 crc kubenswrapper[4824]: I0121 11:13:53.159155 4824 scope.go:117] "RemoveContainer" containerID="963e8246205e0fbf3da3eb5eadb52a32e5194169b632c986a341b4e6c7b913dd" Jan 21 11:13:53 crc kubenswrapper[4824]: E0121 11:13:53.159449 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"963e8246205e0fbf3da3eb5eadb52a32e5194169b632c986a341b4e6c7b913dd\": container with ID starting with 963e8246205e0fbf3da3eb5eadb52a32e5194169b632c986a341b4e6c7b913dd not found: ID does not exist" containerID="963e8246205e0fbf3da3eb5eadb52a32e5194169b632c986a341b4e6c7b913dd" Jan 21 11:13:53 crc kubenswrapper[4824]: I0121 11:13:53.159478 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"963e8246205e0fbf3da3eb5eadb52a32e5194169b632c986a341b4e6c7b913dd"} err="failed to get container status \"963e8246205e0fbf3da3eb5eadb52a32e5194169b632c986a341b4e6c7b913dd\": rpc error: code = NotFound desc = could not find container \"963e8246205e0fbf3da3eb5eadb52a32e5194169b632c986a341b4e6c7b913dd\": container with ID starting with 963e8246205e0fbf3da3eb5eadb52a32e5194169b632c986a341b4e6c7b913dd not found: ID does not exist" Jan 21 11:13:53 crc kubenswrapper[4824]: I0121 11:13:53.159502 4824 scope.go:117] "RemoveContainer" containerID="294b6d02b13eca2d11ff429f1d4ad8e41b5d9b3b5d8e4c71c522e3fcaf16fe06" Jan 21 11:13:53 crc kubenswrapper[4824]: E0121 11:13:53.159758 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"294b6d02b13eca2d11ff429f1d4ad8e41b5d9b3b5d8e4c71c522e3fcaf16fe06\": container with ID starting with 294b6d02b13eca2d11ff429f1d4ad8e41b5d9b3b5d8e4c71c522e3fcaf16fe06 not found: ID does not exist" containerID="294b6d02b13eca2d11ff429f1d4ad8e41b5d9b3b5d8e4c71c522e3fcaf16fe06" Jan 21 11:13:53 crc kubenswrapper[4824]: I0121 11:13:53.159781 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"294b6d02b13eca2d11ff429f1d4ad8e41b5d9b3b5d8e4c71c522e3fcaf16fe06"} err="failed to get container status \"294b6d02b13eca2d11ff429f1d4ad8e41b5d9b3b5d8e4c71c522e3fcaf16fe06\": rpc error: code = NotFound desc = could not find container \"294b6d02b13eca2d11ff429f1d4ad8e41b5d9b3b5d8e4c71c522e3fcaf16fe06\": container with ID starting with 294b6d02b13eca2d11ff429f1d4ad8e41b5d9b3b5d8e4c71c522e3fcaf16fe06 not found: ID does not exist" Jan 21 11:13:53 crc kubenswrapper[4824]: I0121 11:13:53.159793 4824 scope.go:117] "RemoveContainer" containerID="a9c629df6b9dc7cf2e177356c8434961e0a88d27574a92b65159b29366efbef3" Jan 21 11:13:53 crc kubenswrapper[4824]: E0121 11:13:53.160297 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a9c629df6b9dc7cf2e177356c8434961e0a88d27574a92b65159b29366efbef3\": container with ID starting with a9c629df6b9dc7cf2e177356c8434961e0a88d27574a92b65159b29366efbef3 not found: ID does not exist" containerID="a9c629df6b9dc7cf2e177356c8434961e0a88d27574a92b65159b29366efbef3" Jan 21 11:13:53 crc kubenswrapper[4824]: I0121 11:13:53.160315 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a9c629df6b9dc7cf2e177356c8434961e0a88d27574a92b65159b29366efbef3"} err="failed to get container status \"a9c629df6b9dc7cf2e177356c8434961e0a88d27574a92b65159b29366efbef3\": rpc error: code = NotFound desc = could not find container \"a9c629df6b9dc7cf2e177356c8434961e0a88d27574a92b65159b29366efbef3\": container with ID starting with a9c629df6b9dc7cf2e177356c8434961e0a88d27574a92b65159b29366efbef3 not found: ID does not exist" Jan 21 11:13:54 crc kubenswrapper[4824]: I0121 11:13:54.054048 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0939eda4-fff7-447b-ad52-ddb427bf1272" path="/var/lib/kubelet/pods/0939eda4-fff7-447b-ad52-ddb427bf1272/volumes" Jan 21 11:13:54 crc kubenswrapper[4824]: I0121 11:13:54.054570 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ff7a0886-d523-486a-9bf0-bae3456b85c2" path="/var/lib/kubelet/pods/ff7a0886-d523-486a-9bf0-bae3456b85c2/volumes" Jan 21 11:13:54 crc kubenswrapper[4824]: I0121 11:13:54.774562 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-2qhk6"] Jan 21 11:13:54 crc kubenswrapper[4824]: I0121 11:13:54.774788 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-2qhk6" podUID="30363b0b-09e8-4fb6-88e4-4e941d833ec8" containerName="registry-server" containerID="cri-o://812b638a67ec88d3b10766fb23536e55485cc458e7ca4b815bcac4f9331b7868" gracePeriod=2 Jan 21 11:13:55 crc kubenswrapper[4824]: I0121 11:13:55.123372 4824 generic.go:334] "Generic (PLEG): container finished" podID="30363b0b-09e8-4fb6-88e4-4e941d833ec8" containerID="812b638a67ec88d3b10766fb23536e55485cc458e7ca4b815bcac4f9331b7868" exitCode=0 Jan 21 11:13:55 crc kubenswrapper[4824]: I0121 11:13:55.123411 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2qhk6" event={"ID":"30363b0b-09e8-4fb6-88e4-4e941d833ec8","Type":"ContainerDied","Data":"812b638a67ec88d3b10766fb23536e55485cc458e7ca4b815bcac4f9331b7868"} Jan 21 11:13:55 crc kubenswrapper[4824]: I0121 11:13:55.123432 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2qhk6" event={"ID":"30363b0b-09e8-4fb6-88e4-4e941d833ec8","Type":"ContainerDied","Data":"e23807ae2071e0cd34acdbc48558665e201952b8472f9a00085254fc560b0ace"} Jan 21 11:13:55 crc kubenswrapper[4824]: I0121 11:13:55.123442 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e23807ae2071e0cd34acdbc48558665e201952b8472f9a00085254fc560b0ace" Jan 21 11:13:55 crc kubenswrapper[4824]: I0121 11:13:55.142142 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2qhk6" Jan 21 11:13:55 crc kubenswrapper[4824]: I0121 11:13:55.193361 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcjz2\" (UniqueName: \"kubernetes.io/projected/30363b0b-09e8-4fb6-88e4-4e941d833ec8-kube-api-access-xcjz2\") pod \"30363b0b-09e8-4fb6-88e4-4e941d833ec8\" (UID: \"30363b0b-09e8-4fb6-88e4-4e941d833ec8\") " Jan 21 11:13:55 crc kubenswrapper[4824]: I0121 11:13:55.193453 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/30363b0b-09e8-4fb6-88e4-4e941d833ec8-catalog-content\") pod \"30363b0b-09e8-4fb6-88e4-4e941d833ec8\" (UID: \"30363b0b-09e8-4fb6-88e4-4e941d833ec8\") " Jan 21 11:13:55 crc kubenswrapper[4824]: I0121 11:13:55.193482 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/30363b0b-09e8-4fb6-88e4-4e941d833ec8-utilities\") pod \"30363b0b-09e8-4fb6-88e4-4e941d833ec8\" (UID: \"30363b0b-09e8-4fb6-88e4-4e941d833ec8\") " Jan 21 11:13:55 crc kubenswrapper[4824]: I0121 11:13:55.194280 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/30363b0b-09e8-4fb6-88e4-4e941d833ec8-utilities" (OuterVolumeSpecName: "utilities") pod "30363b0b-09e8-4fb6-88e4-4e941d833ec8" (UID: "30363b0b-09e8-4fb6-88e4-4e941d833ec8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:13:55 crc kubenswrapper[4824]: I0121 11:13:55.198068 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/30363b0b-09e8-4fb6-88e4-4e941d833ec8-kube-api-access-xcjz2" (OuterVolumeSpecName: "kube-api-access-xcjz2") pod "30363b0b-09e8-4fb6-88e4-4e941d833ec8" (UID: "30363b0b-09e8-4fb6-88e4-4e941d833ec8"). InnerVolumeSpecName "kube-api-access-xcjz2". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:13:55 crc kubenswrapper[4824]: I0121 11:13:55.288457 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/30363b0b-09e8-4fb6-88e4-4e941d833ec8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "30363b0b-09e8-4fb6-88e4-4e941d833ec8" (UID: "30363b0b-09e8-4fb6-88e4-4e941d833ec8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:13:55 crc kubenswrapper[4824]: I0121 11:13:55.294256 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/30363b0b-09e8-4fb6-88e4-4e941d833ec8-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 21 11:13:55 crc kubenswrapper[4824]: I0121 11:13:55.294276 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/30363b0b-09e8-4fb6-88e4-4e941d833ec8-utilities\") on node \"crc\" DevicePath \"\"" Jan 21 11:13:55 crc kubenswrapper[4824]: I0121 11:13:55.294286 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcjz2\" (UniqueName: \"kubernetes.io/projected/30363b0b-09e8-4fb6-88e4-4e941d833ec8-kube-api-access-xcjz2\") on node \"crc\" DevicePath \"\"" Jan 21 11:13:56 crc kubenswrapper[4824]: I0121 11:13:56.127175 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2qhk6" Jan 21 11:13:56 crc kubenswrapper[4824]: I0121 11:13:56.139286 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-2qhk6"] Jan 21 11:13:56 crc kubenswrapper[4824]: I0121 11:13:56.141565 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-2qhk6"] Jan 21 11:13:58 crc kubenswrapper[4824]: I0121 11:13:58.053290 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="30363b0b-09e8-4fb6-88e4-4e941d833ec8" path="/var/lib/kubelet/pods/30363b0b-09e8-4fb6-88e4-4e941d833ec8/volumes" Jan 21 11:13:58 crc kubenswrapper[4824]: I0121 11:13:58.603991 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Jan 21 11:13:58 crc kubenswrapper[4824]: E0121 11:13:58.604319 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30363b0b-09e8-4fb6-88e4-4e941d833ec8" containerName="extract-content" Jan 21 11:13:58 crc kubenswrapper[4824]: I0121 11:13:58.604336 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="30363b0b-09e8-4fb6-88e4-4e941d833ec8" containerName="extract-content" Jan 21 11:13:58 crc kubenswrapper[4824]: E0121 11:13:58.604348 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b5068e5-8cc3-4ed1-811c-835a08eefaa0" containerName="pruner" Jan 21 11:13:58 crc kubenswrapper[4824]: I0121 11:13:58.604354 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b5068e5-8cc3-4ed1-811c-835a08eefaa0" containerName="pruner" Jan 21 11:13:58 crc kubenswrapper[4824]: E0121 11:13:58.604360 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30363b0b-09e8-4fb6-88e4-4e941d833ec8" containerName="extract-utilities" Jan 21 11:13:58 crc kubenswrapper[4824]: I0121 11:13:58.604365 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="30363b0b-09e8-4fb6-88e4-4e941d833ec8" containerName="extract-utilities" Jan 21 11:13:58 crc kubenswrapper[4824]: E0121 11:13:58.604374 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff7a0886-d523-486a-9bf0-bae3456b85c2" containerName="extract-utilities" Jan 21 11:13:58 crc kubenswrapper[4824]: I0121 11:13:58.604380 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff7a0886-d523-486a-9bf0-bae3456b85c2" containerName="extract-utilities" Jan 21 11:13:58 crc kubenswrapper[4824]: E0121 11:13:58.604391 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4205fbde-9c6c-40c8-8c8c-d34a607923e4" containerName="extract-content" Jan 21 11:13:58 crc kubenswrapper[4824]: I0121 11:13:58.604396 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="4205fbde-9c6c-40c8-8c8c-d34a607923e4" containerName="extract-content" Jan 21 11:13:58 crc kubenswrapper[4824]: E0121 11:13:58.604404 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4205fbde-9c6c-40c8-8c8c-d34a607923e4" containerName="registry-server" Jan 21 11:13:58 crc kubenswrapper[4824]: I0121 11:13:58.604409 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="4205fbde-9c6c-40c8-8c8c-d34a607923e4" containerName="registry-server" Jan 21 11:13:58 crc kubenswrapper[4824]: E0121 11:13:58.604416 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0939eda4-fff7-447b-ad52-ddb427bf1272" containerName="registry-server" Jan 21 11:13:58 crc kubenswrapper[4824]: I0121 11:13:58.604422 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="0939eda4-fff7-447b-ad52-ddb427bf1272" containerName="registry-server" Jan 21 11:13:58 crc kubenswrapper[4824]: E0121 11:13:58.604430 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0939eda4-fff7-447b-ad52-ddb427bf1272" containerName="extract-content" Jan 21 11:13:58 crc kubenswrapper[4824]: I0121 11:13:58.604435 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="0939eda4-fff7-447b-ad52-ddb427bf1272" containerName="extract-content" Jan 21 11:13:58 crc kubenswrapper[4824]: E0121 11:13:58.604446 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30363b0b-09e8-4fb6-88e4-4e941d833ec8" containerName="registry-server" Jan 21 11:13:58 crc kubenswrapper[4824]: I0121 11:13:58.604451 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="30363b0b-09e8-4fb6-88e4-4e941d833ec8" containerName="registry-server" Jan 21 11:13:58 crc kubenswrapper[4824]: E0121 11:13:58.604458 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff7a0886-d523-486a-9bf0-bae3456b85c2" containerName="extract-content" Jan 21 11:13:58 crc kubenswrapper[4824]: I0121 11:13:58.604463 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff7a0886-d523-486a-9bf0-bae3456b85c2" containerName="extract-content" Jan 21 11:13:58 crc kubenswrapper[4824]: E0121 11:13:58.604471 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4205fbde-9c6c-40c8-8c8c-d34a607923e4" containerName="extract-utilities" Jan 21 11:13:58 crc kubenswrapper[4824]: I0121 11:13:58.604476 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="4205fbde-9c6c-40c8-8c8c-d34a607923e4" containerName="extract-utilities" Jan 21 11:13:58 crc kubenswrapper[4824]: E0121 11:13:58.604484 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff7a0886-d523-486a-9bf0-bae3456b85c2" containerName="registry-server" Jan 21 11:13:58 crc kubenswrapper[4824]: I0121 11:13:58.604489 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff7a0886-d523-486a-9bf0-bae3456b85c2" containerName="registry-server" Jan 21 11:13:58 crc kubenswrapper[4824]: E0121 11:13:58.604497 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5a6cefc5-7c75-40ca-b14e-d0f7615b7cec" containerName="pruner" Jan 21 11:13:58 crc kubenswrapper[4824]: I0121 11:13:58.604501 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="5a6cefc5-7c75-40ca-b14e-d0f7615b7cec" containerName="pruner" Jan 21 11:13:58 crc kubenswrapper[4824]: E0121 11:13:58.604507 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0939eda4-fff7-447b-ad52-ddb427bf1272" containerName="extract-utilities" Jan 21 11:13:58 crc kubenswrapper[4824]: I0121 11:13:58.604512 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="0939eda4-fff7-447b-ad52-ddb427bf1272" containerName="extract-utilities" Jan 21 11:13:58 crc kubenswrapper[4824]: I0121 11:13:58.604602 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="5a6cefc5-7c75-40ca-b14e-d0f7615b7cec" containerName="pruner" Jan 21 11:13:58 crc kubenswrapper[4824]: I0121 11:13:58.604610 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="ff7a0886-d523-486a-9bf0-bae3456b85c2" containerName="registry-server" Jan 21 11:13:58 crc kubenswrapper[4824]: I0121 11:13:58.604616 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="0939eda4-fff7-447b-ad52-ddb427bf1272" containerName="registry-server" Jan 21 11:13:58 crc kubenswrapper[4824]: I0121 11:13:58.604623 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="5b5068e5-8cc3-4ed1-811c-835a08eefaa0" containerName="pruner" Jan 21 11:13:58 crc kubenswrapper[4824]: I0121 11:13:58.604632 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="4205fbde-9c6c-40c8-8c8c-d34a607923e4" containerName="registry-server" Jan 21 11:13:58 crc kubenswrapper[4824]: I0121 11:13:58.604640 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="30363b0b-09e8-4fb6-88e4-4e941d833ec8" containerName="registry-server" Jan 21 11:13:58 crc kubenswrapper[4824]: I0121 11:13:58.604973 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 21 11:13:58 crc kubenswrapper[4824]: I0121 11:13:58.607015 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Jan 21 11:13:58 crc kubenswrapper[4824]: I0121 11:13:58.607122 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Jan 21 11:13:58 crc kubenswrapper[4824]: I0121 11:13:58.611951 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Jan 21 11:13:58 crc kubenswrapper[4824]: I0121 11:13:58.628474 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/bb3ac9c9-8f2c-46e7-bd3c-425f42d66a2a-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"bb3ac9c9-8f2c-46e7-bd3c-425f42d66a2a\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 21 11:13:58 crc kubenswrapper[4824]: I0121 11:13:58.628552 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/bb3ac9c9-8f2c-46e7-bd3c-425f42d66a2a-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"bb3ac9c9-8f2c-46e7-bd3c-425f42d66a2a\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 21 11:13:58 crc kubenswrapper[4824]: I0121 11:13:58.729297 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/bb3ac9c9-8f2c-46e7-bd3c-425f42d66a2a-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"bb3ac9c9-8f2c-46e7-bd3c-425f42d66a2a\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 21 11:13:58 crc kubenswrapper[4824]: I0121 11:13:58.729391 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/bb3ac9c9-8f2c-46e7-bd3c-425f42d66a2a-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"bb3ac9c9-8f2c-46e7-bd3c-425f42d66a2a\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 21 11:13:58 crc kubenswrapper[4824]: I0121 11:13:58.729393 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/bb3ac9c9-8f2c-46e7-bd3c-425f42d66a2a-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"bb3ac9c9-8f2c-46e7-bd3c-425f42d66a2a\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 21 11:13:58 crc kubenswrapper[4824]: I0121 11:13:58.743680 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/bb3ac9c9-8f2c-46e7-bd3c-425f42d66a2a-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"bb3ac9c9-8f2c-46e7-bd3c-425f42d66a2a\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 21 11:13:58 crc kubenswrapper[4824]: I0121 11:13:58.924739 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 21 11:13:59 crc kubenswrapper[4824]: I0121 11:13:59.252083 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Jan 21 11:13:59 crc kubenswrapper[4824]: W0121 11:13:59.256669 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-podbb3ac9c9_8f2c_46e7_bd3c_425f42d66a2a.slice/crio-37bf124b210287b69472d2d75fa4a168942f9d0045a31894cabe7d2285f31fc7 WatchSource:0}: Error finding container 37bf124b210287b69472d2d75fa4a168942f9d0045a31894cabe7d2285f31fc7: Status 404 returned error can't find the container with id 37bf124b210287b69472d2d75fa4a168942f9d0045a31894cabe7d2285f31fc7 Jan 21 11:14:00 crc kubenswrapper[4824]: I0121 11:14:00.142510 4824 generic.go:334] "Generic (PLEG): container finished" podID="bb3ac9c9-8f2c-46e7-bd3c-425f42d66a2a" containerID="a0e75d242bdfcf7d63ffb16f6b896e2df3292d373940f672bdb93872e312173c" exitCode=0 Jan 21 11:14:00 crc kubenswrapper[4824]: I0121 11:14:00.142543 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"bb3ac9c9-8f2c-46e7-bd3c-425f42d66a2a","Type":"ContainerDied","Data":"a0e75d242bdfcf7d63ffb16f6b896e2df3292d373940f672bdb93872e312173c"} Jan 21 11:14:00 crc kubenswrapper[4824]: I0121 11:14:00.142701 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"bb3ac9c9-8f2c-46e7-bd3c-425f42d66a2a","Type":"ContainerStarted","Data":"37bf124b210287b69472d2d75fa4a168942f9d0045a31894cabe7d2285f31fc7"} Jan 21 11:14:01 crc kubenswrapper[4824]: I0121 11:14:01.394362 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 21 11:14:01 crc kubenswrapper[4824]: I0121 11:14:01.560108 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/bb3ac9c9-8f2c-46e7-bd3c-425f42d66a2a-kube-api-access\") pod \"bb3ac9c9-8f2c-46e7-bd3c-425f42d66a2a\" (UID: \"bb3ac9c9-8f2c-46e7-bd3c-425f42d66a2a\") " Jan 21 11:14:01 crc kubenswrapper[4824]: I0121 11:14:01.560202 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/bb3ac9c9-8f2c-46e7-bd3c-425f42d66a2a-kubelet-dir\") pod \"bb3ac9c9-8f2c-46e7-bd3c-425f42d66a2a\" (UID: \"bb3ac9c9-8f2c-46e7-bd3c-425f42d66a2a\") " Jan 21 11:14:01 crc kubenswrapper[4824]: I0121 11:14:01.560317 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bb3ac9c9-8f2c-46e7-bd3c-425f42d66a2a-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "bb3ac9c9-8f2c-46e7-bd3c-425f42d66a2a" (UID: "bb3ac9c9-8f2c-46e7-bd3c-425f42d66a2a"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 21 11:14:01 crc kubenswrapper[4824]: I0121 11:14:01.560491 4824 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/bb3ac9c9-8f2c-46e7-bd3c-425f42d66a2a-kubelet-dir\") on node \"crc\" DevicePath \"\"" Jan 21 11:14:01 crc kubenswrapper[4824]: I0121 11:14:01.564494 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bb3ac9c9-8f2c-46e7-bd3c-425f42d66a2a-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "bb3ac9c9-8f2c-46e7-bd3c-425f42d66a2a" (UID: "bb3ac9c9-8f2c-46e7-bd3c-425f42d66a2a"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:14:01 crc kubenswrapper[4824]: I0121 11:14:01.661434 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/bb3ac9c9-8f2c-46e7-bd3c-425f42d66a2a-kube-api-access\") on node \"crc\" DevicePath \"\"" Jan 21 11:14:02 crc kubenswrapper[4824]: I0121 11:14:02.151653 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"bb3ac9c9-8f2c-46e7-bd3c-425f42d66a2a","Type":"ContainerDied","Data":"37bf124b210287b69472d2d75fa4a168942f9d0045a31894cabe7d2285f31fc7"} Jan 21 11:14:02 crc kubenswrapper[4824]: I0121 11:14:02.151687 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="37bf124b210287b69472d2d75fa4a168942f9d0045a31894cabe7d2285f31fc7" Jan 21 11:14:02 crc kubenswrapper[4824]: I0121 11:14:02.151734 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 21 11:14:03 crc kubenswrapper[4824]: I0121 11:14:03.369576 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 21 11:14:05 crc kubenswrapper[4824]: I0121 11:14:05.603320 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Jan 21 11:14:05 crc kubenswrapper[4824]: E0121 11:14:05.603508 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb3ac9c9-8f2c-46e7-bd3c-425f42d66a2a" containerName="pruner" Jan 21 11:14:05 crc kubenswrapper[4824]: I0121 11:14:05.603519 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb3ac9c9-8f2c-46e7-bd3c-425f42d66a2a" containerName="pruner" Jan 21 11:14:05 crc kubenswrapper[4824]: I0121 11:14:05.603608 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="bb3ac9c9-8f2c-46e7-bd3c-425f42d66a2a" containerName="pruner" Jan 21 11:14:05 crc kubenswrapper[4824]: I0121 11:14:05.603986 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Jan 21 11:14:05 crc kubenswrapper[4824]: I0121 11:14:05.605367 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Jan 21 11:14:05 crc kubenswrapper[4824]: I0121 11:14:05.605415 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Jan 21 11:14:05 crc kubenswrapper[4824]: I0121 11:14:05.611040 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Jan 21 11:14:05 crc kubenswrapper[4824]: I0121 11:14:05.699109 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a83fe5bb-7790-45ea-8f0b-56799a5c4b89-kubelet-dir\") pod \"installer-9-crc\" (UID: \"a83fe5bb-7790-45ea-8f0b-56799a5c4b89\") " pod="openshift-kube-apiserver/installer-9-crc" Jan 21 11:14:05 crc kubenswrapper[4824]: I0121 11:14:05.699164 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a83fe5bb-7790-45ea-8f0b-56799a5c4b89-kube-api-access\") pod \"installer-9-crc\" (UID: \"a83fe5bb-7790-45ea-8f0b-56799a5c4b89\") " pod="openshift-kube-apiserver/installer-9-crc" Jan 21 11:14:05 crc kubenswrapper[4824]: I0121 11:14:05.699183 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/a83fe5bb-7790-45ea-8f0b-56799a5c4b89-var-lock\") pod \"installer-9-crc\" (UID: \"a83fe5bb-7790-45ea-8f0b-56799a5c4b89\") " pod="openshift-kube-apiserver/installer-9-crc" Jan 21 11:14:05 crc kubenswrapper[4824]: I0121 11:14:05.800612 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a83fe5bb-7790-45ea-8f0b-56799a5c4b89-kubelet-dir\") pod \"installer-9-crc\" (UID: \"a83fe5bb-7790-45ea-8f0b-56799a5c4b89\") " pod="openshift-kube-apiserver/installer-9-crc" Jan 21 11:14:05 crc kubenswrapper[4824]: I0121 11:14:05.800668 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a83fe5bb-7790-45ea-8f0b-56799a5c4b89-kube-api-access\") pod \"installer-9-crc\" (UID: \"a83fe5bb-7790-45ea-8f0b-56799a5c4b89\") " pod="openshift-kube-apiserver/installer-9-crc" Jan 21 11:14:05 crc kubenswrapper[4824]: I0121 11:14:05.800687 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/a83fe5bb-7790-45ea-8f0b-56799a5c4b89-var-lock\") pod \"installer-9-crc\" (UID: \"a83fe5bb-7790-45ea-8f0b-56799a5c4b89\") " pod="openshift-kube-apiserver/installer-9-crc" Jan 21 11:14:05 crc kubenswrapper[4824]: I0121 11:14:05.800775 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/a83fe5bb-7790-45ea-8f0b-56799a5c4b89-var-lock\") pod \"installer-9-crc\" (UID: \"a83fe5bb-7790-45ea-8f0b-56799a5c4b89\") " pod="openshift-kube-apiserver/installer-9-crc" Jan 21 11:14:05 crc kubenswrapper[4824]: I0121 11:14:05.800969 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a83fe5bb-7790-45ea-8f0b-56799a5c4b89-kubelet-dir\") pod \"installer-9-crc\" (UID: \"a83fe5bb-7790-45ea-8f0b-56799a5c4b89\") " pod="openshift-kube-apiserver/installer-9-crc" Jan 21 11:14:05 crc kubenswrapper[4824]: I0121 11:14:05.815664 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a83fe5bb-7790-45ea-8f0b-56799a5c4b89-kube-api-access\") pod \"installer-9-crc\" (UID: \"a83fe5bb-7790-45ea-8f0b-56799a5c4b89\") " pod="openshift-kube-apiserver/installer-9-crc" Jan 21 11:14:05 crc kubenswrapper[4824]: I0121 11:14:05.915085 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Jan 21 11:14:06 crc kubenswrapper[4824]: I0121 11:14:06.258768 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Jan 21 11:14:07 crc kubenswrapper[4824]: I0121 11:14:07.175057 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"a83fe5bb-7790-45ea-8f0b-56799a5c4b89","Type":"ContainerStarted","Data":"8ea1c8d6b88fff12b2a363688bd7b4e3c72b3e77fe714e1ec1237fd44548b134"} Jan 21 11:14:07 crc kubenswrapper[4824]: I0121 11:14:07.175256 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"a83fe5bb-7790-45ea-8f0b-56799a5c4b89","Type":"ContainerStarted","Data":"939da8a2d6bcb80f9bcbb9fa2a8ace318e8eb04c7ccffba9384e787401a5fb5e"} Jan 21 11:14:07 crc kubenswrapper[4824]: I0121 11:14:07.187862 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/installer-9-crc" podStartSLOduration=2.187849945 podStartE2EDuration="2.187849945s" podCreationTimestamp="2026-01-21 11:14:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:14:07.184792603 +0000 UTC m=+189.477821896" watchObservedRunningTime="2026-01-21 11:14:07.187849945 +0000 UTC m=+189.480879238" Jan 21 11:14:10 crc kubenswrapper[4824]: I0121 11:14:10.813577 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-fz7gx"] Jan 21 11:14:16 crc kubenswrapper[4824]: I0121 11:14:16.065033 4824 patch_prober.go:28] interesting pod/machine-config-daemon-zqjg9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 21 11:14:16 crc kubenswrapper[4824]: I0121 11:14:16.065403 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 21 11:14:35 crc kubenswrapper[4824]: I0121 11:14:35.832531 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-fz7gx" podUID="86b5dca2-989f-4b6b-983b-87763ce0d469" containerName="oauth-openshift" containerID="cri-o://839b4db59e6f15755832bf687632140307e984c4933ab5994cda3f84fb56f80f" gracePeriod=15 Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.114887 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-fz7gx" Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.137631 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-7fdb5b7d8f-lf5gd"] Jan 21 11:14:36 crc kubenswrapper[4824]: E0121 11:14:36.137806 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="86b5dca2-989f-4b6b-983b-87763ce0d469" containerName="oauth-openshift" Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.137823 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="86b5dca2-989f-4b6b-983b-87763ce0d469" containerName="oauth-openshift" Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.137926 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="86b5dca2-989f-4b6b-983b-87763ce0d469" containerName="oauth-openshift" Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.138250 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-7fdb5b7d8f-lf5gd" Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.144610 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-7fdb5b7d8f-lf5gd"] Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.200774 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/86b5dca2-989f-4b6b-983b-87763ce0d469-v4-0-config-system-ocp-branding-template\") pod \"86b5dca2-989f-4b6b-983b-87763ce0d469\" (UID: \"86b5dca2-989f-4b6b-983b-87763ce0d469\") " Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.200809 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/86b5dca2-989f-4b6b-983b-87763ce0d469-v4-0-config-system-service-ca\") pod \"86b5dca2-989f-4b6b-983b-87763ce0d469\" (UID: \"86b5dca2-989f-4b6b-983b-87763ce0d469\") " Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.200829 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/86b5dca2-989f-4b6b-983b-87763ce0d469-v4-0-config-user-template-error\") pod \"86b5dca2-989f-4b6b-983b-87763ce0d469\" (UID: \"86b5dca2-989f-4b6b-983b-87763ce0d469\") " Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.200850 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bhtz6\" (UniqueName: \"kubernetes.io/projected/86b5dca2-989f-4b6b-983b-87763ce0d469-kube-api-access-bhtz6\") pod \"86b5dca2-989f-4b6b-983b-87763ce0d469\" (UID: \"86b5dca2-989f-4b6b-983b-87763ce0d469\") " Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.200875 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/86b5dca2-989f-4b6b-983b-87763ce0d469-v4-0-config-system-serving-cert\") pod \"86b5dca2-989f-4b6b-983b-87763ce0d469\" (UID: \"86b5dca2-989f-4b6b-983b-87763ce0d469\") " Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.200915 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/86b5dca2-989f-4b6b-983b-87763ce0d469-v4-0-config-system-router-certs\") pod \"86b5dca2-989f-4b6b-983b-87763ce0d469\" (UID: \"86b5dca2-989f-4b6b-983b-87763ce0d469\") " Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.200941 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/86b5dca2-989f-4b6b-983b-87763ce0d469-audit-dir\") pod \"86b5dca2-989f-4b6b-983b-87763ce0d469\" (UID: \"86b5dca2-989f-4b6b-983b-87763ce0d469\") " Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.200978 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/86b5dca2-989f-4b6b-983b-87763ce0d469-v4-0-config-user-template-provider-selection\") pod \"86b5dca2-989f-4b6b-983b-87763ce0d469\" (UID: \"86b5dca2-989f-4b6b-983b-87763ce0d469\") " Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.200996 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/86b5dca2-989f-4b6b-983b-87763ce0d469-audit-policies\") pod \"86b5dca2-989f-4b6b-983b-87763ce0d469\" (UID: \"86b5dca2-989f-4b6b-983b-87763ce0d469\") " Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.201017 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/86b5dca2-989f-4b6b-983b-87763ce0d469-v4-0-config-user-template-login\") pod \"86b5dca2-989f-4b6b-983b-87763ce0d469\" (UID: \"86b5dca2-989f-4b6b-983b-87763ce0d469\") " Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.201049 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/86b5dca2-989f-4b6b-983b-87763ce0d469-v4-0-config-system-session\") pod \"86b5dca2-989f-4b6b-983b-87763ce0d469\" (UID: \"86b5dca2-989f-4b6b-983b-87763ce0d469\") " Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.201075 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/86b5dca2-989f-4b6b-983b-87763ce0d469-v4-0-config-user-idp-0-file-data\") pod \"86b5dca2-989f-4b6b-983b-87763ce0d469\" (UID: \"86b5dca2-989f-4b6b-983b-87763ce0d469\") " Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.201098 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/86b5dca2-989f-4b6b-983b-87763ce0d469-v4-0-config-system-trusted-ca-bundle\") pod \"86b5dca2-989f-4b6b-983b-87763ce0d469\" (UID: \"86b5dca2-989f-4b6b-983b-87763ce0d469\") " Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.201115 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/86b5dca2-989f-4b6b-983b-87763ce0d469-v4-0-config-system-cliconfig\") pod \"86b5dca2-989f-4b6b-983b-87763ce0d469\" (UID: \"86b5dca2-989f-4b6b-983b-87763ce0d469\") " Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.201227 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/2322c755-9e87-4129-92f7-f525fbf333b4-v4-0-config-system-router-certs\") pod \"oauth-openshift-7fdb5b7d8f-lf5gd\" (UID: \"2322c755-9e87-4129-92f7-f525fbf333b4\") " pod="openshift-authentication/oauth-openshift-7fdb5b7d8f-lf5gd" Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.201249 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/2322c755-9e87-4129-92f7-f525fbf333b4-audit-dir\") pod \"oauth-openshift-7fdb5b7d8f-lf5gd\" (UID: \"2322c755-9e87-4129-92f7-f525fbf333b4\") " pod="openshift-authentication/oauth-openshift-7fdb5b7d8f-lf5gd" Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.201265 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2322c755-9e87-4129-92f7-f525fbf333b4-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-7fdb5b7d8f-lf5gd\" (UID: \"2322c755-9e87-4129-92f7-f525fbf333b4\") " pod="openshift-authentication/oauth-openshift-7fdb5b7d8f-lf5gd" Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.201284 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/2322c755-9e87-4129-92f7-f525fbf333b4-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-7fdb5b7d8f-lf5gd\" (UID: \"2322c755-9e87-4129-92f7-f525fbf333b4\") " pod="openshift-authentication/oauth-openshift-7fdb5b7d8f-lf5gd" Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.201305 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/2322c755-9e87-4129-92f7-f525fbf333b4-audit-policies\") pod \"oauth-openshift-7fdb5b7d8f-lf5gd\" (UID: \"2322c755-9e87-4129-92f7-f525fbf333b4\") " pod="openshift-authentication/oauth-openshift-7fdb5b7d8f-lf5gd" Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.201321 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/2322c755-9e87-4129-92f7-f525fbf333b4-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-7fdb5b7d8f-lf5gd\" (UID: \"2322c755-9e87-4129-92f7-f525fbf333b4\") " pod="openshift-authentication/oauth-openshift-7fdb5b7d8f-lf5gd" Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.201340 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/2322c755-9e87-4129-92f7-f525fbf333b4-v4-0-config-system-serving-cert\") pod \"oauth-openshift-7fdb5b7d8f-lf5gd\" (UID: \"2322c755-9e87-4129-92f7-f525fbf333b4\") " pod="openshift-authentication/oauth-openshift-7fdb5b7d8f-lf5gd" Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.201362 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-52gcl\" (UniqueName: \"kubernetes.io/projected/2322c755-9e87-4129-92f7-f525fbf333b4-kube-api-access-52gcl\") pod \"oauth-openshift-7fdb5b7d8f-lf5gd\" (UID: \"2322c755-9e87-4129-92f7-f525fbf333b4\") " pod="openshift-authentication/oauth-openshift-7fdb5b7d8f-lf5gd" Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.201397 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/2322c755-9e87-4129-92f7-f525fbf333b4-v4-0-config-system-cliconfig\") pod \"oauth-openshift-7fdb5b7d8f-lf5gd\" (UID: \"2322c755-9e87-4129-92f7-f525fbf333b4\") " pod="openshift-authentication/oauth-openshift-7fdb5b7d8f-lf5gd" Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.201418 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/2322c755-9e87-4129-92f7-f525fbf333b4-v4-0-config-user-template-login\") pod \"oauth-openshift-7fdb5b7d8f-lf5gd\" (UID: \"2322c755-9e87-4129-92f7-f525fbf333b4\") " pod="openshift-authentication/oauth-openshift-7fdb5b7d8f-lf5gd" Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.201444 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/2322c755-9e87-4129-92f7-f525fbf333b4-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-7fdb5b7d8f-lf5gd\" (UID: \"2322c755-9e87-4129-92f7-f525fbf333b4\") " pod="openshift-authentication/oauth-openshift-7fdb5b7d8f-lf5gd" Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.201463 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/2322c755-9e87-4129-92f7-f525fbf333b4-v4-0-config-user-template-error\") pod \"oauth-openshift-7fdb5b7d8f-lf5gd\" (UID: \"2322c755-9e87-4129-92f7-f525fbf333b4\") " pod="openshift-authentication/oauth-openshift-7fdb5b7d8f-lf5gd" Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.201481 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/2322c755-9e87-4129-92f7-f525fbf333b4-v4-0-config-system-session\") pod \"oauth-openshift-7fdb5b7d8f-lf5gd\" (UID: \"2322c755-9e87-4129-92f7-f525fbf333b4\") " pod="openshift-authentication/oauth-openshift-7fdb5b7d8f-lf5gd" Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.201495 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/2322c755-9e87-4129-92f7-f525fbf333b4-v4-0-config-system-service-ca\") pod \"oauth-openshift-7fdb5b7d8f-lf5gd\" (UID: \"2322c755-9e87-4129-92f7-f525fbf333b4\") " pod="openshift-authentication/oauth-openshift-7fdb5b7d8f-lf5gd" Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.201574 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/86b5dca2-989f-4b6b-983b-87763ce0d469-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "86b5dca2-989f-4b6b-983b-87763ce0d469" (UID: "86b5dca2-989f-4b6b-983b-87763ce0d469"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.202020 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/86b5dca2-989f-4b6b-983b-87763ce0d469-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "86b5dca2-989f-4b6b-983b-87763ce0d469" (UID: "86b5dca2-989f-4b6b-983b-87763ce0d469"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.202111 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/86b5dca2-989f-4b6b-983b-87763ce0d469-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "86b5dca2-989f-4b6b-983b-87763ce0d469" (UID: "86b5dca2-989f-4b6b-983b-87763ce0d469"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.202108 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/86b5dca2-989f-4b6b-983b-87763ce0d469-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "86b5dca2-989f-4b6b-983b-87763ce0d469" (UID: "86b5dca2-989f-4b6b-983b-87763ce0d469"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.202300 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/86b5dca2-989f-4b6b-983b-87763ce0d469-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "86b5dca2-989f-4b6b-983b-87763ce0d469" (UID: "86b5dca2-989f-4b6b-983b-87763ce0d469"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.208986 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/86b5dca2-989f-4b6b-983b-87763ce0d469-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "86b5dca2-989f-4b6b-983b-87763ce0d469" (UID: "86b5dca2-989f-4b6b-983b-87763ce0d469"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.209003 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/86b5dca2-989f-4b6b-983b-87763ce0d469-kube-api-access-bhtz6" (OuterVolumeSpecName: "kube-api-access-bhtz6") pod "86b5dca2-989f-4b6b-983b-87763ce0d469" (UID: "86b5dca2-989f-4b6b-983b-87763ce0d469"). InnerVolumeSpecName "kube-api-access-bhtz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.209241 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/86b5dca2-989f-4b6b-983b-87763ce0d469-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "86b5dca2-989f-4b6b-983b-87763ce0d469" (UID: "86b5dca2-989f-4b6b-983b-87763ce0d469"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.209494 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/86b5dca2-989f-4b6b-983b-87763ce0d469-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "86b5dca2-989f-4b6b-983b-87763ce0d469" (UID: "86b5dca2-989f-4b6b-983b-87763ce0d469"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.209664 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/86b5dca2-989f-4b6b-983b-87763ce0d469-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "86b5dca2-989f-4b6b-983b-87763ce0d469" (UID: "86b5dca2-989f-4b6b-983b-87763ce0d469"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.209895 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/86b5dca2-989f-4b6b-983b-87763ce0d469-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "86b5dca2-989f-4b6b-983b-87763ce0d469" (UID: "86b5dca2-989f-4b6b-983b-87763ce0d469"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.210116 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/86b5dca2-989f-4b6b-983b-87763ce0d469-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "86b5dca2-989f-4b6b-983b-87763ce0d469" (UID: "86b5dca2-989f-4b6b-983b-87763ce0d469"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.210227 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/86b5dca2-989f-4b6b-983b-87763ce0d469-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "86b5dca2-989f-4b6b-983b-87763ce0d469" (UID: "86b5dca2-989f-4b6b-983b-87763ce0d469"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.210417 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/86b5dca2-989f-4b6b-983b-87763ce0d469-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "86b5dca2-989f-4b6b-983b-87763ce0d469" (UID: "86b5dca2-989f-4b6b-983b-87763ce0d469"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.282622 4824 generic.go:334] "Generic (PLEG): container finished" podID="86b5dca2-989f-4b6b-983b-87763ce0d469" containerID="839b4db59e6f15755832bf687632140307e984c4933ab5994cda3f84fb56f80f" exitCode=0 Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.282646 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-fz7gx" Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.282659 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-fz7gx" event={"ID":"86b5dca2-989f-4b6b-983b-87763ce0d469","Type":"ContainerDied","Data":"839b4db59e6f15755832bf687632140307e984c4933ab5994cda3f84fb56f80f"} Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.282709 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-fz7gx" event={"ID":"86b5dca2-989f-4b6b-983b-87763ce0d469","Type":"ContainerDied","Data":"046a7d0640fbf37bae3b158d9284d146391f1ce208f167ac80fea14196ea051f"} Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.282725 4824 scope.go:117] "RemoveContainer" containerID="839b4db59e6f15755832bf687632140307e984c4933ab5994cda3f84fb56f80f" Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.295916 4824 scope.go:117] "RemoveContainer" containerID="839b4db59e6f15755832bf687632140307e984c4933ab5994cda3f84fb56f80f" Jan 21 11:14:36 crc kubenswrapper[4824]: E0121 11:14:36.296227 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"839b4db59e6f15755832bf687632140307e984c4933ab5994cda3f84fb56f80f\": container with ID starting with 839b4db59e6f15755832bf687632140307e984c4933ab5994cda3f84fb56f80f not found: ID does not exist" containerID="839b4db59e6f15755832bf687632140307e984c4933ab5994cda3f84fb56f80f" Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.296255 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"839b4db59e6f15755832bf687632140307e984c4933ab5994cda3f84fb56f80f"} err="failed to get container status \"839b4db59e6f15755832bf687632140307e984c4933ab5994cda3f84fb56f80f\": rpc error: code = NotFound desc = could not find container \"839b4db59e6f15755832bf687632140307e984c4933ab5994cda3f84fb56f80f\": container with ID starting with 839b4db59e6f15755832bf687632140307e984c4933ab5994cda3f84fb56f80f not found: ID does not exist" Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.302871 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/2322c755-9e87-4129-92f7-f525fbf333b4-v4-0-config-user-template-error\") pod \"oauth-openshift-7fdb5b7d8f-lf5gd\" (UID: \"2322c755-9e87-4129-92f7-f525fbf333b4\") " pod="openshift-authentication/oauth-openshift-7fdb5b7d8f-lf5gd" Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.303446 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-fz7gx"] Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.303490 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/2322c755-9e87-4129-92f7-f525fbf333b4-v4-0-config-system-session\") pod \"oauth-openshift-7fdb5b7d8f-lf5gd\" (UID: \"2322c755-9e87-4129-92f7-f525fbf333b4\") " pod="openshift-authentication/oauth-openshift-7fdb5b7d8f-lf5gd" Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.303512 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/2322c755-9e87-4129-92f7-f525fbf333b4-v4-0-config-system-service-ca\") pod \"oauth-openshift-7fdb5b7d8f-lf5gd\" (UID: \"2322c755-9e87-4129-92f7-f525fbf333b4\") " pod="openshift-authentication/oauth-openshift-7fdb5b7d8f-lf5gd" Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.303557 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/2322c755-9e87-4129-92f7-f525fbf333b4-v4-0-config-system-router-certs\") pod \"oauth-openshift-7fdb5b7d8f-lf5gd\" (UID: \"2322c755-9e87-4129-92f7-f525fbf333b4\") " pod="openshift-authentication/oauth-openshift-7fdb5b7d8f-lf5gd" Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.303573 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/2322c755-9e87-4129-92f7-f525fbf333b4-audit-dir\") pod \"oauth-openshift-7fdb5b7d8f-lf5gd\" (UID: \"2322c755-9e87-4129-92f7-f525fbf333b4\") " pod="openshift-authentication/oauth-openshift-7fdb5b7d8f-lf5gd" Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.303591 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2322c755-9e87-4129-92f7-f525fbf333b4-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-7fdb5b7d8f-lf5gd\" (UID: \"2322c755-9e87-4129-92f7-f525fbf333b4\") " pod="openshift-authentication/oauth-openshift-7fdb5b7d8f-lf5gd" Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.303611 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/2322c755-9e87-4129-92f7-f525fbf333b4-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-7fdb5b7d8f-lf5gd\" (UID: \"2322c755-9e87-4129-92f7-f525fbf333b4\") " pod="openshift-authentication/oauth-openshift-7fdb5b7d8f-lf5gd" Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.303631 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/2322c755-9e87-4129-92f7-f525fbf333b4-audit-policies\") pod \"oauth-openshift-7fdb5b7d8f-lf5gd\" (UID: \"2322c755-9e87-4129-92f7-f525fbf333b4\") " pod="openshift-authentication/oauth-openshift-7fdb5b7d8f-lf5gd" Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.303648 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/2322c755-9e87-4129-92f7-f525fbf333b4-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-7fdb5b7d8f-lf5gd\" (UID: \"2322c755-9e87-4129-92f7-f525fbf333b4\") " pod="openshift-authentication/oauth-openshift-7fdb5b7d8f-lf5gd" Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.303664 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/2322c755-9e87-4129-92f7-f525fbf333b4-v4-0-config-system-serving-cert\") pod \"oauth-openshift-7fdb5b7d8f-lf5gd\" (UID: \"2322c755-9e87-4129-92f7-f525fbf333b4\") " pod="openshift-authentication/oauth-openshift-7fdb5b7d8f-lf5gd" Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.303682 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-52gcl\" (UniqueName: \"kubernetes.io/projected/2322c755-9e87-4129-92f7-f525fbf333b4-kube-api-access-52gcl\") pod \"oauth-openshift-7fdb5b7d8f-lf5gd\" (UID: \"2322c755-9e87-4129-92f7-f525fbf333b4\") " pod="openshift-authentication/oauth-openshift-7fdb5b7d8f-lf5gd" Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.303709 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/2322c755-9e87-4129-92f7-f525fbf333b4-v4-0-config-system-cliconfig\") pod \"oauth-openshift-7fdb5b7d8f-lf5gd\" (UID: \"2322c755-9e87-4129-92f7-f525fbf333b4\") " pod="openshift-authentication/oauth-openshift-7fdb5b7d8f-lf5gd" Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.303730 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/2322c755-9e87-4129-92f7-f525fbf333b4-v4-0-config-user-template-login\") pod \"oauth-openshift-7fdb5b7d8f-lf5gd\" (UID: \"2322c755-9e87-4129-92f7-f525fbf333b4\") " pod="openshift-authentication/oauth-openshift-7fdb5b7d8f-lf5gd" Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.303748 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/2322c755-9e87-4129-92f7-f525fbf333b4-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-7fdb5b7d8f-lf5gd\" (UID: \"2322c755-9e87-4129-92f7-f525fbf333b4\") " pod="openshift-authentication/oauth-openshift-7fdb5b7d8f-lf5gd" Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.303781 4824 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/86b5dca2-989f-4b6b-983b-87763ce0d469-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.303790 4824 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/86b5dca2-989f-4b6b-983b-87763ce0d469-audit-dir\") on node \"crc\" DevicePath \"\"" Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.303802 4824 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/86b5dca2-989f-4b6b-983b-87763ce0d469-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.303810 4824 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/86b5dca2-989f-4b6b-983b-87763ce0d469-audit-policies\") on node \"crc\" DevicePath \"\"" Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.303840 4824 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/86b5dca2-989f-4b6b-983b-87763ce0d469-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.303850 4824 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/86b5dca2-989f-4b6b-983b-87763ce0d469-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.303859 4824 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/86b5dca2-989f-4b6b-983b-87763ce0d469-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.303867 4824 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/86b5dca2-989f-4b6b-983b-87763ce0d469-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.303876 4824 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/86b5dca2-989f-4b6b-983b-87763ce0d469-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.303885 4824 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/86b5dca2-989f-4b6b-983b-87763ce0d469-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.303893 4824 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/86b5dca2-989f-4b6b-983b-87763ce0d469-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.303903 4824 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/86b5dca2-989f-4b6b-983b-87763ce0d469-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.303911 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bhtz6\" (UniqueName: \"kubernetes.io/projected/86b5dca2-989f-4b6b-983b-87763ce0d469-kube-api-access-bhtz6\") on node \"crc\" DevicePath \"\"" Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.303920 4824 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/86b5dca2-989f-4b6b-983b-87763ce0d469-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.304423 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/2322c755-9e87-4129-92f7-f525fbf333b4-v4-0-config-system-service-ca\") pod \"oauth-openshift-7fdb5b7d8f-lf5gd\" (UID: \"2322c755-9e87-4129-92f7-f525fbf333b4\") " pod="openshift-authentication/oauth-openshift-7fdb5b7d8f-lf5gd" Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.304927 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/2322c755-9e87-4129-92f7-f525fbf333b4-audit-dir\") pod \"oauth-openshift-7fdb5b7d8f-lf5gd\" (UID: \"2322c755-9e87-4129-92f7-f525fbf333b4\") " pod="openshift-authentication/oauth-openshift-7fdb5b7d8f-lf5gd" Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.305205 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/2322c755-9e87-4129-92f7-f525fbf333b4-v4-0-config-system-cliconfig\") pod \"oauth-openshift-7fdb5b7d8f-lf5gd\" (UID: \"2322c755-9e87-4129-92f7-f525fbf333b4\") " pod="openshift-authentication/oauth-openshift-7fdb5b7d8f-lf5gd" Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.305712 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/2322c755-9e87-4129-92f7-f525fbf333b4-audit-policies\") pod \"oauth-openshift-7fdb5b7d8f-lf5gd\" (UID: \"2322c755-9e87-4129-92f7-f525fbf333b4\") " pod="openshift-authentication/oauth-openshift-7fdb5b7d8f-lf5gd" Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.306209 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-fz7gx"] Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.306244 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/2322c755-9e87-4129-92f7-f525fbf333b4-v4-0-config-system-session\") pod \"oauth-openshift-7fdb5b7d8f-lf5gd\" (UID: \"2322c755-9e87-4129-92f7-f525fbf333b4\") " pod="openshift-authentication/oauth-openshift-7fdb5b7d8f-lf5gd" Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.306347 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2322c755-9e87-4129-92f7-f525fbf333b4-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-7fdb5b7d8f-lf5gd\" (UID: \"2322c755-9e87-4129-92f7-f525fbf333b4\") " pod="openshift-authentication/oauth-openshift-7fdb5b7d8f-lf5gd" Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.306422 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/2322c755-9e87-4129-92f7-f525fbf333b4-v4-0-config-user-template-error\") pod \"oauth-openshift-7fdb5b7d8f-lf5gd\" (UID: \"2322c755-9e87-4129-92f7-f525fbf333b4\") " pod="openshift-authentication/oauth-openshift-7fdb5b7d8f-lf5gd" Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.306729 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/2322c755-9e87-4129-92f7-f525fbf333b4-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-7fdb5b7d8f-lf5gd\" (UID: \"2322c755-9e87-4129-92f7-f525fbf333b4\") " pod="openshift-authentication/oauth-openshift-7fdb5b7d8f-lf5gd" Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.307217 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/2322c755-9e87-4129-92f7-f525fbf333b4-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-7fdb5b7d8f-lf5gd\" (UID: \"2322c755-9e87-4129-92f7-f525fbf333b4\") " pod="openshift-authentication/oauth-openshift-7fdb5b7d8f-lf5gd" Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.307672 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/2322c755-9e87-4129-92f7-f525fbf333b4-v4-0-config-system-serving-cert\") pod \"oauth-openshift-7fdb5b7d8f-lf5gd\" (UID: \"2322c755-9e87-4129-92f7-f525fbf333b4\") " pod="openshift-authentication/oauth-openshift-7fdb5b7d8f-lf5gd" Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.307859 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/2322c755-9e87-4129-92f7-f525fbf333b4-v4-0-config-system-router-certs\") pod \"oauth-openshift-7fdb5b7d8f-lf5gd\" (UID: \"2322c755-9e87-4129-92f7-f525fbf333b4\") " pod="openshift-authentication/oauth-openshift-7fdb5b7d8f-lf5gd" Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.308616 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/2322c755-9e87-4129-92f7-f525fbf333b4-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-7fdb5b7d8f-lf5gd\" (UID: \"2322c755-9e87-4129-92f7-f525fbf333b4\") " pod="openshift-authentication/oauth-openshift-7fdb5b7d8f-lf5gd" Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.308816 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/2322c755-9e87-4129-92f7-f525fbf333b4-v4-0-config-user-template-login\") pod \"oauth-openshift-7fdb5b7d8f-lf5gd\" (UID: \"2322c755-9e87-4129-92f7-f525fbf333b4\") " pod="openshift-authentication/oauth-openshift-7fdb5b7d8f-lf5gd" Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.316562 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-52gcl\" (UniqueName: \"kubernetes.io/projected/2322c755-9e87-4129-92f7-f525fbf333b4-kube-api-access-52gcl\") pod \"oauth-openshift-7fdb5b7d8f-lf5gd\" (UID: \"2322c755-9e87-4129-92f7-f525fbf333b4\") " pod="openshift-authentication/oauth-openshift-7fdb5b7d8f-lf5gd" Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.448681 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-7fdb5b7d8f-lf5gd" Jan 21 11:14:36 crc kubenswrapper[4824]: I0121 11:14:36.772423 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-7fdb5b7d8f-lf5gd"] Jan 21 11:14:37 crc kubenswrapper[4824]: I0121 11:14:37.291082 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-7fdb5b7d8f-lf5gd" event={"ID":"2322c755-9e87-4129-92f7-f525fbf333b4","Type":"ContainerStarted","Data":"8f3d9e6c373bc746bb74551e87be2165a2396d15a6ad7f180620c55bda160f7e"} Jan 21 11:14:37 crc kubenswrapper[4824]: I0121 11:14:37.292185 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-7fdb5b7d8f-lf5gd" event={"ID":"2322c755-9e87-4129-92f7-f525fbf333b4","Type":"ContainerStarted","Data":"cd68f36f8eb5256d8f2edc35114238f1f3b7c75238f18b05388f0d06cac882df"} Jan 21 11:14:37 crc kubenswrapper[4824]: I0121 11:14:37.292465 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-7fdb5b7d8f-lf5gd" Jan 21 11:14:37 crc kubenswrapper[4824]: I0121 11:14:37.296225 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-7fdb5b7d8f-lf5gd" Jan 21 11:14:37 crc kubenswrapper[4824]: I0121 11:14:37.308157 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-7fdb5b7d8f-lf5gd" podStartSLOduration=27.308143671 podStartE2EDuration="27.308143671s" podCreationTimestamp="2026-01-21 11:14:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:14:37.306332502 +0000 UTC m=+219.599361814" watchObservedRunningTime="2026-01-21 11:14:37.308143671 +0000 UTC m=+219.601172963" Jan 21 11:14:38 crc kubenswrapper[4824]: I0121 11:14:38.053367 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="86b5dca2-989f-4b6b-983b-87763ce0d469" path="/var/lib/kubelet/pods/86b5dca2-989f-4b6b-983b-87763ce0d469/volumes" Jan 21 11:14:43 crc kubenswrapper[4824]: I0121 11:14:43.892564 4824 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Jan 21 11:14:43 crc kubenswrapper[4824]: I0121 11:14:43.894169 4824 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Jan 21 11:14:43 crc kubenswrapper[4824]: I0121 11:14:43.894314 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 21 11:14:43 crc kubenswrapper[4824]: I0121 11:14:43.894525 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" containerID="cri-o://a66b5c33884af00408dc35cf02f08ebd8570bd695f9f1fe558bf54fd06930431" gracePeriod=15 Jan 21 11:14:43 crc kubenswrapper[4824]: I0121 11:14:43.894685 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" containerID="cri-o://3a7023cfe4c098bcdd9c0477a8187f4ed7a0d2abdbe1e143436863948f441b4f" gracePeriod=15 Jan 21 11:14:43 crc kubenswrapper[4824]: I0121 11:14:43.894891 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" containerID="cri-o://48070c09fad511108ff22e44211cf26911509b8406d66948985cbfc3009d8233" gracePeriod=15 Jan 21 11:14:43 crc kubenswrapper[4824]: I0121 11:14:43.894899 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" containerID="cri-o://61b14b6846f94cdeba8f1f61b3754e7663041f92c70356c4ee6cbb988a2e680f" gracePeriod=15 Jan 21 11:14:43 crc kubenswrapper[4824]: I0121 11:14:43.894870 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" containerID="cri-o://d7514786602746dabfa690da48d48e233fe68d6dd6674aa5ecd489c0b3185f08" gracePeriod=15 Jan 21 11:14:43 crc kubenswrapper[4824]: I0121 11:14:43.894879 4824 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Jan 21 11:14:43 crc kubenswrapper[4824]: E0121 11:14:43.895215 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Jan 21 11:14:43 crc kubenswrapper[4824]: I0121 11:14:43.895230 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Jan 21 11:14:43 crc kubenswrapper[4824]: E0121 11:14:43.895241 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Jan 21 11:14:43 crc kubenswrapper[4824]: I0121 11:14:43.895246 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Jan 21 11:14:43 crc kubenswrapper[4824]: E0121 11:14:43.895257 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Jan 21 11:14:43 crc kubenswrapper[4824]: I0121 11:14:43.895262 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Jan 21 11:14:43 crc kubenswrapper[4824]: E0121 11:14:43.895271 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Jan 21 11:14:43 crc kubenswrapper[4824]: I0121 11:14:43.895276 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Jan 21 11:14:43 crc kubenswrapper[4824]: E0121 11:14:43.895300 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Jan 21 11:14:43 crc kubenswrapper[4824]: I0121 11:14:43.895306 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Jan 21 11:14:43 crc kubenswrapper[4824]: E0121 11:14:43.895313 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Jan 21 11:14:43 crc kubenswrapper[4824]: I0121 11:14:43.895319 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Jan 21 11:14:43 crc kubenswrapper[4824]: I0121 11:14:43.895409 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Jan 21 11:14:43 crc kubenswrapper[4824]: I0121 11:14:43.895419 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Jan 21 11:14:43 crc kubenswrapper[4824]: I0121 11:14:43.895426 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Jan 21 11:14:43 crc kubenswrapper[4824]: I0121 11:14:43.895436 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Jan 21 11:14:43 crc kubenswrapper[4824]: I0121 11:14:43.895459 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Jan 21 11:14:43 crc kubenswrapper[4824]: E0121 11:14:43.895555 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Jan 21 11:14:43 crc kubenswrapper[4824]: I0121 11:14:43.895563 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Jan 21 11:14:43 crc kubenswrapper[4824]: I0121 11:14:43.895652 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Jan 21 11:14:43 crc kubenswrapper[4824]: I0121 11:14:43.917930 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Jan 21 11:14:43 crc kubenswrapper[4824]: I0121 11:14:43.978494 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 21 11:14:43 crc kubenswrapper[4824]: I0121 11:14:43.978853 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 21 11:14:43 crc kubenswrapper[4824]: I0121 11:14:43.979024 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 21 11:14:43 crc kubenswrapper[4824]: I0121 11:14:43.979147 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 21 11:14:43 crc kubenswrapper[4824]: I0121 11:14:43.979264 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 21 11:14:43 crc kubenswrapper[4824]: I0121 11:14:43.979403 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 21 11:14:43 crc kubenswrapper[4824]: I0121 11:14:43.979563 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 21 11:14:43 crc kubenswrapper[4824]: I0121 11:14:43.979681 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 21 11:14:44 crc kubenswrapper[4824]: I0121 11:14:44.080520 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 21 11:14:44 crc kubenswrapper[4824]: I0121 11:14:44.080645 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 21 11:14:44 crc kubenswrapper[4824]: I0121 11:14:44.080748 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 21 11:14:44 crc kubenswrapper[4824]: I0121 11:14:44.080781 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 21 11:14:44 crc kubenswrapper[4824]: I0121 11:14:44.080806 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 21 11:14:44 crc kubenswrapper[4824]: I0121 11:14:44.080839 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 21 11:14:44 crc kubenswrapper[4824]: I0121 11:14:44.080865 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 21 11:14:44 crc kubenswrapper[4824]: I0121 11:14:44.080892 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 21 11:14:44 crc kubenswrapper[4824]: I0121 11:14:44.080902 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 21 11:14:44 crc kubenswrapper[4824]: I0121 11:14:44.080922 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 21 11:14:44 crc kubenswrapper[4824]: I0121 11:14:44.080924 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 21 11:14:44 crc kubenswrapper[4824]: I0121 11:14:44.080939 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 21 11:14:44 crc kubenswrapper[4824]: I0121 11:14:44.080944 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 21 11:14:44 crc kubenswrapper[4824]: I0121 11:14:44.080987 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 21 11:14:44 crc kubenswrapper[4824]: I0121 11:14:44.081051 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 21 11:14:44 crc kubenswrapper[4824]: I0121 11:14:44.081097 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 21 11:14:44 crc kubenswrapper[4824]: I0121 11:14:44.215816 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 21 11:14:44 crc kubenswrapper[4824]: E0121 11:14:44.230697 4824 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 192.168.26.120:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.188cbabf404def34 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-01-21 11:14:44.230131508 +0000 UTC m=+226.523160800,LastTimestamp:2026-01-21 11:14:44.230131508 +0000 UTC m=+226.523160800,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Jan 21 11:14:44 crc kubenswrapper[4824]: I0121 11:14:44.319669 4824 generic.go:334] "Generic (PLEG): container finished" podID="a83fe5bb-7790-45ea-8f0b-56799a5c4b89" containerID="8ea1c8d6b88fff12b2a363688bd7b4e3c72b3e77fe714e1ec1237fd44548b134" exitCode=0 Jan 21 11:14:44 crc kubenswrapper[4824]: I0121 11:14:44.319735 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"a83fe5bb-7790-45ea-8f0b-56799a5c4b89","Type":"ContainerDied","Data":"8ea1c8d6b88fff12b2a363688bd7b4e3c72b3e77fe714e1ec1237fd44548b134"} Jan 21 11:14:44 crc kubenswrapper[4824]: I0121 11:14:44.320261 4824 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 192.168.26.120:6443: connect: connection refused" Jan 21 11:14:44 crc kubenswrapper[4824]: I0121 11:14:44.320506 4824 status_manager.go:851] "Failed to get status for pod" podUID="a83fe5bb-7790-45ea-8f0b-56799a5c4b89" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 192.168.26.120:6443: connect: connection refused" Jan 21 11:14:44 crc kubenswrapper[4824]: I0121 11:14:44.321015 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"0dfee8c9edc60f4beca30cb6369cc26121c4a22045aa85ba62e4dcbcb9455144"} Jan 21 11:14:44 crc kubenswrapper[4824]: I0121 11:14:44.322790 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Jan 21 11:14:44 crc kubenswrapper[4824]: I0121 11:14:44.323631 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Jan 21 11:14:44 crc kubenswrapper[4824]: I0121 11:14:44.324104 4824 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="3a7023cfe4c098bcdd9c0477a8187f4ed7a0d2abdbe1e143436863948f441b4f" exitCode=0 Jan 21 11:14:44 crc kubenswrapper[4824]: I0121 11:14:44.324132 4824 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="48070c09fad511108ff22e44211cf26911509b8406d66948985cbfc3009d8233" exitCode=0 Jan 21 11:14:44 crc kubenswrapper[4824]: I0121 11:14:44.324140 4824 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="d7514786602746dabfa690da48d48e233fe68d6dd6674aa5ecd489c0b3185f08" exitCode=0 Jan 21 11:14:44 crc kubenswrapper[4824]: I0121 11:14:44.324147 4824 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="61b14b6846f94cdeba8f1f61b3754e7663041f92c70356c4ee6cbb988a2e680f" exitCode=2 Jan 21 11:14:44 crc kubenswrapper[4824]: I0121 11:14:44.324174 4824 scope.go:117] "RemoveContainer" containerID="af11ad95bcf0c6bbd5fa4ab99bb592ce1c87f40d29c5ade7832414315820a569" Jan 21 11:14:45 crc kubenswrapper[4824]: I0121 11:14:45.330618 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Jan 21 11:14:45 crc kubenswrapper[4824]: I0121 11:14:45.336070 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"95939db35012b5a5c83445539a3c3ba0348e91a0065268d0f4bf7dc607ee2022"} Jan 21 11:14:45 crc kubenswrapper[4824]: I0121 11:14:45.336589 4824 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 192.168.26.120:6443: connect: connection refused" Jan 21 11:14:45 crc kubenswrapper[4824]: I0121 11:14:45.336782 4824 status_manager.go:851] "Failed to get status for pod" podUID="a83fe5bb-7790-45ea-8f0b-56799a5c4b89" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 192.168.26.120:6443: connect: connection refused" Jan 21 11:14:45 crc kubenswrapper[4824]: I0121 11:14:45.507559 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Jan 21 11:14:45 crc kubenswrapper[4824]: I0121 11:14:45.508211 4824 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 192.168.26.120:6443: connect: connection refused" Jan 21 11:14:45 crc kubenswrapper[4824]: I0121 11:14:45.508509 4824 status_manager.go:851] "Failed to get status for pod" podUID="a83fe5bb-7790-45ea-8f0b-56799a5c4b89" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 192.168.26.120:6443: connect: connection refused" Jan 21 11:14:45 crc kubenswrapper[4824]: I0121 11:14:45.698256 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a83fe5bb-7790-45ea-8f0b-56799a5c4b89-kubelet-dir\") pod \"a83fe5bb-7790-45ea-8f0b-56799a5c4b89\" (UID: \"a83fe5bb-7790-45ea-8f0b-56799a5c4b89\") " Jan 21 11:14:45 crc kubenswrapper[4824]: I0121 11:14:45.698285 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/a83fe5bb-7790-45ea-8f0b-56799a5c4b89-var-lock\") pod \"a83fe5bb-7790-45ea-8f0b-56799a5c4b89\" (UID: \"a83fe5bb-7790-45ea-8f0b-56799a5c4b89\") " Jan 21 11:14:45 crc kubenswrapper[4824]: I0121 11:14:45.698317 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a83fe5bb-7790-45ea-8f0b-56799a5c4b89-kube-api-access\") pod \"a83fe5bb-7790-45ea-8f0b-56799a5c4b89\" (UID: \"a83fe5bb-7790-45ea-8f0b-56799a5c4b89\") " Jan 21 11:14:45 crc kubenswrapper[4824]: I0121 11:14:45.698381 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a83fe5bb-7790-45ea-8f0b-56799a5c4b89-var-lock" (OuterVolumeSpecName: "var-lock") pod "a83fe5bb-7790-45ea-8f0b-56799a5c4b89" (UID: "a83fe5bb-7790-45ea-8f0b-56799a5c4b89"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 21 11:14:45 crc kubenswrapper[4824]: I0121 11:14:45.698377 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a83fe5bb-7790-45ea-8f0b-56799a5c4b89-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "a83fe5bb-7790-45ea-8f0b-56799a5c4b89" (UID: "a83fe5bb-7790-45ea-8f0b-56799a5c4b89"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 21 11:14:45 crc kubenswrapper[4824]: I0121 11:14:45.698540 4824 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a83fe5bb-7790-45ea-8f0b-56799a5c4b89-kubelet-dir\") on node \"crc\" DevicePath \"\"" Jan 21 11:14:45 crc kubenswrapper[4824]: I0121 11:14:45.698559 4824 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/a83fe5bb-7790-45ea-8f0b-56799a5c4b89-var-lock\") on node \"crc\" DevicePath \"\"" Jan 21 11:14:45 crc kubenswrapper[4824]: I0121 11:14:45.702096 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a83fe5bb-7790-45ea-8f0b-56799a5c4b89-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "a83fe5bb-7790-45ea-8f0b-56799a5c4b89" (UID: "a83fe5bb-7790-45ea-8f0b-56799a5c4b89"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:14:45 crc kubenswrapper[4824]: I0121 11:14:45.800018 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a83fe5bb-7790-45ea-8f0b-56799a5c4b89-kube-api-access\") on node \"crc\" DevicePath \"\"" Jan 21 11:14:46 crc kubenswrapper[4824]: I0121 11:14:46.065641 4824 patch_prober.go:28] interesting pod/machine-config-daemon-zqjg9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 21 11:14:46 crc kubenswrapper[4824]: I0121 11:14:46.066027 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 21 11:14:46 crc kubenswrapper[4824]: I0121 11:14:46.066069 4824 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" Jan 21 11:14:46 crc kubenswrapper[4824]: I0121 11:14:46.066596 4824 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"2b1b37c6d2994021cfc30f85e92c40ae35677113b17aa066d52bc7a76ae5c153"} pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 21 11:14:46 crc kubenswrapper[4824]: I0121 11:14:46.066651 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" containerName="machine-config-daemon" containerID="cri-o://2b1b37c6d2994021cfc30f85e92c40ae35677113b17aa066d52bc7a76ae5c153" gracePeriod=600 Jan 21 11:14:46 crc kubenswrapper[4824]: I0121 11:14:46.166113 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Jan 21 11:14:46 crc kubenswrapper[4824]: I0121 11:14:46.166706 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 21 11:14:46 crc kubenswrapper[4824]: I0121 11:14:46.167049 4824 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 192.168.26.120:6443: connect: connection refused" Jan 21 11:14:46 crc kubenswrapper[4824]: I0121 11:14:46.167258 4824 status_manager.go:851] "Failed to get status for pod" podUID="a83fe5bb-7790-45ea-8f0b-56799a5c4b89" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 192.168.26.120:6443: connect: connection refused" Jan 21 11:14:46 crc kubenswrapper[4824]: I0121 11:14:46.167502 4824 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 192.168.26.120:6443: connect: connection refused" Jan 21 11:14:46 crc kubenswrapper[4824]: I0121 11:14:46.305382 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Jan 21 11:14:46 crc kubenswrapper[4824]: I0121 11:14:46.305499 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir" (OuterVolumeSpecName: "cert-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "cert-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 21 11:14:46 crc kubenswrapper[4824]: I0121 11:14:46.305615 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Jan 21 11:14:46 crc kubenswrapper[4824]: I0121 11:14:46.305632 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 21 11:14:46 crc kubenswrapper[4824]: I0121 11:14:46.305653 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Jan 21 11:14:46 crc kubenswrapper[4824]: I0121 11:14:46.305788 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 21 11:14:46 crc kubenswrapper[4824]: I0121 11:14:46.305855 4824 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") on node \"crc\" DevicePath \"\"" Jan 21 11:14:46 crc kubenswrapper[4824]: I0121 11:14:46.305866 4824 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") on node \"crc\" DevicePath \"\"" Jan 21 11:14:46 crc kubenswrapper[4824]: I0121 11:14:46.305873 4824 reconciler_common.go:293] "Volume detached for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") on node \"crc\" DevicePath \"\"" Jan 21 11:14:46 crc kubenswrapper[4824]: I0121 11:14:46.342277 4824 generic.go:334] "Generic (PLEG): container finished" podID="33f3d922-4ffe-409b-a49a-d88c85898260" containerID="2b1b37c6d2994021cfc30f85e92c40ae35677113b17aa066d52bc7a76ae5c153" exitCode=0 Jan 21 11:14:46 crc kubenswrapper[4824]: I0121 11:14:46.342338 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" event={"ID":"33f3d922-4ffe-409b-a49a-d88c85898260","Type":"ContainerDied","Data":"2b1b37c6d2994021cfc30f85e92c40ae35677113b17aa066d52bc7a76ae5c153"} Jan 21 11:14:46 crc kubenswrapper[4824]: I0121 11:14:46.342363 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" event={"ID":"33f3d922-4ffe-409b-a49a-d88c85898260","Type":"ContainerStarted","Data":"04e919af77a770ce30313c5a8770baf390f15a9d134ed04fa80518ff38656c8a"} Jan 21 11:14:46 crc kubenswrapper[4824]: I0121 11:14:46.343093 4824 status_manager.go:851] "Failed to get status for pod" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-machine-config-operator/pods/machine-config-daemon-zqjg9\": dial tcp 192.168.26.120:6443: connect: connection refused" Jan 21 11:14:46 crc kubenswrapper[4824]: I0121 11:14:46.343407 4824 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 192.168.26.120:6443: connect: connection refused" Jan 21 11:14:46 crc kubenswrapper[4824]: I0121 11:14:46.343611 4824 status_manager.go:851] "Failed to get status for pod" podUID="a83fe5bb-7790-45ea-8f0b-56799a5c4b89" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 192.168.26.120:6443: connect: connection refused" Jan 21 11:14:46 crc kubenswrapper[4824]: I0121 11:14:46.343798 4824 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 192.168.26.120:6443: connect: connection refused" Jan 21 11:14:46 crc kubenswrapper[4824]: I0121 11:14:46.344002 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"a83fe5bb-7790-45ea-8f0b-56799a5c4b89","Type":"ContainerDied","Data":"939da8a2d6bcb80f9bcbb9fa2a8ace318e8eb04c7ccffba9384e787401a5fb5e"} Jan 21 11:14:46 crc kubenswrapper[4824]: I0121 11:14:46.344025 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="939da8a2d6bcb80f9bcbb9fa2a8ace318e8eb04c7ccffba9384e787401a5fb5e" Jan 21 11:14:46 crc kubenswrapper[4824]: I0121 11:14:46.344057 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Jan 21 11:14:46 crc kubenswrapper[4824]: I0121 11:14:46.347085 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Jan 21 11:14:46 crc kubenswrapper[4824]: I0121 11:14:46.347272 4824 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 192.168.26.120:6443: connect: connection refused" Jan 21 11:14:46 crc kubenswrapper[4824]: I0121 11:14:46.347594 4824 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="a66b5c33884af00408dc35cf02f08ebd8570bd695f9f1fe558bf54fd06930431" exitCode=0 Jan 21 11:14:46 crc kubenswrapper[4824]: I0121 11:14:46.347633 4824 scope.go:117] "RemoveContainer" containerID="3a7023cfe4c098bcdd9c0477a8187f4ed7a0d2abdbe1e143436863948f441b4f" Jan 21 11:14:46 crc kubenswrapper[4824]: I0121 11:14:46.347581 4824 status_manager.go:851] "Failed to get status for pod" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-machine-config-operator/pods/machine-config-daemon-zqjg9\": dial tcp 192.168.26.120:6443: connect: connection refused" Jan 21 11:14:46 crc kubenswrapper[4824]: I0121 11:14:46.347659 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 21 11:14:46 crc kubenswrapper[4824]: I0121 11:14:46.348046 4824 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 192.168.26.120:6443: connect: connection refused" Jan 21 11:14:46 crc kubenswrapper[4824]: I0121 11:14:46.349166 4824 status_manager.go:851] "Failed to get status for pod" podUID="a83fe5bb-7790-45ea-8f0b-56799a5c4b89" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 192.168.26.120:6443: connect: connection refused" Jan 21 11:14:46 crc kubenswrapper[4824]: I0121 11:14:46.357462 4824 status_manager.go:851] "Failed to get status for pod" podUID="a83fe5bb-7790-45ea-8f0b-56799a5c4b89" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 192.168.26.120:6443: connect: connection refused" Jan 21 11:14:46 crc kubenswrapper[4824]: I0121 11:14:46.357746 4824 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 192.168.26.120:6443: connect: connection refused" Jan 21 11:14:46 crc kubenswrapper[4824]: I0121 11:14:46.358121 4824 scope.go:117] "RemoveContainer" containerID="48070c09fad511108ff22e44211cf26911509b8406d66948985cbfc3009d8233" Jan 21 11:14:46 crc kubenswrapper[4824]: I0121 11:14:46.358129 4824 status_manager.go:851] "Failed to get status for pod" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-machine-config-operator/pods/machine-config-daemon-zqjg9\": dial tcp 192.168.26.120:6443: connect: connection refused" Jan 21 11:14:46 crc kubenswrapper[4824]: I0121 11:14:46.358306 4824 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 192.168.26.120:6443: connect: connection refused" Jan 21 11:14:46 crc kubenswrapper[4824]: I0121 11:14:46.367989 4824 scope.go:117] "RemoveContainer" containerID="d7514786602746dabfa690da48d48e233fe68d6dd6674aa5ecd489c0b3185f08" Jan 21 11:14:46 crc kubenswrapper[4824]: I0121 11:14:46.393224 4824 scope.go:117] "RemoveContainer" containerID="61b14b6846f94cdeba8f1f61b3754e7663041f92c70356c4ee6cbb988a2e680f" Jan 21 11:14:46 crc kubenswrapper[4824]: I0121 11:14:46.402052 4824 scope.go:117] "RemoveContainer" containerID="a66b5c33884af00408dc35cf02f08ebd8570bd695f9f1fe558bf54fd06930431" Jan 21 11:14:46 crc kubenswrapper[4824]: I0121 11:14:46.411972 4824 scope.go:117] "RemoveContainer" containerID="a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc" Jan 21 11:14:46 crc kubenswrapper[4824]: I0121 11:14:46.428380 4824 scope.go:117] "RemoveContainer" containerID="3a7023cfe4c098bcdd9c0477a8187f4ed7a0d2abdbe1e143436863948f441b4f" Jan 21 11:14:46 crc kubenswrapper[4824]: E0121 11:14:46.428900 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3a7023cfe4c098bcdd9c0477a8187f4ed7a0d2abdbe1e143436863948f441b4f\": container with ID starting with 3a7023cfe4c098bcdd9c0477a8187f4ed7a0d2abdbe1e143436863948f441b4f not found: ID does not exist" containerID="3a7023cfe4c098bcdd9c0477a8187f4ed7a0d2abdbe1e143436863948f441b4f" Jan 21 11:14:46 crc kubenswrapper[4824]: I0121 11:14:46.428931 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3a7023cfe4c098bcdd9c0477a8187f4ed7a0d2abdbe1e143436863948f441b4f"} err="failed to get container status \"3a7023cfe4c098bcdd9c0477a8187f4ed7a0d2abdbe1e143436863948f441b4f\": rpc error: code = NotFound desc = could not find container \"3a7023cfe4c098bcdd9c0477a8187f4ed7a0d2abdbe1e143436863948f441b4f\": container with ID starting with 3a7023cfe4c098bcdd9c0477a8187f4ed7a0d2abdbe1e143436863948f441b4f not found: ID does not exist" Jan 21 11:14:46 crc kubenswrapper[4824]: I0121 11:14:46.428951 4824 scope.go:117] "RemoveContainer" containerID="48070c09fad511108ff22e44211cf26911509b8406d66948985cbfc3009d8233" Jan 21 11:14:46 crc kubenswrapper[4824]: E0121 11:14:46.429163 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"48070c09fad511108ff22e44211cf26911509b8406d66948985cbfc3009d8233\": container with ID starting with 48070c09fad511108ff22e44211cf26911509b8406d66948985cbfc3009d8233 not found: ID does not exist" containerID="48070c09fad511108ff22e44211cf26911509b8406d66948985cbfc3009d8233" Jan 21 11:14:46 crc kubenswrapper[4824]: I0121 11:14:46.429195 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"48070c09fad511108ff22e44211cf26911509b8406d66948985cbfc3009d8233"} err="failed to get container status \"48070c09fad511108ff22e44211cf26911509b8406d66948985cbfc3009d8233\": rpc error: code = NotFound desc = could not find container \"48070c09fad511108ff22e44211cf26911509b8406d66948985cbfc3009d8233\": container with ID starting with 48070c09fad511108ff22e44211cf26911509b8406d66948985cbfc3009d8233 not found: ID does not exist" Jan 21 11:14:46 crc kubenswrapper[4824]: I0121 11:14:46.429207 4824 scope.go:117] "RemoveContainer" containerID="d7514786602746dabfa690da48d48e233fe68d6dd6674aa5ecd489c0b3185f08" Jan 21 11:14:46 crc kubenswrapper[4824]: E0121 11:14:46.429422 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d7514786602746dabfa690da48d48e233fe68d6dd6674aa5ecd489c0b3185f08\": container with ID starting with d7514786602746dabfa690da48d48e233fe68d6dd6674aa5ecd489c0b3185f08 not found: ID does not exist" containerID="d7514786602746dabfa690da48d48e233fe68d6dd6674aa5ecd489c0b3185f08" Jan 21 11:14:46 crc kubenswrapper[4824]: I0121 11:14:46.429464 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d7514786602746dabfa690da48d48e233fe68d6dd6674aa5ecd489c0b3185f08"} err="failed to get container status \"d7514786602746dabfa690da48d48e233fe68d6dd6674aa5ecd489c0b3185f08\": rpc error: code = NotFound desc = could not find container \"d7514786602746dabfa690da48d48e233fe68d6dd6674aa5ecd489c0b3185f08\": container with ID starting with d7514786602746dabfa690da48d48e233fe68d6dd6674aa5ecd489c0b3185f08 not found: ID does not exist" Jan 21 11:14:46 crc kubenswrapper[4824]: I0121 11:14:46.429478 4824 scope.go:117] "RemoveContainer" containerID="61b14b6846f94cdeba8f1f61b3754e7663041f92c70356c4ee6cbb988a2e680f" Jan 21 11:14:46 crc kubenswrapper[4824]: E0121 11:14:46.429641 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"61b14b6846f94cdeba8f1f61b3754e7663041f92c70356c4ee6cbb988a2e680f\": container with ID starting with 61b14b6846f94cdeba8f1f61b3754e7663041f92c70356c4ee6cbb988a2e680f not found: ID does not exist" containerID="61b14b6846f94cdeba8f1f61b3754e7663041f92c70356c4ee6cbb988a2e680f" Jan 21 11:14:46 crc kubenswrapper[4824]: I0121 11:14:46.429660 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"61b14b6846f94cdeba8f1f61b3754e7663041f92c70356c4ee6cbb988a2e680f"} err="failed to get container status \"61b14b6846f94cdeba8f1f61b3754e7663041f92c70356c4ee6cbb988a2e680f\": rpc error: code = NotFound desc = could not find container \"61b14b6846f94cdeba8f1f61b3754e7663041f92c70356c4ee6cbb988a2e680f\": container with ID starting with 61b14b6846f94cdeba8f1f61b3754e7663041f92c70356c4ee6cbb988a2e680f not found: ID does not exist" Jan 21 11:14:46 crc kubenswrapper[4824]: I0121 11:14:46.429671 4824 scope.go:117] "RemoveContainer" containerID="a66b5c33884af00408dc35cf02f08ebd8570bd695f9f1fe558bf54fd06930431" Jan 21 11:14:46 crc kubenswrapper[4824]: E0121 11:14:46.429842 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a66b5c33884af00408dc35cf02f08ebd8570bd695f9f1fe558bf54fd06930431\": container with ID starting with a66b5c33884af00408dc35cf02f08ebd8570bd695f9f1fe558bf54fd06930431 not found: ID does not exist" containerID="a66b5c33884af00408dc35cf02f08ebd8570bd695f9f1fe558bf54fd06930431" Jan 21 11:14:46 crc kubenswrapper[4824]: I0121 11:14:46.429862 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a66b5c33884af00408dc35cf02f08ebd8570bd695f9f1fe558bf54fd06930431"} err="failed to get container status \"a66b5c33884af00408dc35cf02f08ebd8570bd695f9f1fe558bf54fd06930431\": rpc error: code = NotFound desc = could not find container \"a66b5c33884af00408dc35cf02f08ebd8570bd695f9f1fe558bf54fd06930431\": container with ID starting with a66b5c33884af00408dc35cf02f08ebd8570bd695f9f1fe558bf54fd06930431 not found: ID does not exist" Jan 21 11:14:46 crc kubenswrapper[4824]: I0121 11:14:46.429873 4824 scope.go:117] "RemoveContainer" containerID="a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc" Jan 21 11:14:46 crc kubenswrapper[4824]: E0121 11:14:46.430043 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc\": container with ID starting with a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc not found: ID does not exist" containerID="a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc" Jan 21 11:14:46 crc kubenswrapper[4824]: I0121 11:14:46.430061 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc"} err="failed to get container status \"a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc\": rpc error: code = NotFound desc = could not find container \"a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc\": container with ID starting with a1e66e91716e648bce76f58c8ad41bbfd0b4ada77b186fd086c6a8b280adacbc not found: ID does not exist" Jan 21 11:14:46 crc kubenswrapper[4824]: E0121 11:14:46.999236 4824 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 192.168.26.120:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.188cbabf404def34 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-01-21 11:14:44.230131508 +0000 UTC m=+226.523160800,LastTimestamp:2026-01-21 11:14:44.230131508 +0000 UTC m=+226.523160800,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Jan 21 11:14:48 crc kubenswrapper[4824]: I0121 11:14:48.056550 4824 status_manager.go:851] "Failed to get status for pod" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-machine-config-operator/pods/machine-config-daemon-zqjg9\": dial tcp 192.168.26.120:6443: connect: connection refused" Jan 21 11:14:48 crc kubenswrapper[4824]: I0121 11:14:48.057002 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4b27818a5e8e43d0dc095d08835c792" path="/var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/volumes" Jan 21 11:14:48 crc kubenswrapper[4824]: I0121 11:14:48.057140 4824 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 192.168.26.120:6443: connect: connection refused" Jan 21 11:14:48 crc kubenswrapper[4824]: I0121 11:14:48.057416 4824 status_manager.go:851] "Failed to get status for pod" podUID="a83fe5bb-7790-45ea-8f0b-56799a5c4b89" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 192.168.26.120:6443: connect: connection refused" Jan 21 11:14:48 crc kubenswrapper[4824]: I0121 11:14:48.057723 4824 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 192.168.26.120:6443: connect: connection refused" Jan 21 11:14:51 crc kubenswrapper[4824]: E0121 11:14:51.552239 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:14:51Z\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:14:51Z\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:14:51Z\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-21T11:14:51Z\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Patch \"https://api-int.crc.testing:6443/api/v1/nodes/crc/status?timeout=10s\": dial tcp 192.168.26.120:6443: connect: connection refused" Jan 21 11:14:51 crc kubenswrapper[4824]: E0121 11:14:51.553320 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 192.168.26.120:6443: connect: connection refused" Jan 21 11:14:51 crc kubenswrapper[4824]: E0121 11:14:51.553602 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 192.168.26.120:6443: connect: connection refused" Jan 21 11:14:51 crc kubenswrapper[4824]: E0121 11:14:51.553797 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 192.168.26.120:6443: connect: connection refused" Jan 21 11:14:51 crc kubenswrapper[4824]: E0121 11:14:51.554135 4824 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 192.168.26.120:6443: connect: connection refused" Jan 21 11:14:51 crc kubenswrapper[4824]: E0121 11:14:51.554158 4824 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Jan 21 11:14:52 crc kubenswrapper[4824]: E0121 11:14:52.285305 4824 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 192.168.26.120:6443: connect: connection refused" Jan 21 11:14:52 crc kubenswrapper[4824]: E0121 11:14:52.285662 4824 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 192.168.26.120:6443: connect: connection refused" Jan 21 11:14:52 crc kubenswrapper[4824]: E0121 11:14:52.285870 4824 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 192.168.26.120:6443: connect: connection refused" Jan 21 11:14:52 crc kubenswrapper[4824]: E0121 11:14:52.286102 4824 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 192.168.26.120:6443: connect: connection refused" Jan 21 11:14:52 crc kubenswrapper[4824]: E0121 11:14:52.286349 4824 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 192.168.26.120:6443: connect: connection refused" Jan 21 11:14:52 crc kubenswrapper[4824]: I0121 11:14:52.286377 4824 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" Jan 21 11:14:52 crc kubenswrapper[4824]: E0121 11:14:52.286606 4824 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 192.168.26.120:6443: connect: connection refused" interval="200ms" Jan 21 11:14:52 crc kubenswrapper[4824]: E0121 11:14:52.487635 4824 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 192.168.26.120:6443: connect: connection refused" interval="400ms" Jan 21 11:14:52 crc kubenswrapper[4824]: E0121 11:14:52.889042 4824 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 192.168.26.120:6443: connect: connection refused" interval="800ms" Jan 21 11:14:53 crc kubenswrapper[4824]: E0121 11:14:53.690217 4824 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 192.168.26.120:6443: connect: connection refused" interval="1.6s" Jan 21 11:14:54 crc kubenswrapper[4824]: E0121 11:14:54.084837 4824 desired_state_of_world_populator.go:312] "Error processing volume" err="error processing PVC openshift-image-registry/crc-image-registry-storage: failed to fetch PVC from API server: Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-image-registry/persistentvolumeclaims/crc-image-registry-storage\": dial tcp 192.168.26.120:6443: connect: connection refused" pod="openshift-image-registry/image-registry-697d97f7c8-wxlxx" volumeName="registry-storage" Jan 21 11:14:55 crc kubenswrapper[4824]: I0121 11:14:55.048671 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 21 11:14:55 crc kubenswrapper[4824]: I0121 11:14:55.049426 4824 status_manager.go:851] "Failed to get status for pod" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-machine-config-operator/pods/machine-config-daemon-zqjg9\": dial tcp 192.168.26.120:6443: connect: connection refused" Jan 21 11:14:55 crc kubenswrapper[4824]: I0121 11:14:55.049823 4824 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 192.168.26.120:6443: connect: connection refused" Jan 21 11:14:55 crc kubenswrapper[4824]: I0121 11:14:55.050066 4824 status_manager.go:851] "Failed to get status for pod" podUID="a83fe5bb-7790-45ea-8f0b-56799a5c4b89" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 192.168.26.120:6443: connect: connection refused" Jan 21 11:14:55 crc kubenswrapper[4824]: I0121 11:14:55.059477 4824 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="6d31ab61-f94c-4f49-85d0-c9e91b83069b" Jan 21 11:14:55 crc kubenswrapper[4824]: I0121 11:14:55.059503 4824 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="6d31ab61-f94c-4f49-85d0-c9e91b83069b" Jan 21 11:14:55 crc kubenswrapper[4824]: E0121 11:14:55.059763 4824 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 192.168.26.120:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 21 11:14:55 crc kubenswrapper[4824]: I0121 11:14:55.060137 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 21 11:14:55 crc kubenswrapper[4824]: W0121 11:14:55.074402 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod71bb4a3aecc4ba5b26c4b7318770ce13.slice/crio-25e32e75e80b54c012f4f719d3849a2eed22b09fc73b90c2e946e0de64b5a368 WatchSource:0}: Error finding container 25e32e75e80b54c012f4f719d3849a2eed22b09fc73b90c2e946e0de64b5a368: Status 404 returned error can't find the container with id 25e32e75e80b54c012f4f719d3849a2eed22b09fc73b90c2e946e0de64b5a368 Jan 21 11:14:55 crc kubenswrapper[4824]: E0121 11:14:55.290822 4824 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 192.168.26.120:6443: connect: connection refused" interval="3.2s" Jan 21 11:14:55 crc kubenswrapper[4824]: I0121 11:14:55.381230 4824 generic.go:334] "Generic (PLEG): container finished" podID="71bb4a3aecc4ba5b26c4b7318770ce13" containerID="39ba2e9f1acb7a8d89c60567635c87d812b096ba076466971d576e55c18f1b39" exitCode=0 Jan 21 11:14:55 crc kubenswrapper[4824]: I0121 11:14:55.381270 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerDied","Data":"39ba2e9f1acb7a8d89c60567635c87d812b096ba076466971d576e55c18f1b39"} Jan 21 11:14:55 crc kubenswrapper[4824]: I0121 11:14:55.381292 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"25e32e75e80b54c012f4f719d3849a2eed22b09fc73b90c2e946e0de64b5a368"} Jan 21 11:14:55 crc kubenswrapper[4824]: I0121 11:14:55.381507 4824 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="6d31ab61-f94c-4f49-85d0-c9e91b83069b" Jan 21 11:14:55 crc kubenswrapper[4824]: I0121 11:14:55.381524 4824 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="6d31ab61-f94c-4f49-85d0-c9e91b83069b" Jan 21 11:14:55 crc kubenswrapper[4824]: I0121 11:14:55.381764 4824 status_manager.go:851] "Failed to get status for pod" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-machine-config-operator/pods/machine-config-daemon-zqjg9\": dial tcp 192.168.26.120:6443: connect: connection refused" Jan 21 11:14:55 crc kubenswrapper[4824]: E0121 11:14:55.381762 4824 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 192.168.26.120:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 21 11:14:55 crc kubenswrapper[4824]: I0121 11:14:55.381999 4824 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 192.168.26.120:6443: connect: connection refused" Jan 21 11:14:55 crc kubenswrapper[4824]: I0121 11:14:55.382218 4824 status_manager.go:851] "Failed to get status for pod" podUID="a83fe5bb-7790-45ea-8f0b-56799a5c4b89" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 192.168.26.120:6443: connect: connection refused" Jan 21 11:14:56 crc kubenswrapper[4824]: I0121 11:14:56.394182 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"a593ed3c66b1442fd2eeb7e1fbafe41c24a6922b534938a59f914e5e6f561f51"} Jan 21 11:14:56 crc kubenswrapper[4824]: I0121 11:14:56.394473 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"8e1c447530e12d995e9dfbdd4718c96d63cb8891dab598bfa0d0879bc4cf7f00"} Jan 21 11:14:56 crc kubenswrapper[4824]: I0121 11:14:56.394484 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"b778529cbbba0136d4fe6b7f1c03d9e3b7f49b108c456ff9a9c6ec582a7276a7"} Jan 21 11:14:56 crc kubenswrapper[4824]: I0121 11:14:56.394493 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"3cffa5e9fd6593f2c2eda3c289b3e16ac1a3866c802c737d57d64747711588f8"} Jan 21 11:14:56 crc kubenswrapper[4824]: I0121 11:14:56.394503 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"b804f5b3c0033796f7109c566616c8c8430ccb2d6d9f5f537dca260fca6f87cb"} Jan 21 11:14:56 crc kubenswrapper[4824]: I0121 11:14:56.394675 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 21 11:14:56 crc kubenswrapper[4824]: I0121 11:14:56.394744 4824 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="6d31ab61-f94c-4f49-85d0-c9e91b83069b" Jan 21 11:14:56 crc kubenswrapper[4824]: I0121 11:14:56.394758 4824 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="6d31ab61-f94c-4f49-85d0-c9e91b83069b" Jan 21 11:14:58 crc kubenswrapper[4824]: I0121 11:14:58.404543 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Jan 21 11:14:58 crc kubenswrapper[4824]: I0121 11:14:58.404756 4824 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="43529d1b9e1e5e22b0d56b023946119da8aacbae54b68c26ff0a1a82dc890c70" exitCode=1 Jan 21 11:14:58 crc kubenswrapper[4824]: I0121 11:14:58.404782 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"43529d1b9e1e5e22b0d56b023946119da8aacbae54b68c26ff0a1a82dc890c70"} Jan 21 11:14:58 crc kubenswrapper[4824]: I0121 11:14:58.405089 4824 scope.go:117] "RemoveContainer" containerID="43529d1b9e1e5e22b0d56b023946119da8aacbae54b68c26ff0a1a82dc890c70" Jan 21 11:14:59 crc kubenswrapper[4824]: I0121 11:14:59.412547 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Jan 21 11:14:59 crc kubenswrapper[4824]: I0121 11:14:59.412802 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"fead566ce315e0429506998b40292a6c707d4e650412b0c7978957b011746333"} Jan 21 11:15:00 crc kubenswrapper[4824]: I0121 11:15:00.061004 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 21 11:15:00 crc kubenswrapper[4824]: I0121 11:15:00.061035 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 21 11:15:00 crc kubenswrapper[4824]: I0121 11:15:00.065057 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 21 11:15:00 crc kubenswrapper[4824]: I0121 11:15:00.860275 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 21 11:15:00 crc kubenswrapper[4824]: I0121 11:15:00.863780 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 21 11:15:01 crc kubenswrapper[4824]: I0121 11:15:01.419933 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 21 11:15:01 crc kubenswrapper[4824]: I0121 11:15:01.550292 4824 kubelet.go:1914] "Deleted mirror pod because it is outdated" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 21 11:15:01 crc kubenswrapper[4824]: I0121 11:15:01.598296 4824 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="3fef5d00-9d10-4b25-bca4-6482ab4c492f" Jan 21 11:15:02 crc kubenswrapper[4824]: I0121 11:15:02.423028 4824 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="6d31ab61-f94c-4f49-85d0-c9e91b83069b" Jan 21 11:15:02 crc kubenswrapper[4824]: I0121 11:15:02.423054 4824 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="6d31ab61-f94c-4f49-85d0-c9e91b83069b" Jan 21 11:15:02 crc kubenswrapper[4824]: I0121 11:15:02.425063 4824 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="3fef5d00-9d10-4b25-bca4-6482ab4c492f" Jan 21 11:15:08 crc kubenswrapper[4824]: I0121 11:15:08.424148 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Jan 21 11:15:08 crc kubenswrapper[4824]: I0121 11:15:08.861616 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Jan 21 11:15:09 crc kubenswrapper[4824]: I0121 11:15:09.088353 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Jan 21 11:15:09 crc kubenswrapper[4824]: I0121 11:15:09.110353 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Jan 21 11:15:09 crc kubenswrapper[4824]: I0121 11:15:09.271289 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Jan 21 11:15:09 crc kubenswrapper[4824]: I0121 11:15:09.296717 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Jan 21 11:15:09 crc kubenswrapper[4824]: I0121 11:15:09.332647 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Jan 21 11:15:09 crc kubenswrapper[4824]: I0121 11:15:09.383431 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Jan 21 11:15:09 crc kubenswrapper[4824]: I0121 11:15:09.421771 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Jan 21 11:15:09 crc kubenswrapper[4824]: I0121 11:15:09.517897 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Jan 21 11:15:09 crc kubenswrapper[4824]: I0121 11:15:09.754038 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Jan 21 11:15:09 crc kubenswrapper[4824]: I0121 11:15:09.788995 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Jan 21 11:15:09 crc kubenswrapper[4824]: I0121 11:15:09.826785 4824 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Jan 21 11:15:09 crc kubenswrapper[4824]: I0121 11:15:09.862917 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Jan 21 11:15:10 crc kubenswrapper[4824]: I0121 11:15:10.206885 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Jan 21 11:15:10 crc kubenswrapper[4824]: I0121 11:15:10.296676 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Jan 21 11:15:10 crc kubenswrapper[4824]: I0121 11:15:10.706471 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Jan 21 11:15:10 crc kubenswrapper[4824]: I0121 11:15:10.846383 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Jan 21 11:15:10 crc kubenswrapper[4824]: I0121 11:15:10.987211 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Jan 21 11:15:11 crc kubenswrapper[4824]: I0121 11:15:11.203030 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Jan 21 11:15:11 crc kubenswrapper[4824]: I0121 11:15:11.441187 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Jan 21 11:15:11 crc kubenswrapper[4824]: I0121 11:15:11.496752 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Jan 21 11:15:11 crc kubenswrapper[4824]: I0121 11:15:11.968442 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Jan 21 11:15:12 crc kubenswrapper[4824]: I0121 11:15:12.570555 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Jan 21 11:15:12 crc kubenswrapper[4824]: I0121 11:15:12.635523 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Jan 21 11:15:13 crc kubenswrapper[4824]: I0121 11:15:13.142930 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Jan 21 11:15:13 crc kubenswrapper[4824]: I0121 11:15:13.146029 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Jan 21 11:15:13 crc kubenswrapper[4824]: I0121 11:15:13.204407 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Jan 21 11:15:13 crc kubenswrapper[4824]: I0121 11:15:13.402258 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Jan 21 11:15:13 crc kubenswrapper[4824]: I0121 11:15:13.471537 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Jan 21 11:15:13 crc kubenswrapper[4824]: I0121 11:15:13.538415 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Jan 21 11:15:13 crc kubenswrapper[4824]: I0121 11:15:13.698434 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 21 11:15:13 crc kubenswrapper[4824]: I0121 11:15:13.807767 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Jan 21 11:15:13 crc kubenswrapper[4824]: I0121 11:15:13.941377 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Jan 21 11:15:14 crc kubenswrapper[4824]: I0121 11:15:14.138039 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Jan 21 11:15:14 crc kubenswrapper[4824]: I0121 11:15:14.161256 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Jan 21 11:15:14 crc kubenswrapper[4824]: I0121 11:15:14.174452 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Jan 21 11:15:14 crc kubenswrapper[4824]: I0121 11:15:14.281221 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Jan 21 11:15:14 crc kubenswrapper[4824]: I0121 11:15:14.313582 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Jan 21 11:15:14 crc kubenswrapper[4824]: I0121 11:15:14.440326 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Jan 21 11:15:14 crc kubenswrapper[4824]: I0121 11:15:14.723166 4824 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Jan 21 11:15:15 crc kubenswrapper[4824]: I0121 11:15:15.405999 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Jan 21 11:15:15 crc kubenswrapper[4824]: I0121 11:15:15.722113 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Jan 21 11:15:15 crc kubenswrapper[4824]: I0121 11:15:15.754282 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Jan 21 11:15:16 crc kubenswrapper[4824]: I0121 11:15:16.063754 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Jan 21 11:15:16 crc kubenswrapper[4824]: I0121 11:15:16.114714 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Jan 21 11:15:16 crc kubenswrapper[4824]: I0121 11:15:16.194263 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Jan 21 11:15:16 crc kubenswrapper[4824]: I0121 11:15:16.270729 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Jan 21 11:15:16 crc kubenswrapper[4824]: I0121 11:15:16.288735 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Jan 21 11:15:16 crc kubenswrapper[4824]: I0121 11:15:16.412348 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Jan 21 11:15:16 crc kubenswrapper[4824]: I0121 11:15:16.636109 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Jan 21 11:15:16 crc kubenswrapper[4824]: I0121 11:15:16.682137 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Jan 21 11:15:16 crc kubenswrapper[4824]: I0121 11:15:16.743942 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Jan 21 11:15:16 crc kubenswrapper[4824]: I0121 11:15:16.832660 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Jan 21 11:15:16 crc kubenswrapper[4824]: I0121 11:15:16.833257 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Jan 21 11:15:17 crc kubenswrapper[4824]: I0121 11:15:17.003801 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Jan 21 11:15:17 crc kubenswrapper[4824]: I0121 11:15:17.006805 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Jan 21 11:15:17 crc kubenswrapper[4824]: I0121 11:15:17.061339 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Jan 21 11:15:17 crc kubenswrapper[4824]: I0121 11:15:17.165701 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Jan 21 11:15:17 crc kubenswrapper[4824]: I0121 11:15:17.292356 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Jan 21 11:15:17 crc kubenswrapper[4824]: I0121 11:15:17.335771 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Jan 21 11:15:17 crc kubenswrapper[4824]: I0121 11:15:17.372069 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Jan 21 11:15:17 crc kubenswrapper[4824]: I0121 11:15:17.399041 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Jan 21 11:15:17 crc kubenswrapper[4824]: I0121 11:15:17.466828 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Jan 21 11:15:17 crc kubenswrapper[4824]: I0121 11:15:17.627782 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Jan 21 11:15:17 crc kubenswrapper[4824]: I0121 11:15:17.635673 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Jan 21 11:15:17 crc kubenswrapper[4824]: I0121 11:15:17.723530 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Jan 21 11:15:17 crc kubenswrapper[4824]: I0121 11:15:17.733600 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Jan 21 11:15:17 crc kubenswrapper[4824]: I0121 11:15:17.860717 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Jan 21 11:15:17 crc kubenswrapper[4824]: I0121 11:15:17.952594 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Jan 21 11:15:18 crc kubenswrapper[4824]: I0121 11:15:18.141527 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Jan 21 11:15:18 crc kubenswrapper[4824]: I0121 11:15:18.283305 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Jan 21 11:15:18 crc kubenswrapper[4824]: I0121 11:15:18.366524 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Jan 21 11:15:18 crc kubenswrapper[4824]: I0121 11:15:18.369208 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Jan 21 11:15:18 crc kubenswrapper[4824]: I0121 11:15:18.429527 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Jan 21 11:15:18 crc kubenswrapper[4824]: I0121 11:15:18.490534 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Jan 21 11:15:18 crc kubenswrapper[4824]: I0121 11:15:18.575109 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Jan 21 11:15:18 crc kubenswrapper[4824]: I0121 11:15:18.576739 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Jan 21 11:15:18 crc kubenswrapper[4824]: I0121 11:15:18.701334 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Jan 21 11:15:18 crc kubenswrapper[4824]: I0121 11:15:18.715483 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Jan 21 11:15:18 crc kubenswrapper[4824]: I0121 11:15:18.741452 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Jan 21 11:15:18 crc kubenswrapper[4824]: I0121 11:15:18.765931 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Jan 21 11:15:18 crc kubenswrapper[4824]: I0121 11:15:18.796038 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Jan 21 11:15:18 crc kubenswrapper[4824]: I0121 11:15:18.836995 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Jan 21 11:15:18 crc kubenswrapper[4824]: I0121 11:15:18.927232 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Jan 21 11:15:19 crc kubenswrapper[4824]: I0121 11:15:19.068057 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Jan 21 11:15:19 crc kubenswrapper[4824]: I0121 11:15:19.111687 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Jan 21 11:15:19 crc kubenswrapper[4824]: I0121 11:15:19.195651 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Jan 21 11:15:19 crc kubenswrapper[4824]: I0121 11:15:19.277316 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Jan 21 11:15:19 crc kubenswrapper[4824]: I0121 11:15:19.279188 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Jan 21 11:15:19 crc kubenswrapper[4824]: I0121 11:15:19.294987 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Jan 21 11:15:19 crc kubenswrapper[4824]: I0121 11:15:19.301026 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Jan 21 11:15:19 crc kubenswrapper[4824]: I0121 11:15:19.357457 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Jan 21 11:15:19 crc kubenswrapper[4824]: I0121 11:15:19.400529 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Jan 21 11:15:19 crc kubenswrapper[4824]: I0121 11:15:19.417456 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Jan 21 11:15:19 crc kubenswrapper[4824]: I0121 11:15:19.548237 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Jan 21 11:15:19 crc kubenswrapper[4824]: I0121 11:15:19.604564 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Jan 21 11:15:19 crc kubenswrapper[4824]: I0121 11:15:19.729577 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Jan 21 11:15:19 crc kubenswrapper[4824]: I0121 11:15:19.767823 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Jan 21 11:15:19 crc kubenswrapper[4824]: I0121 11:15:19.882491 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Jan 21 11:15:19 crc kubenswrapper[4824]: I0121 11:15:19.907692 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Jan 21 11:15:19 crc kubenswrapper[4824]: I0121 11:15:19.949904 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Jan 21 11:15:19 crc kubenswrapper[4824]: I0121 11:15:19.962071 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Jan 21 11:15:20 crc kubenswrapper[4824]: I0121 11:15:20.029426 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Jan 21 11:15:20 crc kubenswrapper[4824]: I0121 11:15:20.068133 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Jan 21 11:15:20 crc kubenswrapper[4824]: I0121 11:15:20.075603 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Jan 21 11:15:20 crc kubenswrapper[4824]: I0121 11:15:20.093926 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Jan 21 11:15:20 crc kubenswrapper[4824]: I0121 11:15:20.133264 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Jan 21 11:15:20 crc kubenswrapper[4824]: I0121 11:15:20.178782 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Jan 21 11:15:20 crc kubenswrapper[4824]: I0121 11:15:20.270048 4824 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Jan 21 11:15:20 crc kubenswrapper[4824]: I0121 11:15:20.286240 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Jan 21 11:15:20 crc kubenswrapper[4824]: I0121 11:15:20.313346 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Jan 21 11:15:20 crc kubenswrapper[4824]: I0121 11:15:20.336663 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Jan 21 11:15:20 crc kubenswrapper[4824]: I0121 11:15:20.345823 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Jan 21 11:15:20 crc kubenswrapper[4824]: I0121 11:15:20.373000 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Jan 21 11:15:20 crc kubenswrapper[4824]: I0121 11:15:20.404996 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Jan 21 11:15:20 crc kubenswrapper[4824]: I0121 11:15:20.461016 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Jan 21 11:15:20 crc kubenswrapper[4824]: I0121 11:15:20.525112 4824 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Jan 21 11:15:20 crc kubenswrapper[4824]: I0121 11:15:20.526771 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Jan 21 11:15:20 crc kubenswrapper[4824]: I0121 11:15:20.526887 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Jan 21 11:15:20 crc kubenswrapper[4824]: I0121 11:15:20.671061 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Jan 21 11:15:20 crc kubenswrapper[4824]: I0121 11:15:20.714580 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Jan 21 11:15:20 crc kubenswrapper[4824]: I0121 11:15:20.813379 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Jan 21 11:15:20 crc kubenswrapper[4824]: I0121 11:15:20.933791 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Jan 21 11:15:20 crc kubenswrapper[4824]: I0121 11:15:20.967106 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.038570 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.101023 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.166745 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.271414 4824 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.272495 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podStartSLOduration=38.272480966 podStartE2EDuration="38.272480966s" podCreationTimestamp="2026-01-21 11:14:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:15:01.569012902 +0000 UTC m=+243.862042193" watchObservedRunningTime="2026-01-21 11:15:21.272480966 +0000 UTC m=+263.565510259" Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.274913 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.274952 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29483235-4zbf5","openshift-kube-apiserver/kube-apiserver-crc","openshift-marketplace/marketplace-operator-79b997595-xcxs9"] Jan 21 11:15:21 crc kubenswrapper[4824]: E0121 11:15:21.275159 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a83fe5bb-7790-45ea-8f0b-56799a5c4b89" containerName="installer" Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.275177 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="a83fe5bb-7790-45ea-8f0b-56799a5c4b89" containerName="installer" Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.275270 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="a83fe5bb-7790-45ea-8f0b-56799a5c4b89" containerName="installer" Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.275731 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29483235-4zbf5" Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.275925 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-gjhp5","openshift-marketplace/redhat-operators-hqs52","openshift-marketplace/community-operators-7q9f2","openshift-marketplace/redhat-marketplace-4w5m6","openshift-marketplace/certified-operators-b922x"] Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.276108 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-b922x" podUID="3f11911f-e365-4151-b536-3988186b8757" containerName="registry-server" containerID="cri-o://f4386ff52380b3c795405ecbb9387d4d9b5fca2cffa267eb0ddb32961b0a7f10" gracePeriod=30 Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.276114 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-xcxs9" Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.276280 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-hqs52" podUID="c3b56f2d-96c0-4597-9235-147fc4802d19" containerName="registry-server" containerID="cri-o://b1068da5b4766f19c34535e8e58a22325fb8a2d31ce06ee547c973539c887a47" gracePeriod=30 Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.276409 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-7q9f2" podUID="1b040d3d-f4a0-451a-97de-08bea41c6dce" containerName="registry-server" containerID="cri-o://728ea008f822360aaa1493d68beaf88c96b2cdf76a7eeb252c13acb6bf91c9d5" gracePeriod=30 Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.276481 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-gjhp5" podUID="14d99a1f-fca3-421a-8b47-ab4cda2f77fc" containerName="marketplace-operator" containerID="cri-o://605262189eb74a464efb5cb3518becda7f59cb81cb5a3ce736fbc1a29ea45dc4" gracePeriod=30 Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.276617 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-4w5m6" podUID="1bc8c652-fb4f-48b7-86ce-c45522dfd476" containerName="registry-server" containerID="cri-o://fa2ad6f2a50b9350a988947c908be8adb082c0d1e40507b63b02b300f1699aae" gracePeriod=30 Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.277698 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.278136 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.278820 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.278921 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.285645 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.294175 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=20.294162522 podStartE2EDuration="20.294162522s" podCreationTimestamp="2026-01-21 11:15:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:15:21.293228597 +0000 UTC m=+263.586257889" watchObservedRunningTime="2026-01-21 11:15:21.294162522 +0000 UTC m=+263.587191814" Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.345044 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.376091 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b4c29122-d9b2-4f85-8b29-c0e7d888667f-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-xcxs9\" (UID: \"b4c29122-d9b2-4f85-8b29-c0e7d888667f\") " pod="openshift-marketplace/marketplace-operator-79b997595-xcxs9" Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.376190 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/123c89e0-3113-4b28-9129-3848a7288052-config-volume\") pod \"collect-profiles-29483235-4zbf5\" (UID: \"123c89e0-3113-4b28-9129-3848a7288052\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29483235-4zbf5" Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.376265 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/123c89e0-3113-4b28-9129-3848a7288052-secret-volume\") pod \"collect-profiles-29483235-4zbf5\" (UID: \"123c89e0-3113-4b28-9129-3848a7288052\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29483235-4zbf5" Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.376282 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xbqmv\" (UniqueName: \"kubernetes.io/projected/b4c29122-d9b2-4f85-8b29-c0e7d888667f-kube-api-access-xbqmv\") pod \"marketplace-operator-79b997595-xcxs9\" (UID: \"b4c29122-d9b2-4f85-8b29-c0e7d888667f\") " pod="openshift-marketplace/marketplace-operator-79b997595-xcxs9" Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.376305 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fxbxn\" (UniqueName: \"kubernetes.io/projected/123c89e0-3113-4b28-9129-3848a7288052-kube-api-access-fxbxn\") pod \"collect-profiles-29483235-4zbf5\" (UID: \"123c89e0-3113-4b28-9129-3848a7288052\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29483235-4zbf5" Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.376321 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b4c29122-d9b2-4f85-8b29-c0e7d888667f-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-xcxs9\" (UID: \"b4c29122-d9b2-4f85-8b29-c0e7d888667f\") " pod="openshift-marketplace/marketplace-operator-79b997595-xcxs9" Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.386204 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.477128 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/123c89e0-3113-4b28-9129-3848a7288052-secret-volume\") pod \"collect-profiles-29483235-4zbf5\" (UID: \"123c89e0-3113-4b28-9129-3848a7288052\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29483235-4zbf5" Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.477189 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xbqmv\" (UniqueName: \"kubernetes.io/projected/b4c29122-d9b2-4f85-8b29-c0e7d888667f-kube-api-access-xbqmv\") pod \"marketplace-operator-79b997595-xcxs9\" (UID: \"b4c29122-d9b2-4f85-8b29-c0e7d888667f\") " pod="openshift-marketplace/marketplace-operator-79b997595-xcxs9" Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.477223 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fxbxn\" (UniqueName: \"kubernetes.io/projected/123c89e0-3113-4b28-9129-3848a7288052-kube-api-access-fxbxn\") pod \"collect-profiles-29483235-4zbf5\" (UID: \"123c89e0-3113-4b28-9129-3848a7288052\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29483235-4zbf5" Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.477242 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b4c29122-d9b2-4f85-8b29-c0e7d888667f-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-xcxs9\" (UID: \"b4c29122-d9b2-4f85-8b29-c0e7d888667f\") " pod="openshift-marketplace/marketplace-operator-79b997595-xcxs9" Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.477256 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b4c29122-d9b2-4f85-8b29-c0e7d888667f-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-xcxs9\" (UID: \"b4c29122-d9b2-4f85-8b29-c0e7d888667f\") " pod="openshift-marketplace/marketplace-operator-79b997595-xcxs9" Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.477289 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/123c89e0-3113-4b28-9129-3848a7288052-config-volume\") pod \"collect-profiles-29483235-4zbf5\" (UID: \"123c89e0-3113-4b28-9129-3848a7288052\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29483235-4zbf5" Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.478328 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b4c29122-d9b2-4f85-8b29-c0e7d888667f-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-xcxs9\" (UID: \"b4c29122-d9b2-4f85-8b29-c0e7d888667f\") " pod="openshift-marketplace/marketplace-operator-79b997595-xcxs9" Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.478710 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/123c89e0-3113-4b28-9129-3848a7288052-config-volume\") pod \"collect-profiles-29483235-4zbf5\" (UID: \"123c89e0-3113-4b28-9129-3848a7288052\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29483235-4zbf5" Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.482400 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b4c29122-d9b2-4f85-8b29-c0e7d888667f-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-xcxs9\" (UID: \"b4c29122-d9b2-4f85-8b29-c0e7d888667f\") " pod="openshift-marketplace/marketplace-operator-79b997595-xcxs9" Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.482418 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/123c89e0-3113-4b28-9129-3848a7288052-secret-volume\") pod \"collect-profiles-29483235-4zbf5\" (UID: \"123c89e0-3113-4b28-9129-3848a7288052\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29483235-4zbf5" Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.490400 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xbqmv\" (UniqueName: \"kubernetes.io/projected/b4c29122-d9b2-4f85-8b29-c0e7d888667f-kube-api-access-xbqmv\") pod \"marketplace-operator-79b997595-xcxs9\" (UID: \"b4c29122-d9b2-4f85-8b29-c0e7d888667f\") " pod="openshift-marketplace/marketplace-operator-79b997595-xcxs9" Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.494263 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fxbxn\" (UniqueName: \"kubernetes.io/projected/123c89e0-3113-4b28-9129-3848a7288052-kube-api-access-fxbxn\") pod \"collect-profiles-29483235-4zbf5\" (UID: \"123c89e0-3113-4b28-9129-3848a7288052\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29483235-4zbf5" Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.495979 4824 generic.go:334] "Generic (PLEG): container finished" podID="c3b56f2d-96c0-4597-9235-147fc4802d19" containerID="b1068da5b4766f19c34535e8e58a22325fb8a2d31ce06ee547c973539c887a47" exitCode=0 Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.496040 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hqs52" event={"ID":"c3b56f2d-96c0-4597-9235-147fc4802d19","Type":"ContainerDied","Data":"b1068da5b4766f19c34535e8e58a22325fb8a2d31ce06ee547c973539c887a47"} Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.498180 4824 generic.go:334] "Generic (PLEG): container finished" podID="1bc8c652-fb4f-48b7-86ce-c45522dfd476" containerID="fa2ad6f2a50b9350a988947c908be8adb082c0d1e40507b63b02b300f1699aae" exitCode=0 Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.498246 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4w5m6" event={"ID":"1bc8c652-fb4f-48b7-86ce-c45522dfd476","Type":"ContainerDied","Data":"fa2ad6f2a50b9350a988947c908be8adb082c0d1e40507b63b02b300f1699aae"} Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.501816 4824 generic.go:334] "Generic (PLEG): container finished" podID="3f11911f-e365-4151-b536-3988186b8757" containerID="f4386ff52380b3c795405ecbb9387d4d9b5fca2cffa267eb0ddb32961b0a7f10" exitCode=0 Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.501872 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-b922x" event={"ID":"3f11911f-e365-4151-b536-3988186b8757","Type":"ContainerDied","Data":"f4386ff52380b3c795405ecbb9387d4d9b5fca2cffa267eb0ddb32961b0a7f10"} Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.503828 4824 generic.go:334] "Generic (PLEG): container finished" podID="1b040d3d-f4a0-451a-97de-08bea41c6dce" containerID="728ea008f822360aaa1493d68beaf88c96b2cdf76a7eeb252c13acb6bf91c9d5" exitCode=0 Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.503907 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7q9f2" event={"ID":"1b040d3d-f4a0-451a-97de-08bea41c6dce","Type":"ContainerDied","Data":"728ea008f822360aaa1493d68beaf88c96b2cdf76a7eeb252c13acb6bf91c9d5"} Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.505062 4824 generic.go:334] "Generic (PLEG): container finished" podID="14d99a1f-fca3-421a-8b47-ab4cda2f77fc" containerID="605262189eb74a464efb5cb3518becda7f59cb81cb5a3ce736fbc1a29ea45dc4" exitCode=0 Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.505800 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-gjhp5" event={"ID":"14d99a1f-fca3-421a-8b47-ab4cda2f77fc","Type":"ContainerDied","Data":"605262189eb74a464efb5cb3518becda7f59cb81cb5a3ce736fbc1a29ea45dc4"} Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.557613 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29483235-4zbf5"] Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.558229 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.560146 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-xcxs9"] Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.595291 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.597680 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29483235-4zbf5" Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.634450 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.680991 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-xcxs9" Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.707660 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-b922x" Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.708103 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7q9f2" Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.710728 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4w5m6" Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.715496 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hqs52" Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.718394 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-gjhp5" Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.756785 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.787081 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.821322 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.841520 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.881602 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kjjkx\" (UniqueName: \"kubernetes.io/projected/14d99a1f-fca3-421a-8b47-ab4cda2f77fc-kube-api-access-kjjkx\") pod \"14d99a1f-fca3-421a-8b47-ab4cda2f77fc\" (UID: \"14d99a1f-fca3-421a-8b47-ab4cda2f77fc\") " Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.881643 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/14d99a1f-fca3-421a-8b47-ab4cda2f77fc-marketplace-trusted-ca\") pod \"14d99a1f-fca3-421a-8b47-ab4cda2f77fc\" (UID: \"14d99a1f-fca3-421a-8b47-ab4cda2f77fc\") " Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.881667 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lchhw\" (UniqueName: \"kubernetes.io/projected/1bc8c652-fb4f-48b7-86ce-c45522dfd476-kube-api-access-lchhw\") pod \"1bc8c652-fb4f-48b7-86ce-c45522dfd476\" (UID: \"1bc8c652-fb4f-48b7-86ce-c45522dfd476\") " Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.881694 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3f11911f-e365-4151-b536-3988186b8757-utilities\") pod \"3f11911f-e365-4151-b536-3988186b8757\" (UID: \"3f11911f-e365-4151-b536-3988186b8757\") " Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.882255 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c3b56f2d-96c0-4597-9235-147fc4802d19-catalog-content\") pod \"c3b56f2d-96c0-4597-9235-147fc4802d19\" (UID: \"c3b56f2d-96c0-4597-9235-147fc4802d19\") " Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.882290 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3f11911f-e365-4151-b536-3988186b8757-catalog-content\") pod \"3f11911f-e365-4151-b536-3988186b8757\" (UID: \"3f11911f-e365-4151-b536-3988186b8757\") " Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.882317 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1b040d3d-f4a0-451a-97de-08bea41c6dce-utilities\") pod \"1b040d3d-f4a0-451a-97de-08bea41c6dce\" (UID: \"1b040d3d-f4a0-451a-97de-08bea41c6dce\") " Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.882342 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/14d99a1f-fca3-421a-8b47-ab4cda2f77fc-marketplace-operator-metrics\") pod \"14d99a1f-fca3-421a-8b47-ab4cda2f77fc\" (UID: \"14d99a1f-fca3-421a-8b47-ab4cda2f77fc\") " Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.882349 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/14d99a1f-fca3-421a-8b47-ab4cda2f77fc-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "14d99a1f-fca3-421a-8b47-ab4cda2f77fc" (UID: "14d99a1f-fca3-421a-8b47-ab4cda2f77fc"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.882409 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mxt8j\" (UniqueName: \"kubernetes.io/projected/1b040d3d-f4a0-451a-97de-08bea41c6dce-kube-api-access-mxt8j\") pod \"1b040d3d-f4a0-451a-97de-08bea41c6dce\" (UID: \"1b040d3d-f4a0-451a-97de-08bea41c6dce\") " Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.882429 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c3b56f2d-96c0-4597-9235-147fc4802d19-utilities\") pod \"c3b56f2d-96c0-4597-9235-147fc4802d19\" (UID: \"c3b56f2d-96c0-4597-9235-147fc4802d19\") " Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.882456 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8656b\" (UniqueName: \"kubernetes.io/projected/c3b56f2d-96c0-4597-9235-147fc4802d19-kube-api-access-8656b\") pod \"c3b56f2d-96c0-4597-9235-147fc4802d19\" (UID: \"c3b56f2d-96c0-4597-9235-147fc4802d19\") " Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.882490 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1bc8c652-fb4f-48b7-86ce-c45522dfd476-catalog-content\") pod \"1bc8c652-fb4f-48b7-86ce-c45522dfd476\" (UID: \"1bc8c652-fb4f-48b7-86ce-c45522dfd476\") " Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.882517 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1bc8c652-fb4f-48b7-86ce-c45522dfd476-utilities\") pod \"1bc8c652-fb4f-48b7-86ce-c45522dfd476\" (UID: \"1bc8c652-fb4f-48b7-86ce-c45522dfd476\") " Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.882545 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xjbt6\" (UniqueName: \"kubernetes.io/projected/3f11911f-e365-4151-b536-3988186b8757-kube-api-access-xjbt6\") pod \"3f11911f-e365-4151-b536-3988186b8757\" (UID: \"3f11911f-e365-4151-b536-3988186b8757\") " Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.882565 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1b040d3d-f4a0-451a-97de-08bea41c6dce-catalog-content\") pod \"1b040d3d-f4a0-451a-97de-08bea41c6dce\" (UID: \"1b040d3d-f4a0-451a-97de-08bea41c6dce\") " Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.882509 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3f11911f-e365-4151-b536-3988186b8757-utilities" (OuterVolumeSpecName: "utilities") pod "3f11911f-e365-4151-b536-3988186b8757" (UID: "3f11911f-e365-4151-b536-3988186b8757"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.882808 4824 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/14d99a1f-fca3-421a-8b47-ab4cda2f77fc-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.882823 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3f11911f-e365-4151-b536-3988186b8757-utilities\") on node \"crc\" DevicePath \"\"" Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.883199 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1bc8c652-fb4f-48b7-86ce-c45522dfd476-utilities" (OuterVolumeSpecName: "utilities") pod "1bc8c652-fb4f-48b7-86ce-c45522dfd476" (UID: "1bc8c652-fb4f-48b7-86ce-c45522dfd476"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.883259 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c3b56f2d-96c0-4597-9235-147fc4802d19-utilities" (OuterVolumeSpecName: "utilities") pod "c3b56f2d-96c0-4597-9235-147fc4802d19" (UID: "c3b56f2d-96c0-4597-9235-147fc4802d19"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.883835 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1b040d3d-f4a0-451a-97de-08bea41c6dce-utilities" (OuterVolumeSpecName: "utilities") pod "1b040d3d-f4a0-451a-97de-08bea41c6dce" (UID: "1b040d3d-f4a0-451a-97de-08bea41c6dce"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.884502 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bc8c652-fb4f-48b7-86ce-c45522dfd476-kube-api-access-lchhw" (OuterVolumeSpecName: "kube-api-access-lchhw") pod "1bc8c652-fb4f-48b7-86ce-c45522dfd476" (UID: "1bc8c652-fb4f-48b7-86ce-c45522dfd476"). InnerVolumeSpecName "kube-api-access-lchhw". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.884679 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1b040d3d-f4a0-451a-97de-08bea41c6dce-kube-api-access-mxt8j" (OuterVolumeSpecName: "kube-api-access-mxt8j") pod "1b040d3d-f4a0-451a-97de-08bea41c6dce" (UID: "1b040d3d-f4a0-451a-97de-08bea41c6dce"). InnerVolumeSpecName "kube-api-access-mxt8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.885028 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3f11911f-e365-4151-b536-3988186b8757-kube-api-access-xjbt6" (OuterVolumeSpecName: "kube-api-access-xjbt6") pod "3f11911f-e365-4151-b536-3988186b8757" (UID: "3f11911f-e365-4151-b536-3988186b8757"). InnerVolumeSpecName "kube-api-access-xjbt6". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.885129 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/14d99a1f-fca3-421a-8b47-ab4cda2f77fc-kube-api-access-kjjkx" (OuterVolumeSpecName: "kube-api-access-kjjkx") pod "14d99a1f-fca3-421a-8b47-ab4cda2f77fc" (UID: "14d99a1f-fca3-421a-8b47-ab4cda2f77fc"). InnerVolumeSpecName "kube-api-access-kjjkx". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.885354 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14d99a1f-fca3-421a-8b47-ab4cda2f77fc-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "14d99a1f-fca3-421a-8b47-ab4cda2f77fc" (UID: "14d99a1f-fca3-421a-8b47-ab4cda2f77fc"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.885494 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c3b56f2d-96c0-4597-9235-147fc4802d19-kube-api-access-8656b" (OuterVolumeSpecName: "kube-api-access-8656b") pod "c3b56f2d-96c0-4597-9235-147fc4802d19" (UID: "c3b56f2d-96c0-4597-9235-147fc4802d19"). InnerVolumeSpecName "kube-api-access-8656b". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.901449 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1bc8c652-fb4f-48b7-86ce-c45522dfd476-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1bc8c652-fb4f-48b7-86ce-c45522dfd476" (UID: "1bc8c652-fb4f-48b7-86ce-c45522dfd476"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.905383 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.924401 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3f11911f-e365-4151-b536-3988186b8757-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3f11911f-e365-4151-b536-3988186b8757" (UID: "3f11911f-e365-4151-b536-3988186b8757"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.930840 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1b040d3d-f4a0-451a-97de-08bea41c6dce-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1b040d3d-f4a0-451a-97de-08bea41c6dce" (UID: "1b040d3d-f4a0-451a-97de-08bea41c6dce"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.979834 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c3b56f2d-96c0-4597-9235-147fc4802d19-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c3b56f2d-96c0-4597-9235-147fc4802d19" (UID: "c3b56f2d-96c0-4597-9235-147fc4802d19"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.983508 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mxt8j\" (UniqueName: \"kubernetes.io/projected/1b040d3d-f4a0-451a-97de-08bea41c6dce-kube-api-access-mxt8j\") on node \"crc\" DevicePath \"\"" Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.983533 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c3b56f2d-96c0-4597-9235-147fc4802d19-utilities\") on node \"crc\" DevicePath \"\"" Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.983545 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8656b\" (UniqueName: \"kubernetes.io/projected/c3b56f2d-96c0-4597-9235-147fc4802d19-kube-api-access-8656b\") on node \"crc\" DevicePath \"\"" Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.983572 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1bc8c652-fb4f-48b7-86ce-c45522dfd476-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.983581 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1bc8c652-fb4f-48b7-86ce-c45522dfd476-utilities\") on node \"crc\" DevicePath \"\"" Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.983589 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xjbt6\" (UniqueName: \"kubernetes.io/projected/3f11911f-e365-4151-b536-3988186b8757-kube-api-access-xjbt6\") on node \"crc\" DevicePath \"\"" Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.983597 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1b040d3d-f4a0-451a-97de-08bea41c6dce-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.983605 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kjjkx\" (UniqueName: \"kubernetes.io/projected/14d99a1f-fca3-421a-8b47-ab4cda2f77fc-kube-api-access-kjjkx\") on node \"crc\" DevicePath \"\"" Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.983613 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lchhw\" (UniqueName: \"kubernetes.io/projected/1bc8c652-fb4f-48b7-86ce-c45522dfd476-kube-api-access-lchhw\") on node \"crc\" DevicePath \"\"" Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.983620 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c3b56f2d-96c0-4597-9235-147fc4802d19-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.983628 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3f11911f-e365-4151-b536-3988186b8757-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.983635 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1b040d3d-f4a0-451a-97de-08bea41c6dce-utilities\") on node \"crc\" DevicePath \"\"" Jan 21 11:15:21 crc kubenswrapper[4824]: I0121 11:15:21.983643 4824 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/14d99a1f-fca3-421a-8b47-ab4cda2f77fc-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Jan 21 11:15:22 crc kubenswrapper[4824]: I0121 11:15:22.040256 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Jan 21 11:15:22 crc kubenswrapper[4824]: I0121 11:15:22.117406 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Jan 21 11:15:22 crc kubenswrapper[4824]: I0121 11:15:22.166872 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Jan 21 11:15:22 crc kubenswrapper[4824]: I0121 11:15:22.211206 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Jan 21 11:15:22 crc kubenswrapper[4824]: I0121 11:15:22.263054 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Jan 21 11:15:22 crc kubenswrapper[4824]: I0121 11:15:22.295195 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Jan 21 11:15:22 crc kubenswrapper[4824]: I0121 11:15:22.343149 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Jan 21 11:15:22 crc kubenswrapper[4824]: I0121 11:15:22.384940 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Jan 21 11:15:22 crc kubenswrapper[4824]: I0121 11:15:22.424870 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Jan 21 11:15:22 crc kubenswrapper[4824]: I0121 11:15:22.449075 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Jan 21 11:15:22 crc kubenswrapper[4824]: I0121 11:15:22.487776 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Jan 21 11:15:22 crc kubenswrapper[4824]: I0121 11:15:22.496333 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Jan 21 11:15:22 crc kubenswrapper[4824]: I0121 11:15:22.506105 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Jan 21 11:15:22 crc kubenswrapper[4824]: I0121 11:15:22.510722 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-b922x" event={"ID":"3f11911f-e365-4151-b536-3988186b8757","Type":"ContainerDied","Data":"8980ea76a1ff2b34ee51f0465162ae2c030d8714d1e864ebb767d3d7a1da8509"} Jan 21 11:15:22 crc kubenswrapper[4824]: I0121 11:15:22.510753 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-b922x" Jan 21 11:15:22 crc kubenswrapper[4824]: I0121 11:15:22.510769 4824 scope.go:117] "RemoveContainer" containerID="f4386ff52380b3c795405ecbb9387d4d9b5fca2cffa267eb0ddb32961b0a7f10" Jan 21 11:15:22 crc kubenswrapper[4824]: I0121 11:15:22.512801 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7q9f2" event={"ID":"1b040d3d-f4a0-451a-97de-08bea41c6dce","Type":"ContainerDied","Data":"58f9378707e041408b55a43bb2ea6d9092a260cf0c9b8d81dc167499cf1c4702"} Jan 21 11:15:22 crc kubenswrapper[4824]: I0121 11:15:22.512817 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7q9f2" Jan 21 11:15:22 crc kubenswrapper[4824]: I0121 11:15:22.514004 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-gjhp5" event={"ID":"14d99a1f-fca3-421a-8b47-ab4cda2f77fc","Type":"ContainerDied","Data":"3217737690e95d2837429663404e46050468bcb19f6fbf887ccc71adb8493eef"} Jan 21 11:15:22 crc kubenswrapper[4824]: I0121 11:15:22.514018 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-gjhp5" Jan 21 11:15:22 crc kubenswrapper[4824]: I0121 11:15:22.516594 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hqs52" event={"ID":"c3b56f2d-96c0-4597-9235-147fc4802d19","Type":"ContainerDied","Data":"331b281d41eb6e081cb5f5178cf5cdf5ccdc6fe5086d467942c51189e3b27cd1"} Jan 21 11:15:22 crc kubenswrapper[4824]: I0121 11:15:22.516774 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hqs52" Jan 21 11:15:22 crc kubenswrapper[4824]: I0121 11:15:22.518831 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4w5m6" event={"ID":"1bc8c652-fb4f-48b7-86ce-c45522dfd476","Type":"ContainerDied","Data":"1717bf3078cfbe6fee3800d069a35c471757cb03a1eddb1d8ff1e20167f89c9e"} Jan 21 11:15:22 crc kubenswrapper[4824]: I0121 11:15:22.518986 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4w5m6" Jan 21 11:15:22 crc kubenswrapper[4824]: I0121 11:15:22.522480 4824 scope.go:117] "RemoveContainer" containerID="bc5314fd2d078e4f04072ac4dda373754c451434b3d4dce44df1b40cbeecd1ba" Jan 21 11:15:22 crc kubenswrapper[4824]: I0121 11:15:22.525014 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-b922x"] Jan 21 11:15:22 crc kubenswrapper[4824]: I0121 11:15:22.528098 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-b922x"] Jan 21 11:15:22 crc kubenswrapper[4824]: I0121 11:15:22.530413 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Jan 21 11:15:22 crc kubenswrapper[4824]: I0121 11:15:22.534935 4824 scope.go:117] "RemoveContainer" containerID="fe402f640205024f989b1231ac3c21b31c3d7e53bed03337b9aa7afe32b7848e" Jan 21 11:15:22 crc kubenswrapper[4824]: I0121 11:15:22.535216 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-4w5m6"] Jan 21 11:15:22 crc kubenswrapper[4824]: I0121 11:15:22.537793 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-4w5m6"] Jan 21 11:15:22 crc kubenswrapper[4824]: I0121 11:15:22.541446 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-hqs52"] Jan 21 11:15:22 crc kubenswrapper[4824]: I0121 11:15:22.544890 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-hqs52"] Jan 21 11:15:22 crc kubenswrapper[4824]: I0121 11:15:22.548821 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-gjhp5"] Jan 21 11:15:22 crc kubenswrapper[4824]: I0121 11:15:22.548884 4824 scope.go:117] "RemoveContainer" containerID="728ea008f822360aaa1493d68beaf88c96b2cdf76a7eeb252c13acb6bf91c9d5" Jan 21 11:15:22 crc kubenswrapper[4824]: I0121 11:15:22.552344 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-gjhp5"] Jan 21 11:15:22 crc kubenswrapper[4824]: I0121 11:15:22.554447 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-7q9f2"] Jan 21 11:15:22 crc kubenswrapper[4824]: I0121 11:15:22.556536 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-7q9f2"] Jan 21 11:15:22 crc kubenswrapper[4824]: I0121 11:15:22.561592 4824 scope.go:117] "RemoveContainer" containerID="9f778c1f1c308a32a3e2af072eab8265d05e159ea1a783308d30e563350f290a" Jan 21 11:15:22 crc kubenswrapper[4824]: I0121 11:15:22.573044 4824 scope.go:117] "RemoveContainer" containerID="ca4b551b3e03faea4b176d4748d261efd694766575aea5c759f3fe832098d94c" Jan 21 11:15:22 crc kubenswrapper[4824]: I0121 11:15:22.588167 4824 scope.go:117] "RemoveContainer" containerID="605262189eb74a464efb5cb3518becda7f59cb81cb5a3ce736fbc1a29ea45dc4" Jan 21 11:15:22 crc kubenswrapper[4824]: I0121 11:15:22.597842 4824 scope.go:117] "RemoveContainer" containerID="b1068da5b4766f19c34535e8e58a22325fb8a2d31ce06ee547c973539c887a47" Jan 21 11:15:22 crc kubenswrapper[4824]: I0121 11:15:22.606281 4824 scope.go:117] "RemoveContainer" containerID="2f3db218d18e31256264bc1967baec36484f60d00de32760d6c6503fe4e7b9fc" Jan 21 11:15:22 crc kubenswrapper[4824]: I0121 11:15:22.617856 4824 scope.go:117] "RemoveContainer" containerID="550d4757a2a4f775989081e05280aba8d60bf329b7f947d0664b2247319cf4a8" Jan 21 11:15:22 crc kubenswrapper[4824]: I0121 11:15:22.630233 4824 scope.go:117] "RemoveContainer" containerID="fa2ad6f2a50b9350a988947c908be8adb082c0d1e40507b63b02b300f1699aae" Jan 21 11:15:22 crc kubenswrapper[4824]: I0121 11:15:22.641873 4824 scope.go:117] "RemoveContainer" containerID="122ae0144dee544f8ce79c06f524513fd679f1334f79370a0f9e8b920f8d7ee4" Jan 21 11:15:22 crc kubenswrapper[4824]: I0121 11:15:22.657924 4824 scope.go:117] "RemoveContainer" containerID="d5b0c30b7ab307dc38d2c9ff03ad628bf5226a1bb36f64f546ca1dedef2c81fc" Jan 21 11:15:22 crc kubenswrapper[4824]: I0121 11:15:22.704234 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Jan 21 11:15:22 crc kubenswrapper[4824]: I0121 11:15:22.802211 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Jan 21 11:15:22 crc kubenswrapper[4824]: I0121 11:15:22.804359 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Jan 21 11:15:22 crc kubenswrapper[4824]: I0121 11:15:22.876143 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Jan 21 11:15:22 crc kubenswrapper[4824]: I0121 11:15:22.933560 4824 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Jan 21 11:15:22 crc kubenswrapper[4824]: I0121 11:15:22.933946 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" containerID="cri-o://95939db35012b5a5c83445539a3c3ba0348e91a0065268d0f4bf7dc607ee2022" gracePeriod=5 Jan 21 11:15:23 crc kubenswrapper[4824]: I0121 11:15:23.018476 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Jan 21 11:15:23 crc kubenswrapper[4824]: I0121 11:15:23.024585 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Jan 21 11:15:23 crc kubenswrapper[4824]: I0121 11:15:23.154203 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Jan 21 11:15:23 crc kubenswrapper[4824]: I0121 11:15:23.275172 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Jan 21 11:15:23 crc kubenswrapper[4824]: I0121 11:15:23.286295 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Jan 21 11:15:23 crc kubenswrapper[4824]: I0121 11:15:23.319859 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Jan 21 11:15:23 crc kubenswrapper[4824]: I0121 11:15:23.340864 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Jan 21 11:15:23 crc kubenswrapper[4824]: I0121 11:15:23.455659 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Jan 21 11:15:23 crc kubenswrapper[4824]: I0121 11:15:23.554579 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Jan 21 11:15:23 crc kubenswrapper[4824]: I0121 11:15:23.589545 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Jan 21 11:15:23 crc kubenswrapper[4824]: I0121 11:15:23.626203 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Jan 21 11:15:23 crc kubenswrapper[4824]: I0121 11:15:23.884421 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Jan 21 11:15:23 crc kubenswrapper[4824]: I0121 11:15:23.907364 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Jan 21 11:15:24 crc kubenswrapper[4824]: I0121 11:15:24.033290 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Jan 21 11:15:24 crc kubenswrapper[4824]: I0121 11:15:24.034600 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Jan 21 11:15:24 crc kubenswrapper[4824]: I0121 11:15:24.054297 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="14d99a1f-fca3-421a-8b47-ab4cda2f77fc" path="/var/lib/kubelet/pods/14d99a1f-fca3-421a-8b47-ab4cda2f77fc/volumes" Jan 21 11:15:24 crc kubenswrapper[4824]: I0121 11:15:24.054781 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1b040d3d-f4a0-451a-97de-08bea41c6dce" path="/var/lib/kubelet/pods/1b040d3d-f4a0-451a-97de-08bea41c6dce/volumes" Jan 21 11:15:24 crc kubenswrapper[4824]: I0121 11:15:24.055390 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bc8c652-fb4f-48b7-86ce-c45522dfd476" path="/var/lib/kubelet/pods/1bc8c652-fb4f-48b7-86ce-c45522dfd476/volumes" Jan 21 11:15:24 crc kubenswrapper[4824]: I0121 11:15:24.056321 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3f11911f-e365-4151-b536-3988186b8757" path="/var/lib/kubelet/pods/3f11911f-e365-4151-b536-3988186b8757/volumes" Jan 21 11:15:24 crc kubenswrapper[4824]: I0121 11:15:24.056872 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c3b56f2d-96c0-4597-9235-147fc4802d19" path="/var/lib/kubelet/pods/c3b56f2d-96c0-4597-9235-147fc4802d19/volumes" Jan 21 11:15:24 crc kubenswrapper[4824]: I0121 11:15:24.161104 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Jan 21 11:15:24 crc kubenswrapper[4824]: I0121 11:15:24.364592 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Jan 21 11:15:24 crc kubenswrapper[4824]: I0121 11:15:24.381786 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Jan 21 11:15:24 crc kubenswrapper[4824]: E0121 11:15:24.415754 4824 log.go:32] "RunPodSandbox from runtime service failed" err=< Jan 21 11:15:24 crc kubenswrapper[4824]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_collect-profiles-29483235-4zbf5_openshift-operator-lifecycle-manager_123c89e0-3113-4b28-9129-3848a7288052_0(d44158efa73ad40b50e170b5fee95ff78fbf38ca5ab12b3c01f809937cd9545c): error adding pod openshift-operator-lifecycle-manager_collect-profiles-29483235-4zbf5 to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"d44158efa73ad40b50e170b5fee95ff78fbf38ca5ab12b3c01f809937cd9545c" Netns:"/var/run/netns/1adc7576-ade7-4a77-ada5-4aec576f0dcb" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-operator-lifecycle-manager;K8S_POD_NAME=collect-profiles-29483235-4zbf5;K8S_POD_INFRA_CONTAINER_ID=d44158efa73ad40b50e170b5fee95ff78fbf38ca5ab12b3c01f809937cd9545c;K8S_POD_UID=123c89e0-3113-4b28-9129-3848a7288052" Path:"" ERRORED: error configuring pod [openshift-operator-lifecycle-manager/collect-profiles-29483235-4zbf5] networking: Multus: [openshift-operator-lifecycle-manager/collect-profiles-29483235-4zbf5/123c89e0-3113-4b28-9129-3848a7288052]: error setting the networks status, pod was already deleted: SetPodNetworkStatusAnnotation: failed to query the pod collect-profiles-29483235-4zbf5 in out of cluster comm: pod "collect-profiles-29483235-4zbf5" not found Jan 21 11:15:24 crc kubenswrapper[4824]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Jan 21 11:15:24 crc kubenswrapper[4824]: > Jan 21 11:15:24 crc kubenswrapper[4824]: E0121 11:15:24.415823 4824 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err=< Jan 21 11:15:24 crc kubenswrapper[4824]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_collect-profiles-29483235-4zbf5_openshift-operator-lifecycle-manager_123c89e0-3113-4b28-9129-3848a7288052_0(d44158efa73ad40b50e170b5fee95ff78fbf38ca5ab12b3c01f809937cd9545c): error adding pod openshift-operator-lifecycle-manager_collect-profiles-29483235-4zbf5 to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"d44158efa73ad40b50e170b5fee95ff78fbf38ca5ab12b3c01f809937cd9545c" Netns:"/var/run/netns/1adc7576-ade7-4a77-ada5-4aec576f0dcb" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-operator-lifecycle-manager;K8S_POD_NAME=collect-profiles-29483235-4zbf5;K8S_POD_INFRA_CONTAINER_ID=d44158efa73ad40b50e170b5fee95ff78fbf38ca5ab12b3c01f809937cd9545c;K8S_POD_UID=123c89e0-3113-4b28-9129-3848a7288052" Path:"" ERRORED: error configuring pod [openshift-operator-lifecycle-manager/collect-profiles-29483235-4zbf5] networking: Multus: [openshift-operator-lifecycle-manager/collect-profiles-29483235-4zbf5/123c89e0-3113-4b28-9129-3848a7288052]: error setting the networks status, pod was already deleted: SetPodNetworkStatusAnnotation: failed to query the pod collect-profiles-29483235-4zbf5 in out of cluster comm: pod "collect-profiles-29483235-4zbf5" not found Jan 21 11:15:24 crc kubenswrapper[4824]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Jan 21 11:15:24 crc kubenswrapper[4824]: > pod="openshift-operator-lifecycle-manager/collect-profiles-29483235-4zbf5" Jan 21 11:15:24 crc kubenswrapper[4824]: E0121 11:15:24.415843 4824 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err=< Jan 21 11:15:24 crc kubenswrapper[4824]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_collect-profiles-29483235-4zbf5_openshift-operator-lifecycle-manager_123c89e0-3113-4b28-9129-3848a7288052_0(d44158efa73ad40b50e170b5fee95ff78fbf38ca5ab12b3c01f809937cd9545c): error adding pod openshift-operator-lifecycle-manager_collect-profiles-29483235-4zbf5 to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"d44158efa73ad40b50e170b5fee95ff78fbf38ca5ab12b3c01f809937cd9545c" Netns:"/var/run/netns/1adc7576-ade7-4a77-ada5-4aec576f0dcb" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-operator-lifecycle-manager;K8S_POD_NAME=collect-profiles-29483235-4zbf5;K8S_POD_INFRA_CONTAINER_ID=d44158efa73ad40b50e170b5fee95ff78fbf38ca5ab12b3c01f809937cd9545c;K8S_POD_UID=123c89e0-3113-4b28-9129-3848a7288052" Path:"" ERRORED: error configuring pod [openshift-operator-lifecycle-manager/collect-profiles-29483235-4zbf5] networking: Multus: [openshift-operator-lifecycle-manager/collect-profiles-29483235-4zbf5/123c89e0-3113-4b28-9129-3848a7288052]: error setting the networks status, pod was already deleted: SetPodNetworkStatusAnnotation: failed to query the pod collect-profiles-29483235-4zbf5 in out of cluster comm: pod "collect-profiles-29483235-4zbf5" not found Jan 21 11:15:24 crc kubenswrapper[4824]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Jan 21 11:15:24 crc kubenswrapper[4824]: > pod="openshift-operator-lifecycle-manager/collect-profiles-29483235-4zbf5" Jan 21 11:15:24 crc kubenswrapper[4824]: E0121 11:15:24.415907 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"collect-profiles-29483235-4zbf5_openshift-operator-lifecycle-manager(123c89e0-3113-4b28-9129-3848a7288052)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"collect-profiles-29483235-4zbf5_openshift-operator-lifecycle-manager(123c89e0-3113-4b28-9129-3848a7288052)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_collect-profiles-29483235-4zbf5_openshift-operator-lifecycle-manager_123c89e0-3113-4b28-9129-3848a7288052_0(d44158efa73ad40b50e170b5fee95ff78fbf38ca5ab12b3c01f809937cd9545c): error adding pod openshift-operator-lifecycle-manager_collect-profiles-29483235-4zbf5 to CNI network \\\"multus-cni-network\\\": plugin type=\\\"multus-shim\\\" name=\\\"multus-cni-network\\\" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:\\\"d44158efa73ad40b50e170b5fee95ff78fbf38ca5ab12b3c01f809937cd9545c\\\" Netns:\\\"/var/run/netns/1adc7576-ade7-4a77-ada5-4aec576f0dcb\\\" IfName:\\\"eth0\\\" Args:\\\"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-operator-lifecycle-manager;K8S_POD_NAME=collect-profiles-29483235-4zbf5;K8S_POD_INFRA_CONTAINER_ID=d44158efa73ad40b50e170b5fee95ff78fbf38ca5ab12b3c01f809937cd9545c;K8S_POD_UID=123c89e0-3113-4b28-9129-3848a7288052\\\" Path:\\\"\\\" ERRORED: error configuring pod [openshift-operator-lifecycle-manager/collect-profiles-29483235-4zbf5] networking: Multus: [openshift-operator-lifecycle-manager/collect-profiles-29483235-4zbf5/123c89e0-3113-4b28-9129-3848a7288052]: error setting the networks status, pod was already deleted: SetPodNetworkStatusAnnotation: failed to query the pod collect-profiles-29483235-4zbf5 in out of cluster comm: pod \\\"collect-profiles-29483235-4zbf5\\\" not found\\n': StdinData: {\\\"binDir\\\":\\\"/var/lib/cni/bin\\\",\\\"clusterNetwork\\\":\\\"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf\\\",\\\"cniVersion\\\":\\\"0.3.1\\\",\\\"daemonSocketDir\\\":\\\"/run/multus/socket\\\",\\\"globalNamespaces\\\":\\\"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv\\\",\\\"logLevel\\\":\\\"verbose\\\",\\\"logToStderr\\\":true,\\\"name\\\":\\\"multus-cni-network\\\",\\\"namespaceIsolation\\\":true,\\\"type\\\":\\\"multus-shim\\\"}\"" pod="openshift-operator-lifecycle-manager/collect-profiles-29483235-4zbf5" podUID="123c89e0-3113-4b28-9129-3848a7288052" Jan 21 11:15:24 crc kubenswrapper[4824]: I0121 11:15:24.528916 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29483235-4zbf5" Jan 21 11:15:24 crc kubenswrapper[4824]: I0121 11:15:24.529267 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29483235-4zbf5" Jan 21 11:15:24 crc kubenswrapper[4824]: I0121 11:15:24.559004 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Jan 21 11:15:24 crc kubenswrapper[4824]: I0121 11:15:24.588501 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Jan 21 11:15:24 crc kubenswrapper[4824]: I0121 11:15:24.614838 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Jan 21 11:15:24 crc kubenswrapper[4824]: E0121 11:15:24.677260 4824 log.go:32] "RunPodSandbox from runtime service failed" err=< Jan 21 11:15:24 crc kubenswrapper[4824]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_marketplace-operator-79b997595-xcxs9_openshift-marketplace_b4c29122-d9b2-4f85-8b29-c0e7d888667f_0(f263b9e3eef1221c6694629a71faae3eb9597658cf0c30c83ddc359ef4d97f24): error adding pod openshift-marketplace_marketplace-operator-79b997595-xcxs9 to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"f263b9e3eef1221c6694629a71faae3eb9597658cf0c30c83ddc359ef4d97f24" Netns:"/var/run/netns/e298ec93-b3e4-4cb3-8240-38b80e92dc97" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-marketplace;K8S_POD_NAME=marketplace-operator-79b997595-xcxs9;K8S_POD_INFRA_CONTAINER_ID=f263b9e3eef1221c6694629a71faae3eb9597658cf0c30c83ddc359ef4d97f24;K8S_POD_UID=b4c29122-d9b2-4f85-8b29-c0e7d888667f" Path:"" ERRORED: error configuring pod [openshift-marketplace/marketplace-operator-79b997595-xcxs9] networking: Multus: [openshift-marketplace/marketplace-operator-79b997595-xcxs9/b4c29122-d9b2-4f85-8b29-c0e7d888667f]: error setting the networks status, pod was already deleted: SetPodNetworkStatusAnnotation: failed to query the pod marketplace-operator-79b997595-xcxs9 in out of cluster comm: pod "marketplace-operator-79b997595-xcxs9" not found Jan 21 11:15:24 crc kubenswrapper[4824]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Jan 21 11:15:24 crc kubenswrapper[4824]: > Jan 21 11:15:24 crc kubenswrapper[4824]: E0121 11:15:24.677309 4824 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err=< Jan 21 11:15:24 crc kubenswrapper[4824]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_marketplace-operator-79b997595-xcxs9_openshift-marketplace_b4c29122-d9b2-4f85-8b29-c0e7d888667f_0(f263b9e3eef1221c6694629a71faae3eb9597658cf0c30c83ddc359ef4d97f24): error adding pod openshift-marketplace_marketplace-operator-79b997595-xcxs9 to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"f263b9e3eef1221c6694629a71faae3eb9597658cf0c30c83ddc359ef4d97f24" Netns:"/var/run/netns/e298ec93-b3e4-4cb3-8240-38b80e92dc97" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-marketplace;K8S_POD_NAME=marketplace-operator-79b997595-xcxs9;K8S_POD_INFRA_CONTAINER_ID=f263b9e3eef1221c6694629a71faae3eb9597658cf0c30c83ddc359ef4d97f24;K8S_POD_UID=b4c29122-d9b2-4f85-8b29-c0e7d888667f" Path:"" ERRORED: error configuring pod [openshift-marketplace/marketplace-operator-79b997595-xcxs9] networking: Multus: [openshift-marketplace/marketplace-operator-79b997595-xcxs9/b4c29122-d9b2-4f85-8b29-c0e7d888667f]: error setting the networks status, pod was already deleted: SetPodNetworkStatusAnnotation: failed to query the pod marketplace-operator-79b997595-xcxs9 in out of cluster comm: pod "marketplace-operator-79b997595-xcxs9" not found Jan 21 11:15:24 crc kubenswrapper[4824]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Jan 21 11:15:24 crc kubenswrapper[4824]: > pod="openshift-marketplace/marketplace-operator-79b997595-xcxs9" Jan 21 11:15:24 crc kubenswrapper[4824]: E0121 11:15:24.677325 4824 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err=< Jan 21 11:15:24 crc kubenswrapper[4824]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_marketplace-operator-79b997595-xcxs9_openshift-marketplace_b4c29122-d9b2-4f85-8b29-c0e7d888667f_0(f263b9e3eef1221c6694629a71faae3eb9597658cf0c30c83ddc359ef4d97f24): error adding pod openshift-marketplace_marketplace-operator-79b997595-xcxs9 to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"f263b9e3eef1221c6694629a71faae3eb9597658cf0c30c83ddc359ef4d97f24" Netns:"/var/run/netns/e298ec93-b3e4-4cb3-8240-38b80e92dc97" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-marketplace;K8S_POD_NAME=marketplace-operator-79b997595-xcxs9;K8S_POD_INFRA_CONTAINER_ID=f263b9e3eef1221c6694629a71faae3eb9597658cf0c30c83ddc359ef4d97f24;K8S_POD_UID=b4c29122-d9b2-4f85-8b29-c0e7d888667f" Path:"" ERRORED: error configuring pod [openshift-marketplace/marketplace-operator-79b997595-xcxs9] networking: Multus: [openshift-marketplace/marketplace-operator-79b997595-xcxs9/b4c29122-d9b2-4f85-8b29-c0e7d888667f]: error setting the networks status, pod was already deleted: SetPodNetworkStatusAnnotation: failed to query the pod marketplace-operator-79b997595-xcxs9 in out of cluster comm: pod "marketplace-operator-79b997595-xcxs9" not found Jan 21 11:15:24 crc kubenswrapper[4824]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Jan 21 11:15:24 crc kubenswrapper[4824]: > pod="openshift-marketplace/marketplace-operator-79b997595-xcxs9" Jan 21 11:15:24 crc kubenswrapper[4824]: E0121 11:15:24.677369 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"marketplace-operator-79b997595-xcxs9_openshift-marketplace(b4c29122-d9b2-4f85-8b29-c0e7d888667f)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"marketplace-operator-79b997595-xcxs9_openshift-marketplace(b4c29122-d9b2-4f85-8b29-c0e7d888667f)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_marketplace-operator-79b997595-xcxs9_openshift-marketplace_b4c29122-d9b2-4f85-8b29-c0e7d888667f_0(f263b9e3eef1221c6694629a71faae3eb9597658cf0c30c83ddc359ef4d97f24): error adding pod openshift-marketplace_marketplace-operator-79b997595-xcxs9 to CNI network \\\"multus-cni-network\\\": plugin type=\\\"multus-shim\\\" name=\\\"multus-cni-network\\\" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:\\\"f263b9e3eef1221c6694629a71faae3eb9597658cf0c30c83ddc359ef4d97f24\\\" Netns:\\\"/var/run/netns/e298ec93-b3e4-4cb3-8240-38b80e92dc97\\\" IfName:\\\"eth0\\\" Args:\\\"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-marketplace;K8S_POD_NAME=marketplace-operator-79b997595-xcxs9;K8S_POD_INFRA_CONTAINER_ID=f263b9e3eef1221c6694629a71faae3eb9597658cf0c30c83ddc359ef4d97f24;K8S_POD_UID=b4c29122-d9b2-4f85-8b29-c0e7d888667f\\\" Path:\\\"\\\" ERRORED: error configuring pod [openshift-marketplace/marketplace-operator-79b997595-xcxs9] networking: Multus: [openshift-marketplace/marketplace-operator-79b997595-xcxs9/b4c29122-d9b2-4f85-8b29-c0e7d888667f]: error setting the networks status, pod was already deleted: SetPodNetworkStatusAnnotation: failed to query the pod marketplace-operator-79b997595-xcxs9 in out of cluster comm: pod \\\"marketplace-operator-79b997595-xcxs9\\\" not found\\n': StdinData: {\\\"binDir\\\":\\\"/var/lib/cni/bin\\\",\\\"clusterNetwork\\\":\\\"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf\\\",\\\"cniVersion\\\":\\\"0.3.1\\\",\\\"daemonSocketDir\\\":\\\"/run/multus/socket\\\",\\\"globalNamespaces\\\":\\\"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv\\\",\\\"logLevel\\\":\\\"verbose\\\",\\\"logToStderr\\\":true,\\\"name\\\":\\\"multus-cni-network\\\",\\\"namespaceIsolation\\\":true,\\\"type\\\":\\\"multus-shim\\\"}\"" pod="openshift-marketplace/marketplace-operator-79b997595-xcxs9" podUID="b4c29122-d9b2-4f85-8b29-c0e7d888667f" Jan 21 11:15:24 crc kubenswrapper[4824]: I0121 11:15:24.837421 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Jan 21 11:15:24 crc kubenswrapper[4824]: I0121 11:15:24.884749 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29483235-4zbf5"] Jan 21 11:15:24 crc kubenswrapper[4824]: I0121 11:15:24.965758 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Jan 21 11:15:25 crc kubenswrapper[4824]: I0121 11:15:25.125553 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Jan 21 11:15:25 crc kubenswrapper[4824]: I0121 11:15:25.211943 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Jan 21 11:15:25 crc kubenswrapper[4824]: I0121 11:15:25.310823 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Jan 21 11:15:25 crc kubenswrapper[4824]: I0121 11:15:25.312306 4824 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Jan 21 11:15:25 crc kubenswrapper[4824]: I0121 11:15:25.353592 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Jan 21 11:15:25 crc kubenswrapper[4824]: I0121 11:15:25.357187 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Jan 21 11:15:25 crc kubenswrapper[4824]: I0121 11:15:25.533824 4824 generic.go:334] "Generic (PLEG): container finished" podID="123c89e0-3113-4b28-9129-3848a7288052" containerID="1c72418277c3fff2993fb8936f438b1d2049c9f13bdefd24ee7c2900f6904eca" exitCode=0 Jan 21 11:15:25 crc kubenswrapper[4824]: I0121 11:15:25.533862 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29483235-4zbf5" event={"ID":"123c89e0-3113-4b28-9129-3848a7288052","Type":"ContainerDied","Data":"1c72418277c3fff2993fb8936f438b1d2049c9f13bdefd24ee7c2900f6904eca"} Jan 21 11:15:25 crc kubenswrapper[4824]: I0121 11:15:25.533892 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-xcxs9" Jan 21 11:15:25 crc kubenswrapper[4824]: I0121 11:15:25.533902 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29483235-4zbf5" event={"ID":"123c89e0-3113-4b28-9129-3848a7288052","Type":"ContainerStarted","Data":"100dbb4b83b5c4af8945cd603c66cd0fed7d5003f712c7ad794dd486ee3cadf0"} Jan 21 11:15:25 crc kubenswrapper[4824]: I0121 11:15:25.534298 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-xcxs9" Jan 21 11:15:25 crc kubenswrapper[4824]: I0121 11:15:25.543736 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Jan 21 11:15:25 crc kubenswrapper[4824]: I0121 11:15:25.677615 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Jan 21 11:15:25 crc kubenswrapper[4824]: I0121 11:15:25.692944 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Jan 21 11:15:25 crc kubenswrapper[4824]: I0121 11:15:25.802231 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Jan 21 11:15:25 crc kubenswrapper[4824]: I0121 11:15:25.897180 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Jan 21 11:15:25 crc kubenswrapper[4824]: I0121 11:15:25.906235 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-xcxs9"] Jan 21 11:15:25 crc kubenswrapper[4824]: W0121 11:15:25.912529 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb4c29122_d9b2_4f85_8b29_c0e7d888667f.slice/crio-432aa6881d174864fae30b87af6d2da57882f48809f9ba7358a64d46d8bcfa8b WatchSource:0}: Error finding container 432aa6881d174864fae30b87af6d2da57882f48809f9ba7358a64d46d8bcfa8b: Status 404 returned error can't find the container with id 432aa6881d174864fae30b87af6d2da57882f48809f9ba7358a64d46d8bcfa8b Jan 21 11:15:25 crc kubenswrapper[4824]: I0121 11:15:25.931235 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Jan 21 11:15:25 crc kubenswrapper[4824]: I0121 11:15:25.938648 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Jan 21 11:15:26 crc kubenswrapper[4824]: I0121 11:15:26.024540 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Jan 21 11:15:26 crc kubenswrapper[4824]: I0121 11:15:26.270326 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Jan 21 11:15:26 crc kubenswrapper[4824]: I0121 11:15:26.300168 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Jan 21 11:15:26 crc kubenswrapper[4824]: I0121 11:15:26.349663 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Jan 21 11:15:26 crc kubenswrapper[4824]: I0121 11:15:26.364679 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Jan 21 11:15:26 crc kubenswrapper[4824]: I0121 11:15:26.414027 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Jan 21 11:15:26 crc kubenswrapper[4824]: I0121 11:15:26.538295 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Jan 21 11:15:26 crc kubenswrapper[4824]: I0121 11:15:26.538873 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-xcxs9" event={"ID":"b4c29122-d9b2-4f85-8b29-c0e7d888667f","Type":"ContainerStarted","Data":"55b402fb3c0dcd6e053dd4d185841bbc3a89f80a565e0957c47691ca0a96a966"} Jan 21 11:15:26 crc kubenswrapper[4824]: I0121 11:15:26.538931 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-xcxs9" event={"ID":"b4c29122-d9b2-4f85-8b29-c0e7d888667f","Type":"ContainerStarted","Data":"432aa6881d174864fae30b87af6d2da57882f48809f9ba7358a64d46d8bcfa8b"} Jan 21 11:15:26 crc kubenswrapper[4824]: I0121 11:15:26.546980 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Jan 21 11:15:26 crc kubenswrapper[4824]: I0121 11:15:26.702420 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Jan 21 11:15:26 crc kubenswrapper[4824]: I0121 11:15:26.712766 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Jan 21 11:15:26 crc kubenswrapper[4824]: I0121 11:15:26.728022 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29483235-4zbf5" Jan 21 11:15:26 crc kubenswrapper[4824]: I0121 11:15:26.741637 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-xcxs9" podStartSLOduration=11.741620253 podStartE2EDuration="11.741620253s" podCreationTimestamp="2026-01-21 11:15:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:15:26.559199354 +0000 UTC m=+268.852228646" watchObservedRunningTime="2026-01-21 11:15:26.741620253 +0000 UTC m=+269.034649535" Jan 21 11:15:26 crc kubenswrapper[4824]: I0121 11:15:26.770971 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Jan 21 11:15:26 crc kubenswrapper[4824]: I0121 11:15:26.794210 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Jan 21 11:15:26 crc kubenswrapper[4824]: I0121 11:15:26.839336 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/123c89e0-3113-4b28-9129-3848a7288052-config-volume\") pod \"123c89e0-3113-4b28-9129-3848a7288052\" (UID: \"123c89e0-3113-4b28-9129-3848a7288052\") " Jan 21 11:15:26 crc kubenswrapper[4824]: I0121 11:15:26.839368 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/123c89e0-3113-4b28-9129-3848a7288052-secret-volume\") pod \"123c89e0-3113-4b28-9129-3848a7288052\" (UID: \"123c89e0-3113-4b28-9129-3848a7288052\") " Jan 21 11:15:26 crc kubenswrapper[4824]: I0121 11:15:26.839403 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fxbxn\" (UniqueName: \"kubernetes.io/projected/123c89e0-3113-4b28-9129-3848a7288052-kube-api-access-fxbxn\") pod \"123c89e0-3113-4b28-9129-3848a7288052\" (UID: \"123c89e0-3113-4b28-9129-3848a7288052\") " Jan 21 11:15:26 crc kubenswrapper[4824]: I0121 11:15:26.839892 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/123c89e0-3113-4b28-9129-3848a7288052-config-volume" (OuterVolumeSpecName: "config-volume") pod "123c89e0-3113-4b28-9129-3848a7288052" (UID: "123c89e0-3113-4b28-9129-3848a7288052"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:15:26 crc kubenswrapper[4824]: I0121 11:15:26.847045 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/123c89e0-3113-4b28-9129-3848a7288052-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "123c89e0-3113-4b28-9129-3848a7288052" (UID: "123c89e0-3113-4b28-9129-3848a7288052"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:15:26 crc kubenswrapper[4824]: I0121 11:15:26.847079 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/123c89e0-3113-4b28-9129-3848a7288052-kube-api-access-fxbxn" (OuterVolumeSpecName: "kube-api-access-fxbxn") pod "123c89e0-3113-4b28-9129-3848a7288052" (UID: "123c89e0-3113-4b28-9129-3848a7288052"). InnerVolumeSpecName "kube-api-access-fxbxn". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:15:26 crc kubenswrapper[4824]: I0121 11:15:26.858546 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Jan 21 11:15:26 crc kubenswrapper[4824]: I0121 11:15:26.890748 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Jan 21 11:15:26 crc kubenswrapper[4824]: I0121 11:15:26.938845 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Jan 21 11:15:26 crc kubenswrapper[4824]: I0121 11:15:26.940227 4824 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/123c89e0-3113-4b28-9129-3848a7288052-config-volume\") on node \"crc\" DevicePath \"\"" Jan 21 11:15:26 crc kubenswrapper[4824]: I0121 11:15:26.940246 4824 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/123c89e0-3113-4b28-9129-3848a7288052-secret-volume\") on node \"crc\" DevicePath \"\"" Jan 21 11:15:26 crc kubenswrapper[4824]: I0121 11:15:26.940255 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fxbxn\" (UniqueName: \"kubernetes.io/projected/123c89e0-3113-4b28-9129-3848a7288052-kube-api-access-fxbxn\") on node \"crc\" DevicePath \"\"" Jan 21 11:15:26 crc kubenswrapper[4824]: I0121 11:15:26.946587 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Jan 21 11:15:27 crc kubenswrapper[4824]: I0121 11:15:27.221499 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Jan 21 11:15:27 crc kubenswrapper[4824]: I0121 11:15:27.246519 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Jan 21 11:15:27 crc kubenswrapper[4824]: I0121 11:15:27.325177 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Jan 21 11:15:27 crc kubenswrapper[4824]: I0121 11:15:27.325804 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Jan 21 11:15:27 crc kubenswrapper[4824]: I0121 11:15:27.459552 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Jan 21 11:15:27 crc kubenswrapper[4824]: I0121 11:15:27.479021 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Jan 21 11:15:27 crc kubenswrapper[4824]: I0121 11:15:27.506305 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Jan 21 11:15:27 crc kubenswrapper[4824]: I0121 11:15:27.545332 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29483235-4zbf5" Jan 21 11:15:27 crc kubenswrapper[4824]: I0121 11:15:27.545346 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29483235-4zbf5" event={"ID":"123c89e0-3113-4b28-9129-3848a7288052","Type":"ContainerDied","Data":"100dbb4b83b5c4af8945cd603c66cd0fed7d5003f712c7ad794dd486ee3cadf0"} Jan 21 11:15:27 crc kubenswrapper[4824]: I0121 11:15:27.545399 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="100dbb4b83b5c4af8945cd603c66cd0fed7d5003f712c7ad794dd486ee3cadf0" Jan 21 11:15:27 crc kubenswrapper[4824]: I0121 11:15:27.545431 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-xcxs9" Jan 21 11:15:27 crc kubenswrapper[4824]: I0121 11:15:27.548338 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-xcxs9" Jan 21 11:15:27 crc kubenswrapper[4824]: I0121 11:15:27.828563 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Jan 21 11:15:27 crc kubenswrapper[4824]: I0121 11:15:27.855235 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Jan 21 11:15:27 crc kubenswrapper[4824]: I0121 11:15:27.894845 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Jan 21 11:15:28 crc kubenswrapper[4824]: I0121 11:15:28.029895 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Jan 21 11:15:28 crc kubenswrapper[4824]: I0121 11:15:28.030466 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 21 11:15:28 crc kubenswrapper[4824]: I0121 11:15:28.053751 4824 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="" Jan 21 11:15:28 crc kubenswrapper[4824]: I0121 11:15:28.061654 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Jan 21 11:15:28 crc kubenswrapper[4824]: I0121 11:15:28.061679 4824 kubelet.go:2649] "Unable to find pod for mirror pod, skipping" mirrorPod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" mirrorPodUID="8bd8c0d7-b03c-4413-813b-aae449f2f506" Jan 21 11:15:28 crc kubenswrapper[4824]: I0121 11:15:28.064381 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Jan 21 11:15:28 crc kubenswrapper[4824]: I0121 11:15:28.064407 4824 kubelet.go:2673] "Unable to find pod for mirror pod, skipping" mirrorPod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" mirrorPodUID="8bd8c0d7-b03c-4413-813b-aae449f2f506" Jan 21 11:15:28 crc kubenswrapper[4824]: I0121 11:15:28.138535 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Jan 21 11:15:28 crc kubenswrapper[4824]: I0121 11:15:28.140479 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Jan 21 11:15:28 crc kubenswrapper[4824]: I0121 11:15:28.153629 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Jan 21 11:15:28 crc kubenswrapper[4824]: I0121 11:15:28.153682 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Jan 21 11:15:28 crc kubenswrapper[4824]: I0121 11:15:28.153716 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Jan 21 11:15:28 crc kubenswrapper[4824]: I0121 11:15:28.153715 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock" (OuterVolumeSpecName: "var-lock") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 21 11:15:28 crc kubenswrapper[4824]: I0121 11:15:28.153734 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Jan 21 11:15:28 crc kubenswrapper[4824]: I0121 11:15:28.153755 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 21 11:15:28 crc kubenswrapper[4824]: I0121 11:15:28.153768 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Jan 21 11:15:28 crc kubenswrapper[4824]: I0121 11:15:28.153802 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests" (OuterVolumeSpecName: "manifests") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "manifests". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 21 11:15:28 crc kubenswrapper[4824]: I0121 11:15:28.153852 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log" (OuterVolumeSpecName: "var-log") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 21 11:15:28 crc kubenswrapper[4824]: I0121 11:15:28.154101 4824 reconciler_common.go:293] "Volume detached for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") on node \"crc\" DevicePath \"\"" Jan 21 11:15:28 crc kubenswrapper[4824]: I0121 11:15:28.154262 4824 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") on node \"crc\" DevicePath \"\"" Jan 21 11:15:28 crc kubenswrapper[4824]: I0121 11:15:28.154422 4824 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") on node \"crc\" DevicePath \"\"" Jan 21 11:15:28 crc kubenswrapper[4824]: I0121 11:15:28.154450 4824 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") on node \"crc\" DevicePath \"\"" Jan 21 11:15:28 crc kubenswrapper[4824]: I0121 11:15:28.159668 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir" (OuterVolumeSpecName: "pod-resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "pod-resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 21 11:15:28 crc kubenswrapper[4824]: I0121 11:15:28.212651 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Jan 21 11:15:28 crc kubenswrapper[4824]: I0121 11:15:28.254948 4824 reconciler_common.go:293] "Volume detached for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") on node \"crc\" DevicePath \"\"" Jan 21 11:15:28 crc kubenswrapper[4824]: I0121 11:15:28.548221 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Jan 21 11:15:28 crc kubenswrapper[4824]: I0121 11:15:28.548457 4824 generic.go:334] "Generic (PLEG): container finished" podID="f85e55b1a89d02b0cb034b1ea31ed45a" containerID="95939db35012b5a5c83445539a3c3ba0348e91a0065268d0f4bf7dc607ee2022" exitCode=137 Jan 21 11:15:28 crc kubenswrapper[4824]: I0121 11:15:28.548519 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 21 11:15:28 crc kubenswrapper[4824]: I0121 11:15:28.548524 4824 scope.go:117] "RemoveContainer" containerID="95939db35012b5a5c83445539a3c3ba0348e91a0065268d0f4bf7dc607ee2022" Jan 21 11:15:28 crc kubenswrapper[4824]: I0121 11:15:28.552126 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Jan 21 11:15:28 crc kubenswrapper[4824]: I0121 11:15:28.569008 4824 scope.go:117] "RemoveContainer" containerID="95939db35012b5a5c83445539a3c3ba0348e91a0065268d0f4bf7dc607ee2022" Jan 21 11:15:28 crc kubenswrapper[4824]: E0121 11:15:28.569571 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"95939db35012b5a5c83445539a3c3ba0348e91a0065268d0f4bf7dc607ee2022\": container with ID starting with 95939db35012b5a5c83445539a3c3ba0348e91a0065268d0f4bf7dc607ee2022 not found: ID does not exist" containerID="95939db35012b5a5c83445539a3c3ba0348e91a0065268d0f4bf7dc607ee2022" Jan 21 11:15:28 crc kubenswrapper[4824]: I0121 11:15:28.569608 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"95939db35012b5a5c83445539a3c3ba0348e91a0065268d0f4bf7dc607ee2022"} err="failed to get container status \"95939db35012b5a5c83445539a3c3ba0348e91a0065268d0f4bf7dc607ee2022\": rpc error: code = NotFound desc = could not find container \"95939db35012b5a5c83445539a3c3ba0348e91a0065268d0f4bf7dc607ee2022\": container with ID starting with 95939db35012b5a5c83445539a3c3ba0348e91a0065268d0f4bf7dc607ee2022 not found: ID does not exist" Jan 21 11:15:28 crc kubenswrapper[4824]: I0121 11:15:28.677859 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Jan 21 11:15:28 crc kubenswrapper[4824]: I0121 11:15:28.760653 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Jan 21 11:15:28 crc kubenswrapper[4824]: I0121 11:15:28.826973 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Jan 21 11:15:28 crc kubenswrapper[4824]: I0121 11:15:28.974946 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Jan 21 11:15:29 crc kubenswrapper[4824]: I0121 11:15:29.015232 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Jan 21 11:15:29 crc kubenswrapper[4824]: I0121 11:15:29.124849 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Jan 21 11:15:29 crc kubenswrapper[4824]: I0121 11:15:29.150834 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Jan 21 11:15:29 crc kubenswrapper[4824]: I0121 11:15:29.197422 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Jan 21 11:15:29 crc kubenswrapper[4824]: I0121 11:15:29.210630 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Jan 21 11:15:29 crc kubenswrapper[4824]: I0121 11:15:29.284230 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Jan 21 11:15:29 crc kubenswrapper[4824]: I0121 11:15:29.346051 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Jan 21 11:15:29 crc kubenswrapper[4824]: I0121 11:15:29.418897 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Jan 21 11:15:29 crc kubenswrapper[4824]: I0121 11:15:29.535944 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Jan 21 11:15:29 crc kubenswrapper[4824]: I0121 11:15:29.737048 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Jan 21 11:15:30 crc kubenswrapper[4824]: I0121 11:15:30.026181 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Jan 21 11:15:30 crc kubenswrapper[4824]: I0121 11:15:30.030503 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Jan 21 11:15:30 crc kubenswrapper[4824]: I0121 11:15:30.066436 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" path="/var/lib/kubelet/pods/f85e55b1a89d02b0cb034b1ea31ed45a/volumes" Jan 21 11:15:30 crc kubenswrapper[4824]: I0121 11:15:30.229481 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Jan 21 11:15:30 crc kubenswrapper[4824]: I0121 11:15:30.260431 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Jan 21 11:15:30 crc kubenswrapper[4824]: I0121 11:15:30.468469 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Jan 21 11:15:30 crc kubenswrapper[4824]: I0121 11:15:30.797448 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Jan 21 11:15:31 crc kubenswrapper[4824]: I0121 11:15:31.189775 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Jan 21 11:15:31 crc kubenswrapper[4824]: I0121 11:15:31.429857 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Jan 21 11:15:31 crc kubenswrapper[4824]: I0121 11:15:31.458534 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Jan 21 11:15:31 crc kubenswrapper[4824]: I0121 11:15:31.533060 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Jan 21 11:15:31 crc kubenswrapper[4824]: I0121 11:15:31.771303 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Jan 21 11:15:31 crc kubenswrapper[4824]: I0121 11:15:31.995467 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Jan 21 11:15:32 crc kubenswrapper[4824]: I0121 11:15:32.027976 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Jan 21 11:15:32 crc kubenswrapper[4824]: I0121 11:15:32.211152 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Jan 21 11:15:47 crc kubenswrapper[4824]: I0121 11:15:47.828136 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-4dkd8"] Jan 21 11:15:47 crc kubenswrapper[4824]: I0121 11:15:47.828668 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-4dkd8" podUID="d962d0be-438a-4d08-8113-525238fa2156" containerName="controller-manager" containerID="cri-o://13eb81088f05a509df912f982b6e26892b6c0fee8b275279ea88e029d704d428" gracePeriod=30 Jan 21 11:15:47 crc kubenswrapper[4824]: I0121 11:15:47.922393 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-qs8d6"] Jan 21 11:15:47 crc kubenswrapper[4824]: I0121 11:15:47.922565 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qs8d6" podUID="51e27630-d5c1-4eed-8232-3698ab99fc19" containerName="route-controller-manager" containerID="cri-o://e407f561a94b91929f1905dd35fb9fce0694555be001231861d30f794a254728" gracePeriod=30 Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.097770 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-4dkd8" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.178883 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qs8d6" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.251883 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d962d0be-438a-4d08-8113-525238fa2156-serving-cert\") pod \"d962d0be-438a-4d08-8113-525238fa2156\" (UID: \"d962d0be-438a-4d08-8113-525238fa2156\") " Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.252192 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5ksj6\" (UniqueName: \"kubernetes.io/projected/51e27630-d5c1-4eed-8232-3698ab99fc19-kube-api-access-5ksj6\") pod \"51e27630-d5c1-4eed-8232-3698ab99fc19\" (UID: \"51e27630-d5c1-4eed-8232-3698ab99fc19\") " Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.252344 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/51e27630-d5c1-4eed-8232-3698ab99fc19-client-ca\") pod \"51e27630-d5c1-4eed-8232-3698ab99fc19\" (UID: \"51e27630-d5c1-4eed-8232-3698ab99fc19\") " Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.252389 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dg7z5\" (UniqueName: \"kubernetes.io/projected/d962d0be-438a-4d08-8113-525238fa2156-kube-api-access-dg7z5\") pod \"d962d0be-438a-4d08-8113-525238fa2156\" (UID: \"d962d0be-438a-4d08-8113-525238fa2156\") " Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.252410 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/51e27630-d5c1-4eed-8232-3698ab99fc19-serving-cert\") pod \"51e27630-d5c1-4eed-8232-3698ab99fc19\" (UID: \"51e27630-d5c1-4eed-8232-3698ab99fc19\") " Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.252453 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d962d0be-438a-4d08-8113-525238fa2156-client-ca\") pod \"d962d0be-438a-4d08-8113-525238fa2156\" (UID: \"d962d0be-438a-4d08-8113-525238fa2156\") " Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.252499 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/51e27630-d5c1-4eed-8232-3698ab99fc19-config\") pod \"51e27630-d5c1-4eed-8232-3698ab99fc19\" (UID: \"51e27630-d5c1-4eed-8232-3698ab99fc19\") " Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.252521 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d962d0be-438a-4d08-8113-525238fa2156-proxy-ca-bundles\") pod \"d962d0be-438a-4d08-8113-525238fa2156\" (UID: \"d962d0be-438a-4d08-8113-525238fa2156\") " Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.252544 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d962d0be-438a-4d08-8113-525238fa2156-config\") pod \"d962d0be-438a-4d08-8113-525238fa2156\" (UID: \"d962d0be-438a-4d08-8113-525238fa2156\") " Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.252881 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/51e27630-d5c1-4eed-8232-3698ab99fc19-client-ca" (OuterVolumeSpecName: "client-ca") pod "51e27630-d5c1-4eed-8232-3698ab99fc19" (UID: "51e27630-d5c1-4eed-8232-3698ab99fc19"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.253402 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d962d0be-438a-4d08-8113-525238fa2156-client-ca" (OuterVolumeSpecName: "client-ca") pod "d962d0be-438a-4d08-8113-525238fa2156" (UID: "d962d0be-438a-4d08-8113-525238fa2156"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.253410 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d962d0be-438a-4d08-8113-525238fa2156-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "d962d0be-438a-4d08-8113-525238fa2156" (UID: "d962d0be-438a-4d08-8113-525238fa2156"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.253436 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/51e27630-d5c1-4eed-8232-3698ab99fc19-config" (OuterVolumeSpecName: "config") pod "51e27630-d5c1-4eed-8232-3698ab99fc19" (UID: "51e27630-d5c1-4eed-8232-3698ab99fc19"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.253541 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d962d0be-438a-4d08-8113-525238fa2156-config" (OuterVolumeSpecName: "config") pod "d962d0be-438a-4d08-8113-525238fa2156" (UID: "d962d0be-438a-4d08-8113-525238fa2156"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.256782 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d962d0be-438a-4d08-8113-525238fa2156-kube-api-access-dg7z5" (OuterVolumeSpecName: "kube-api-access-dg7z5") pod "d962d0be-438a-4d08-8113-525238fa2156" (UID: "d962d0be-438a-4d08-8113-525238fa2156"). InnerVolumeSpecName "kube-api-access-dg7z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.256803 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/51e27630-d5c1-4eed-8232-3698ab99fc19-kube-api-access-5ksj6" (OuterVolumeSpecName: "kube-api-access-5ksj6") pod "51e27630-d5c1-4eed-8232-3698ab99fc19" (UID: "51e27630-d5c1-4eed-8232-3698ab99fc19"). InnerVolumeSpecName "kube-api-access-5ksj6". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.256789 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/51e27630-d5c1-4eed-8232-3698ab99fc19-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "51e27630-d5c1-4eed-8232-3698ab99fc19" (UID: "51e27630-d5c1-4eed-8232-3698ab99fc19"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.257112 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d962d0be-438a-4d08-8113-525238fa2156-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "d962d0be-438a-4d08-8113-525238fa2156" (UID: "d962d0be-438a-4d08-8113-525238fa2156"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.353659 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/51e27630-d5c1-4eed-8232-3698ab99fc19-config\") on node \"crc\" DevicePath \"\"" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.353695 4824 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d962d0be-438a-4d08-8113-525238fa2156-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.353708 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d962d0be-438a-4d08-8113-525238fa2156-config\") on node \"crc\" DevicePath \"\"" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.353716 4824 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d962d0be-438a-4d08-8113-525238fa2156-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.353726 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5ksj6\" (UniqueName: \"kubernetes.io/projected/51e27630-d5c1-4eed-8232-3698ab99fc19-kube-api-access-5ksj6\") on node \"crc\" DevicePath \"\"" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.353734 4824 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/51e27630-d5c1-4eed-8232-3698ab99fc19-client-ca\") on node \"crc\" DevicePath \"\"" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.353757 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dg7z5\" (UniqueName: \"kubernetes.io/projected/d962d0be-438a-4d08-8113-525238fa2156-kube-api-access-dg7z5\") on node \"crc\" DevicePath \"\"" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.353765 4824 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/51e27630-d5c1-4eed-8232-3698ab99fc19-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.353772 4824 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d962d0be-438a-4d08-8113-525238fa2156-client-ca\") on node \"crc\" DevicePath \"\"" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.462547 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-cmbbh"] Jan 21 11:15:48 crc kubenswrapper[4824]: E0121 11:15:48.462726 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d962d0be-438a-4d08-8113-525238fa2156" containerName="controller-manager" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.462738 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="d962d0be-438a-4d08-8113-525238fa2156" containerName="controller-manager" Jan 21 11:15:48 crc kubenswrapper[4824]: E0121 11:15:48.462749 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14d99a1f-fca3-421a-8b47-ab4cda2f77fc" containerName="marketplace-operator" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.462754 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="14d99a1f-fca3-421a-8b47-ab4cda2f77fc" containerName="marketplace-operator" Jan 21 11:15:48 crc kubenswrapper[4824]: E0121 11:15:48.462761 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1bc8c652-fb4f-48b7-86ce-c45522dfd476" containerName="extract-content" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.462767 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="1bc8c652-fb4f-48b7-86ce-c45522dfd476" containerName="extract-content" Jan 21 11:15:48 crc kubenswrapper[4824]: E0121 11:15:48.462774 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1bc8c652-fb4f-48b7-86ce-c45522dfd476" containerName="registry-server" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.462779 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="1bc8c652-fb4f-48b7-86ce-c45522dfd476" containerName="registry-server" Jan 21 11:15:48 crc kubenswrapper[4824]: E0121 11:15:48.462786 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f11911f-e365-4151-b536-3988186b8757" containerName="extract-content" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.462791 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f11911f-e365-4151-b536-3988186b8757" containerName="extract-content" Jan 21 11:15:48 crc kubenswrapper[4824]: E0121 11:15:48.462799 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1bc8c652-fb4f-48b7-86ce-c45522dfd476" containerName="extract-utilities" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.462803 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="1bc8c652-fb4f-48b7-86ce-c45522dfd476" containerName="extract-utilities" Jan 21 11:15:48 crc kubenswrapper[4824]: E0121 11:15:48.462808 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f11911f-e365-4151-b536-3988186b8757" containerName="registry-server" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.462813 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f11911f-e365-4151-b536-3988186b8757" containerName="registry-server" Jan 21 11:15:48 crc kubenswrapper[4824]: E0121 11:15:48.462820 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c3b56f2d-96c0-4597-9235-147fc4802d19" containerName="extract-content" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.462825 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="c3b56f2d-96c0-4597-9235-147fc4802d19" containerName="extract-content" Jan 21 11:15:48 crc kubenswrapper[4824]: E0121 11:15:48.462830 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.462835 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Jan 21 11:15:48 crc kubenswrapper[4824]: E0121 11:15:48.462842 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f11911f-e365-4151-b536-3988186b8757" containerName="extract-utilities" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.462847 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f11911f-e365-4151-b536-3988186b8757" containerName="extract-utilities" Jan 21 11:15:48 crc kubenswrapper[4824]: E0121 11:15:48.462856 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="123c89e0-3113-4b28-9129-3848a7288052" containerName="collect-profiles" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.462861 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="123c89e0-3113-4b28-9129-3848a7288052" containerName="collect-profiles" Jan 21 11:15:48 crc kubenswrapper[4824]: E0121 11:15:48.462868 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b040d3d-f4a0-451a-97de-08bea41c6dce" containerName="registry-server" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.462873 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b040d3d-f4a0-451a-97de-08bea41c6dce" containerName="registry-server" Jan 21 11:15:48 crc kubenswrapper[4824]: E0121 11:15:48.462880 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="51e27630-d5c1-4eed-8232-3698ab99fc19" containerName="route-controller-manager" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.462884 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="51e27630-d5c1-4eed-8232-3698ab99fc19" containerName="route-controller-manager" Jan 21 11:15:48 crc kubenswrapper[4824]: E0121 11:15:48.462893 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b040d3d-f4a0-451a-97de-08bea41c6dce" containerName="extract-utilities" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.462897 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b040d3d-f4a0-451a-97de-08bea41c6dce" containerName="extract-utilities" Jan 21 11:15:48 crc kubenswrapper[4824]: E0121 11:15:48.462906 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c3b56f2d-96c0-4597-9235-147fc4802d19" containerName="registry-server" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.462912 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="c3b56f2d-96c0-4597-9235-147fc4802d19" containerName="registry-server" Jan 21 11:15:48 crc kubenswrapper[4824]: E0121 11:15:48.462918 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c3b56f2d-96c0-4597-9235-147fc4802d19" containerName="extract-utilities" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.462924 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="c3b56f2d-96c0-4597-9235-147fc4802d19" containerName="extract-utilities" Jan 21 11:15:48 crc kubenswrapper[4824]: E0121 11:15:48.462931 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b040d3d-f4a0-451a-97de-08bea41c6dce" containerName="extract-content" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.462936 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b040d3d-f4a0-451a-97de-08bea41c6dce" containerName="extract-content" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.463056 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="123c89e0-3113-4b28-9129-3848a7288052" containerName="collect-profiles" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.463066 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="c3b56f2d-96c0-4597-9235-147fc4802d19" containerName="registry-server" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.463074 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="51e27630-d5c1-4eed-8232-3698ab99fc19" containerName="route-controller-manager" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.463081 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.463087 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="14d99a1f-fca3-421a-8b47-ab4cda2f77fc" containerName="marketplace-operator" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.463094 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="1b040d3d-f4a0-451a-97de-08bea41c6dce" containerName="registry-server" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.463098 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="1bc8c652-fb4f-48b7-86ce-c45522dfd476" containerName="registry-server" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.463105 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="d962d0be-438a-4d08-8113-525238fa2156" containerName="controller-manager" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.463111 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="3f11911f-e365-4151-b536-3988186b8757" containerName="registry-server" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.463663 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-cmbbh" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.466212 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.469258 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-cmbbh"] Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.625775 4824 generic.go:334] "Generic (PLEG): container finished" podID="51e27630-d5c1-4eed-8232-3698ab99fc19" containerID="e407f561a94b91929f1905dd35fb9fce0694555be001231861d30f794a254728" exitCode=0 Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.625835 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qs8d6" event={"ID":"51e27630-d5c1-4eed-8232-3698ab99fc19","Type":"ContainerDied","Data":"e407f561a94b91929f1905dd35fb9fce0694555be001231861d30f794a254728"} Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.625840 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qs8d6" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.625858 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qs8d6" event={"ID":"51e27630-d5c1-4eed-8232-3698ab99fc19","Type":"ContainerDied","Data":"66caec7ee49f045aa9a142ffed3e9d86fa441cb0eec919f264b1bbebed641a83"} Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.625874 4824 scope.go:117] "RemoveContainer" containerID="e407f561a94b91929f1905dd35fb9fce0694555be001231861d30f794a254728" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.627759 4824 generic.go:334] "Generic (PLEG): container finished" podID="d962d0be-438a-4d08-8113-525238fa2156" containerID="13eb81088f05a509df912f982b6e26892b6c0fee8b275279ea88e029d704d428" exitCode=0 Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.627821 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-4dkd8" event={"ID":"d962d0be-438a-4d08-8113-525238fa2156","Type":"ContainerDied","Data":"13eb81088f05a509df912f982b6e26892b6c0fee8b275279ea88e029d704d428"} Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.627849 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-4dkd8" event={"ID":"d962d0be-438a-4d08-8113-525238fa2156","Type":"ContainerDied","Data":"90b18225419fa36b91ae49d12b220ac91e071e5928e03c8bbf47f6f6ddb46af0"} Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.627849 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-4dkd8" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.638318 4824 scope.go:117] "RemoveContainer" containerID="e407f561a94b91929f1905dd35fb9fce0694555be001231861d30f794a254728" Jan 21 11:15:48 crc kubenswrapper[4824]: E0121 11:15:48.638596 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e407f561a94b91929f1905dd35fb9fce0694555be001231861d30f794a254728\": container with ID starting with e407f561a94b91929f1905dd35fb9fce0694555be001231861d30f794a254728 not found: ID does not exist" containerID="e407f561a94b91929f1905dd35fb9fce0694555be001231861d30f794a254728" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.638624 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e407f561a94b91929f1905dd35fb9fce0694555be001231861d30f794a254728"} err="failed to get container status \"e407f561a94b91929f1905dd35fb9fce0694555be001231861d30f794a254728\": rpc error: code = NotFound desc = could not find container \"e407f561a94b91929f1905dd35fb9fce0694555be001231861d30f794a254728\": container with ID starting with e407f561a94b91929f1905dd35fb9fce0694555be001231861d30f794a254728 not found: ID does not exist" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.638644 4824 scope.go:117] "RemoveContainer" containerID="13eb81088f05a509df912f982b6e26892b6c0fee8b275279ea88e029d704d428" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.648857 4824 scope.go:117] "RemoveContainer" containerID="13eb81088f05a509df912f982b6e26892b6c0fee8b275279ea88e029d704d428" Jan 21 11:15:48 crc kubenswrapper[4824]: E0121 11:15:48.649162 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"13eb81088f05a509df912f982b6e26892b6c0fee8b275279ea88e029d704d428\": container with ID starting with 13eb81088f05a509df912f982b6e26892b6c0fee8b275279ea88e029d704d428 not found: ID does not exist" containerID="13eb81088f05a509df912f982b6e26892b6c0fee8b275279ea88e029d704d428" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.649192 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"13eb81088f05a509df912f982b6e26892b6c0fee8b275279ea88e029d704d428"} err="failed to get container status \"13eb81088f05a509df912f982b6e26892b6c0fee8b275279ea88e029d704d428\": rpc error: code = NotFound desc = could not find container \"13eb81088f05a509df912f982b6e26892b6c0fee8b275279ea88e029d704d428\": container with ID starting with 13eb81088f05a509df912f982b6e26892b6c0fee8b275279ea88e029d704d428 not found: ID does not exist" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.649573 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-4dkd8"] Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.652446 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-4dkd8"] Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.655803 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-qs8d6"] Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.657284 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wwd7d\" (UniqueName: \"kubernetes.io/projected/3447df50-807b-487c-9acf-2a2250780bdc-kube-api-access-wwd7d\") pod \"redhat-operators-cmbbh\" (UID: \"3447df50-807b-487c-9acf-2a2250780bdc\") " pod="openshift-marketplace/redhat-operators-cmbbh" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.657315 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3447df50-807b-487c-9acf-2a2250780bdc-utilities\") pod \"redhat-operators-cmbbh\" (UID: \"3447df50-807b-487c-9acf-2a2250780bdc\") " pod="openshift-marketplace/redhat-operators-cmbbh" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.657350 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3447df50-807b-487c-9acf-2a2250780bdc-catalog-content\") pod \"redhat-operators-cmbbh\" (UID: \"3447df50-807b-487c-9acf-2a2250780bdc\") " pod="openshift-marketplace/redhat-operators-cmbbh" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.658589 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-qs8d6"] Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.660936 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-m4srg"] Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.661700 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-m4srg" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.663105 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.666567 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-m4srg"] Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.758165 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wwd7d\" (UniqueName: \"kubernetes.io/projected/3447df50-807b-487c-9acf-2a2250780bdc-kube-api-access-wwd7d\") pod \"redhat-operators-cmbbh\" (UID: \"3447df50-807b-487c-9acf-2a2250780bdc\") " pod="openshift-marketplace/redhat-operators-cmbbh" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.758204 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/53e1053f-f40b-4afc-ad9c-b07d46fc1f94-catalog-content\") pod \"community-operators-m4srg\" (UID: \"53e1053f-f40b-4afc-ad9c-b07d46fc1f94\") " pod="openshift-marketplace/community-operators-m4srg" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.758224 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-49dkh\" (UniqueName: \"kubernetes.io/projected/53e1053f-f40b-4afc-ad9c-b07d46fc1f94-kube-api-access-49dkh\") pod \"community-operators-m4srg\" (UID: \"53e1053f-f40b-4afc-ad9c-b07d46fc1f94\") " pod="openshift-marketplace/community-operators-m4srg" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.758242 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3447df50-807b-487c-9acf-2a2250780bdc-utilities\") pod \"redhat-operators-cmbbh\" (UID: \"3447df50-807b-487c-9acf-2a2250780bdc\") " pod="openshift-marketplace/redhat-operators-cmbbh" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.758269 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3447df50-807b-487c-9acf-2a2250780bdc-catalog-content\") pod \"redhat-operators-cmbbh\" (UID: \"3447df50-807b-487c-9acf-2a2250780bdc\") " pod="openshift-marketplace/redhat-operators-cmbbh" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.758346 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/53e1053f-f40b-4afc-ad9c-b07d46fc1f94-utilities\") pod \"community-operators-m4srg\" (UID: \"53e1053f-f40b-4afc-ad9c-b07d46fc1f94\") " pod="openshift-marketplace/community-operators-m4srg" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.758733 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3447df50-807b-487c-9acf-2a2250780bdc-catalog-content\") pod \"redhat-operators-cmbbh\" (UID: \"3447df50-807b-487c-9acf-2a2250780bdc\") " pod="openshift-marketplace/redhat-operators-cmbbh" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.758769 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3447df50-807b-487c-9acf-2a2250780bdc-utilities\") pod \"redhat-operators-cmbbh\" (UID: \"3447df50-807b-487c-9acf-2a2250780bdc\") " pod="openshift-marketplace/redhat-operators-cmbbh" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.772571 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wwd7d\" (UniqueName: \"kubernetes.io/projected/3447df50-807b-487c-9acf-2a2250780bdc-kube-api-access-wwd7d\") pod \"redhat-operators-cmbbh\" (UID: \"3447df50-807b-487c-9acf-2a2250780bdc\") " pod="openshift-marketplace/redhat-operators-cmbbh" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.774007 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-cmbbh" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.864110 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-49dkh\" (UniqueName: \"kubernetes.io/projected/53e1053f-f40b-4afc-ad9c-b07d46fc1f94-kube-api-access-49dkh\") pod \"community-operators-m4srg\" (UID: \"53e1053f-f40b-4afc-ad9c-b07d46fc1f94\") " pod="openshift-marketplace/community-operators-m4srg" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.864150 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/53e1053f-f40b-4afc-ad9c-b07d46fc1f94-catalog-content\") pod \"community-operators-m4srg\" (UID: \"53e1053f-f40b-4afc-ad9c-b07d46fc1f94\") " pod="openshift-marketplace/community-operators-m4srg" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.864259 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/53e1053f-f40b-4afc-ad9c-b07d46fc1f94-utilities\") pod \"community-operators-m4srg\" (UID: \"53e1053f-f40b-4afc-ad9c-b07d46fc1f94\") " pod="openshift-marketplace/community-operators-m4srg" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.864627 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/53e1053f-f40b-4afc-ad9c-b07d46fc1f94-utilities\") pod \"community-operators-m4srg\" (UID: \"53e1053f-f40b-4afc-ad9c-b07d46fc1f94\") " pod="openshift-marketplace/community-operators-m4srg" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.864721 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/53e1053f-f40b-4afc-ad9c-b07d46fc1f94-catalog-content\") pod \"community-operators-m4srg\" (UID: \"53e1053f-f40b-4afc-ad9c-b07d46fc1f94\") " pod="openshift-marketplace/community-operators-m4srg" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.878346 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-49dkh\" (UniqueName: \"kubernetes.io/projected/53e1053f-f40b-4afc-ad9c-b07d46fc1f94-kube-api-access-49dkh\") pod \"community-operators-m4srg\" (UID: \"53e1053f-f40b-4afc-ad9c-b07d46fc1f94\") " pod="openshift-marketplace/community-operators-m4srg" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.900320 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-cmbbh"] Jan 21 11:15:48 crc kubenswrapper[4824]: W0121 11:15:48.903246 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3447df50_807b_487c_9acf_2a2250780bdc.slice/crio-8181bda55ac71db98645e6125503975812b2fa3fdcd76d88764acf074b860076 WatchSource:0}: Error finding container 8181bda55ac71db98645e6125503975812b2fa3fdcd76d88764acf074b860076: Status 404 returned error can't find the container with id 8181bda55ac71db98645e6125503975812b2fa3fdcd76d88764acf074b860076 Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.913820 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-7c9b544d8-b9cfc"] Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.914390 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7c9b544d8-b9cfc" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.916721 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.917050 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.917315 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.917539 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.917586 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.917944 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.922106 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.924629 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-7c9b544d8-b9cfc"] Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.966665 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/f32d32e5-b286-475a-bda8-5d79b746e576-proxy-ca-bundles\") pod \"controller-manager-7c9b544d8-b9cfc\" (UID: \"f32d32e5-b286-475a-bda8-5d79b746e576\") " pod="openshift-controller-manager/controller-manager-7c9b544d8-b9cfc" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.966901 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g6d2t\" (UniqueName: \"kubernetes.io/projected/f32d32e5-b286-475a-bda8-5d79b746e576-kube-api-access-g6d2t\") pod \"controller-manager-7c9b544d8-b9cfc\" (UID: \"f32d32e5-b286-475a-bda8-5d79b746e576\") " pod="openshift-controller-manager/controller-manager-7c9b544d8-b9cfc" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.966933 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f32d32e5-b286-475a-bda8-5d79b746e576-client-ca\") pod \"controller-manager-7c9b544d8-b9cfc\" (UID: \"f32d32e5-b286-475a-bda8-5d79b746e576\") " pod="openshift-controller-manager/controller-manager-7c9b544d8-b9cfc" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.966993 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f32d32e5-b286-475a-bda8-5d79b746e576-config\") pod \"controller-manager-7c9b544d8-b9cfc\" (UID: \"f32d32e5-b286-475a-bda8-5d79b746e576\") " pod="openshift-controller-manager/controller-manager-7c9b544d8-b9cfc" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.967059 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f32d32e5-b286-475a-bda8-5d79b746e576-serving-cert\") pod \"controller-manager-7c9b544d8-b9cfc\" (UID: \"f32d32e5-b286-475a-bda8-5d79b746e576\") " pod="openshift-controller-manager/controller-manager-7c9b544d8-b9cfc" Jan 21 11:15:48 crc kubenswrapper[4824]: I0121 11:15:48.974737 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-m4srg" Jan 21 11:15:49 crc kubenswrapper[4824]: I0121 11:15:49.068120 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g6d2t\" (UniqueName: \"kubernetes.io/projected/f32d32e5-b286-475a-bda8-5d79b746e576-kube-api-access-g6d2t\") pod \"controller-manager-7c9b544d8-b9cfc\" (UID: \"f32d32e5-b286-475a-bda8-5d79b746e576\") " pod="openshift-controller-manager/controller-manager-7c9b544d8-b9cfc" Jan 21 11:15:49 crc kubenswrapper[4824]: I0121 11:15:49.068198 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f32d32e5-b286-475a-bda8-5d79b746e576-client-ca\") pod \"controller-manager-7c9b544d8-b9cfc\" (UID: \"f32d32e5-b286-475a-bda8-5d79b746e576\") " pod="openshift-controller-manager/controller-manager-7c9b544d8-b9cfc" Jan 21 11:15:49 crc kubenswrapper[4824]: I0121 11:15:49.068237 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f32d32e5-b286-475a-bda8-5d79b746e576-config\") pod \"controller-manager-7c9b544d8-b9cfc\" (UID: \"f32d32e5-b286-475a-bda8-5d79b746e576\") " pod="openshift-controller-manager/controller-manager-7c9b544d8-b9cfc" Jan 21 11:15:49 crc kubenswrapper[4824]: I0121 11:15:49.068316 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f32d32e5-b286-475a-bda8-5d79b746e576-serving-cert\") pod \"controller-manager-7c9b544d8-b9cfc\" (UID: \"f32d32e5-b286-475a-bda8-5d79b746e576\") " pod="openshift-controller-manager/controller-manager-7c9b544d8-b9cfc" Jan 21 11:15:49 crc kubenswrapper[4824]: I0121 11:15:49.068354 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/f32d32e5-b286-475a-bda8-5d79b746e576-proxy-ca-bundles\") pod \"controller-manager-7c9b544d8-b9cfc\" (UID: \"f32d32e5-b286-475a-bda8-5d79b746e576\") " pod="openshift-controller-manager/controller-manager-7c9b544d8-b9cfc" Jan 21 11:15:49 crc kubenswrapper[4824]: I0121 11:15:49.069155 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f32d32e5-b286-475a-bda8-5d79b746e576-client-ca\") pod \"controller-manager-7c9b544d8-b9cfc\" (UID: \"f32d32e5-b286-475a-bda8-5d79b746e576\") " pod="openshift-controller-manager/controller-manager-7c9b544d8-b9cfc" Jan 21 11:15:49 crc kubenswrapper[4824]: I0121 11:15:49.069500 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f32d32e5-b286-475a-bda8-5d79b746e576-config\") pod \"controller-manager-7c9b544d8-b9cfc\" (UID: \"f32d32e5-b286-475a-bda8-5d79b746e576\") " pod="openshift-controller-manager/controller-manager-7c9b544d8-b9cfc" Jan 21 11:15:49 crc kubenswrapper[4824]: I0121 11:15:49.069616 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/f32d32e5-b286-475a-bda8-5d79b746e576-proxy-ca-bundles\") pod \"controller-manager-7c9b544d8-b9cfc\" (UID: \"f32d32e5-b286-475a-bda8-5d79b746e576\") " pod="openshift-controller-manager/controller-manager-7c9b544d8-b9cfc" Jan 21 11:15:49 crc kubenswrapper[4824]: I0121 11:15:49.072074 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f32d32e5-b286-475a-bda8-5d79b746e576-serving-cert\") pod \"controller-manager-7c9b544d8-b9cfc\" (UID: \"f32d32e5-b286-475a-bda8-5d79b746e576\") " pod="openshift-controller-manager/controller-manager-7c9b544d8-b9cfc" Jan 21 11:15:49 crc kubenswrapper[4824]: I0121 11:15:49.083179 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g6d2t\" (UniqueName: \"kubernetes.io/projected/f32d32e5-b286-475a-bda8-5d79b746e576-kube-api-access-g6d2t\") pod \"controller-manager-7c9b544d8-b9cfc\" (UID: \"f32d32e5-b286-475a-bda8-5d79b746e576\") " pod="openshift-controller-manager/controller-manager-7c9b544d8-b9cfc" Jan 21 11:15:49 crc kubenswrapper[4824]: I0121 11:15:49.230078 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7c9b544d8-b9cfc" Jan 21 11:15:49 crc kubenswrapper[4824]: I0121 11:15:49.305799 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-m4srg"] Jan 21 11:15:49 crc kubenswrapper[4824]: W0121 11:15:49.320229 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod53e1053f_f40b_4afc_ad9c_b07d46fc1f94.slice/crio-311aec34001f914806b85fb01bbd4769449448d553c52662486e8aa5470b5438 WatchSource:0}: Error finding container 311aec34001f914806b85fb01bbd4769449448d553c52662486e8aa5470b5438: Status 404 returned error can't find the container with id 311aec34001f914806b85fb01bbd4769449448d553c52662486e8aa5470b5438 Jan 21 11:15:49 crc kubenswrapper[4824]: I0121 11:15:49.562294 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-7c9b544d8-b9cfc"] Jan 21 11:15:49 crc kubenswrapper[4824]: W0121 11:15:49.567633 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf32d32e5_b286_475a_bda8_5d79b746e576.slice/crio-41f5c0a541981a8fe31ce891bdb4387c75d383bf8c055b1b0de1ae7af35b3bcd WatchSource:0}: Error finding container 41f5c0a541981a8fe31ce891bdb4387c75d383bf8c055b1b0de1ae7af35b3bcd: Status 404 returned error can't find the container with id 41f5c0a541981a8fe31ce891bdb4387c75d383bf8c055b1b0de1ae7af35b3bcd Jan 21 11:15:49 crc kubenswrapper[4824]: I0121 11:15:49.628149 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-7c9b544d8-b9cfc"] Jan 21 11:15:49 crc kubenswrapper[4824]: I0121 11:15:49.632430 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7c9b544d8-b9cfc" event={"ID":"f32d32e5-b286-475a-bda8-5d79b746e576","Type":"ContainerStarted","Data":"41f5c0a541981a8fe31ce891bdb4387c75d383bf8c055b1b0de1ae7af35b3bcd"} Jan 21 11:15:49 crc kubenswrapper[4824]: I0121 11:15:49.633992 4824 generic.go:334] "Generic (PLEG): container finished" podID="53e1053f-f40b-4afc-ad9c-b07d46fc1f94" containerID="d2d9f16df9dc20450c70352d5b5dfc3d357cc5eac304802f73897c9f9f00a93c" exitCode=0 Jan 21 11:15:49 crc kubenswrapper[4824]: I0121 11:15:49.634028 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-m4srg" event={"ID":"53e1053f-f40b-4afc-ad9c-b07d46fc1f94","Type":"ContainerDied","Data":"d2d9f16df9dc20450c70352d5b5dfc3d357cc5eac304802f73897c9f9f00a93c"} Jan 21 11:15:49 crc kubenswrapper[4824]: I0121 11:15:49.634065 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-m4srg" event={"ID":"53e1053f-f40b-4afc-ad9c-b07d46fc1f94","Type":"ContainerStarted","Data":"311aec34001f914806b85fb01bbd4769449448d553c52662486e8aa5470b5438"} Jan 21 11:15:49 crc kubenswrapper[4824]: I0121 11:15:49.637854 4824 generic.go:334] "Generic (PLEG): container finished" podID="3447df50-807b-487c-9acf-2a2250780bdc" containerID="729c6703ff41d585ba645a84b7f8de78c214ea811142e302140468d9881567aa" exitCode=0 Jan 21 11:15:49 crc kubenswrapper[4824]: I0121 11:15:49.637912 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cmbbh" event={"ID":"3447df50-807b-487c-9acf-2a2250780bdc","Type":"ContainerDied","Data":"729c6703ff41d585ba645a84b7f8de78c214ea811142e302140468d9881567aa"} Jan 21 11:15:49 crc kubenswrapper[4824]: I0121 11:15:49.637968 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cmbbh" event={"ID":"3447df50-807b-487c-9acf-2a2250780bdc","Type":"ContainerStarted","Data":"8181bda55ac71db98645e6125503975812b2fa3fdcd76d88764acf074b860076"} Jan 21 11:15:49 crc kubenswrapper[4824]: I0121 11:15:49.657840 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5fdb5cfd56-pqprn"] Jan 21 11:15:49 crc kubenswrapper[4824]: I0121 11:15:49.658430 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5fdb5cfd56-pqprn" Jan 21 11:15:49 crc kubenswrapper[4824]: I0121 11:15:49.661380 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Jan 21 11:15:49 crc kubenswrapper[4824]: I0121 11:15:49.661445 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Jan 21 11:15:49 crc kubenswrapper[4824]: I0121 11:15:49.661593 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Jan 21 11:15:49 crc kubenswrapper[4824]: I0121 11:15:49.661686 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Jan 21 11:15:49 crc kubenswrapper[4824]: I0121 11:15:49.661859 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Jan 21 11:15:49 crc kubenswrapper[4824]: I0121 11:15:49.672296 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5fdb5cfd56-pqprn"] Jan 21 11:15:49 crc kubenswrapper[4824]: I0121 11:15:49.672396 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Jan 21 11:15:49 crc kubenswrapper[4824]: I0121 11:15:49.718334 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5fdb5cfd56-pqprn"] Jan 21 11:15:49 crc kubenswrapper[4824]: E0121 11:15:49.718645 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[client-ca config kube-api-access-5xz7t serving-cert], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openshift-route-controller-manager/route-controller-manager-5fdb5cfd56-pqprn" podUID="66af07e9-d9e9-4584-b458-afd995ac1eb3" Jan 21 11:15:49 crc kubenswrapper[4824]: I0121 11:15:49.774500 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/66af07e9-d9e9-4584-b458-afd995ac1eb3-client-ca\") pod \"route-controller-manager-5fdb5cfd56-pqprn\" (UID: \"66af07e9-d9e9-4584-b458-afd995ac1eb3\") " pod="openshift-route-controller-manager/route-controller-manager-5fdb5cfd56-pqprn" Jan 21 11:15:49 crc kubenswrapper[4824]: I0121 11:15:49.774544 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5xz7t\" (UniqueName: \"kubernetes.io/projected/66af07e9-d9e9-4584-b458-afd995ac1eb3-kube-api-access-5xz7t\") pod \"route-controller-manager-5fdb5cfd56-pqprn\" (UID: \"66af07e9-d9e9-4584-b458-afd995ac1eb3\") " pod="openshift-route-controller-manager/route-controller-manager-5fdb5cfd56-pqprn" Jan 21 11:15:49 crc kubenswrapper[4824]: I0121 11:15:49.774680 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/66af07e9-d9e9-4584-b458-afd995ac1eb3-config\") pod \"route-controller-manager-5fdb5cfd56-pqprn\" (UID: \"66af07e9-d9e9-4584-b458-afd995ac1eb3\") " pod="openshift-route-controller-manager/route-controller-manager-5fdb5cfd56-pqprn" Jan 21 11:15:49 crc kubenswrapper[4824]: I0121 11:15:49.774704 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/66af07e9-d9e9-4584-b458-afd995ac1eb3-serving-cert\") pod \"route-controller-manager-5fdb5cfd56-pqprn\" (UID: \"66af07e9-d9e9-4584-b458-afd995ac1eb3\") " pod="openshift-route-controller-manager/route-controller-manager-5fdb5cfd56-pqprn" Jan 21 11:15:49 crc kubenswrapper[4824]: I0121 11:15:49.875344 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/66af07e9-d9e9-4584-b458-afd995ac1eb3-client-ca\") pod \"route-controller-manager-5fdb5cfd56-pqprn\" (UID: \"66af07e9-d9e9-4584-b458-afd995ac1eb3\") " pod="openshift-route-controller-manager/route-controller-manager-5fdb5cfd56-pqprn" Jan 21 11:15:49 crc kubenswrapper[4824]: I0121 11:15:49.876058 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/66af07e9-d9e9-4584-b458-afd995ac1eb3-client-ca\") pod \"route-controller-manager-5fdb5cfd56-pqprn\" (UID: \"66af07e9-d9e9-4584-b458-afd995ac1eb3\") " pod="openshift-route-controller-manager/route-controller-manager-5fdb5cfd56-pqprn" Jan 21 11:15:49 crc kubenswrapper[4824]: I0121 11:15:49.876119 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5xz7t\" (UniqueName: \"kubernetes.io/projected/66af07e9-d9e9-4584-b458-afd995ac1eb3-kube-api-access-5xz7t\") pod \"route-controller-manager-5fdb5cfd56-pqprn\" (UID: \"66af07e9-d9e9-4584-b458-afd995ac1eb3\") " pod="openshift-route-controller-manager/route-controller-manager-5fdb5cfd56-pqprn" Jan 21 11:15:49 crc kubenswrapper[4824]: I0121 11:15:49.876447 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/66af07e9-d9e9-4584-b458-afd995ac1eb3-config\") pod \"route-controller-manager-5fdb5cfd56-pqprn\" (UID: \"66af07e9-d9e9-4584-b458-afd995ac1eb3\") " pod="openshift-route-controller-manager/route-controller-manager-5fdb5cfd56-pqprn" Jan 21 11:15:49 crc kubenswrapper[4824]: I0121 11:15:49.877228 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/66af07e9-d9e9-4584-b458-afd995ac1eb3-config\") pod \"route-controller-manager-5fdb5cfd56-pqprn\" (UID: \"66af07e9-d9e9-4584-b458-afd995ac1eb3\") " pod="openshift-route-controller-manager/route-controller-manager-5fdb5cfd56-pqprn" Jan 21 11:15:49 crc kubenswrapper[4824]: I0121 11:15:49.876482 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/66af07e9-d9e9-4584-b458-afd995ac1eb3-serving-cert\") pod \"route-controller-manager-5fdb5cfd56-pqprn\" (UID: \"66af07e9-d9e9-4584-b458-afd995ac1eb3\") " pod="openshift-route-controller-manager/route-controller-manager-5fdb5cfd56-pqprn" Jan 21 11:15:49 crc kubenswrapper[4824]: I0121 11:15:49.880356 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/66af07e9-d9e9-4584-b458-afd995ac1eb3-serving-cert\") pod \"route-controller-manager-5fdb5cfd56-pqprn\" (UID: \"66af07e9-d9e9-4584-b458-afd995ac1eb3\") " pod="openshift-route-controller-manager/route-controller-manager-5fdb5cfd56-pqprn" Jan 21 11:15:49 crc kubenswrapper[4824]: I0121 11:15:49.891562 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5xz7t\" (UniqueName: \"kubernetes.io/projected/66af07e9-d9e9-4584-b458-afd995ac1eb3-kube-api-access-5xz7t\") pod \"route-controller-manager-5fdb5cfd56-pqprn\" (UID: \"66af07e9-d9e9-4584-b458-afd995ac1eb3\") " pod="openshift-route-controller-manager/route-controller-manager-5fdb5cfd56-pqprn" Jan 21 11:15:50 crc kubenswrapper[4824]: I0121 11:15:50.055987 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="51e27630-d5c1-4eed-8232-3698ab99fc19" path="/var/lib/kubelet/pods/51e27630-d5c1-4eed-8232-3698ab99fc19/volumes" Jan 21 11:15:50 crc kubenswrapper[4824]: I0121 11:15:50.056819 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d962d0be-438a-4d08-8113-525238fa2156" path="/var/lib/kubelet/pods/d962d0be-438a-4d08-8113-525238fa2156/volumes" Jan 21 11:15:50 crc kubenswrapper[4824]: I0121 11:15:50.647166 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cmbbh" event={"ID":"3447df50-807b-487c-9acf-2a2250780bdc","Type":"ContainerStarted","Data":"affad804ab6c5ef16e0c73a329f54a91b3bdbf2150e0045b7e2a45788bbf14f0"} Jan 21 11:15:50 crc kubenswrapper[4824]: I0121 11:15:50.648890 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7c9b544d8-b9cfc" event={"ID":"f32d32e5-b286-475a-bda8-5d79b746e576","Type":"ContainerStarted","Data":"7ca072731bf39a1cadc5fb56bba2c11e121f2b68f324f36479a7c2338c662422"} Jan 21 11:15:50 crc kubenswrapper[4824]: I0121 11:15:50.648913 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-7c9b544d8-b9cfc" podUID="f32d32e5-b286-475a-bda8-5d79b746e576" containerName="controller-manager" containerID="cri-o://7ca072731bf39a1cadc5fb56bba2c11e121f2b68f324f36479a7c2338c662422" gracePeriod=30 Jan 21 11:15:50 crc kubenswrapper[4824]: I0121 11:15:50.648948 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-7c9b544d8-b9cfc" Jan 21 11:15:50 crc kubenswrapper[4824]: I0121 11:15:50.651248 4824 generic.go:334] "Generic (PLEG): container finished" podID="53e1053f-f40b-4afc-ad9c-b07d46fc1f94" containerID="df6d0dcb0faad05a6d372141361f8ee5fdb7937db0906748cd5b6648219d66ba" exitCode=0 Jan 21 11:15:50 crc kubenswrapper[4824]: I0121 11:15:50.651304 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5fdb5cfd56-pqprn" Jan 21 11:15:50 crc kubenswrapper[4824]: I0121 11:15:50.651292 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-m4srg" event={"ID":"53e1053f-f40b-4afc-ad9c-b07d46fc1f94","Type":"ContainerDied","Data":"df6d0dcb0faad05a6d372141361f8ee5fdb7937db0906748cd5b6648219d66ba"} Jan 21 11:15:50 crc kubenswrapper[4824]: I0121 11:15:50.655682 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-7c9b544d8-b9cfc" Jan 21 11:15:50 crc kubenswrapper[4824]: I0121 11:15:50.657300 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5fdb5cfd56-pqprn" Jan 21 11:15:50 crc kubenswrapper[4824]: I0121 11:15:50.679549 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-7c9b544d8-b9cfc" podStartSLOduration=3.679532556 podStartE2EDuration="3.679532556s" podCreationTimestamp="2026-01-21 11:15:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:15:50.679000858 +0000 UTC m=+292.972030150" watchObservedRunningTime="2026-01-21 11:15:50.679532556 +0000 UTC m=+292.972561848" Jan 21 11:15:50 crc kubenswrapper[4824]: I0121 11:15:50.784690 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/66af07e9-d9e9-4584-b458-afd995ac1eb3-config\") pod \"66af07e9-d9e9-4584-b458-afd995ac1eb3\" (UID: \"66af07e9-d9e9-4584-b458-afd995ac1eb3\") " Jan 21 11:15:50 crc kubenswrapper[4824]: I0121 11:15:50.784761 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/66af07e9-d9e9-4584-b458-afd995ac1eb3-client-ca\") pod \"66af07e9-d9e9-4584-b458-afd995ac1eb3\" (UID: \"66af07e9-d9e9-4584-b458-afd995ac1eb3\") " Jan 21 11:15:50 crc kubenswrapper[4824]: I0121 11:15:50.784801 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/66af07e9-d9e9-4584-b458-afd995ac1eb3-serving-cert\") pod \"66af07e9-d9e9-4584-b458-afd995ac1eb3\" (UID: \"66af07e9-d9e9-4584-b458-afd995ac1eb3\") " Jan 21 11:15:50 crc kubenswrapper[4824]: I0121 11:15:50.784838 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5xz7t\" (UniqueName: \"kubernetes.io/projected/66af07e9-d9e9-4584-b458-afd995ac1eb3-kube-api-access-5xz7t\") pod \"66af07e9-d9e9-4584-b458-afd995ac1eb3\" (UID: \"66af07e9-d9e9-4584-b458-afd995ac1eb3\") " Jan 21 11:15:50 crc kubenswrapper[4824]: I0121 11:15:50.785254 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/66af07e9-d9e9-4584-b458-afd995ac1eb3-client-ca" (OuterVolumeSpecName: "client-ca") pod "66af07e9-d9e9-4584-b458-afd995ac1eb3" (UID: "66af07e9-d9e9-4584-b458-afd995ac1eb3"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:15:50 crc kubenswrapper[4824]: I0121 11:15:50.785288 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/66af07e9-d9e9-4584-b458-afd995ac1eb3-config" (OuterVolumeSpecName: "config") pod "66af07e9-d9e9-4584-b458-afd995ac1eb3" (UID: "66af07e9-d9e9-4584-b458-afd995ac1eb3"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:15:50 crc kubenswrapper[4824]: I0121 11:15:50.788477 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/66af07e9-d9e9-4584-b458-afd995ac1eb3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "66af07e9-d9e9-4584-b458-afd995ac1eb3" (UID: "66af07e9-d9e9-4584-b458-afd995ac1eb3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:15:50 crc kubenswrapper[4824]: I0121 11:15:50.789481 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/66af07e9-d9e9-4584-b458-afd995ac1eb3-kube-api-access-5xz7t" (OuterVolumeSpecName: "kube-api-access-5xz7t") pod "66af07e9-d9e9-4584-b458-afd995ac1eb3" (UID: "66af07e9-d9e9-4584-b458-afd995ac1eb3"). InnerVolumeSpecName "kube-api-access-5xz7t". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:15:50 crc kubenswrapper[4824]: I0121 11:15:50.886270 4824 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/66af07e9-d9e9-4584-b458-afd995ac1eb3-client-ca\") on node \"crc\" DevicePath \"\"" Jan 21 11:15:50 crc kubenswrapper[4824]: I0121 11:15:50.886303 4824 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/66af07e9-d9e9-4584-b458-afd995ac1eb3-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 21 11:15:50 crc kubenswrapper[4824]: I0121 11:15:50.886313 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5xz7t\" (UniqueName: \"kubernetes.io/projected/66af07e9-d9e9-4584-b458-afd995ac1eb3-kube-api-access-5xz7t\") on node \"crc\" DevicePath \"\"" Jan 21 11:15:50 crc kubenswrapper[4824]: I0121 11:15:50.886324 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/66af07e9-d9e9-4584-b458-afd995ac1eb3-config\") on node \"crc\" DevicePath \"\"" Jan 21 11:15:50 crc kubenswrapper[4824]: I0121 11:15:50.928092 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7c9b544d8-b9cfc" Jan 21 11:15:50 crc kubenswrapper[4824]: I0121 11:15:50.959036 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-76bbb855b9-ct4ct"] Jan 21 11:15:50 crc kubenswrapper[4824]: E0121 11:15:50.959235 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f32d32e5-b286-475a-bda8-5d79b746e576" containerName="controller-manager" Jan 21 11:15:50 crc kubenswrapper[4824]: I0121 11:15:50.959249 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="f32d32e5-b286-475a-bda8-5d79b746e576" containerName="controller-manager" Jan 21 11:15:50 crc kubenswrapper[4824]: I0121 11:15:50.959338 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="f32d32e5-b286-475a-bda8-5d79b746e576" containerName="controller-manager" Jan 21 11:15:50 crc kubenswrapper[4824]: I0121 11:15:50.959671 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-76bbb855b9-ct4ct" Jan 21 11:15:50 crc kubenswrapper[4824]: I0121 11:15:50.965523 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-76bbb855b9-ct4ct"] Jan 21 11:15:51 crc kubenswrapper[4824]: I0121 11:15:51.059435 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-rgl7w"] Jan 21 11:15:51 crc kubenswrapper[4824]: I0121 11:15:51.060267 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rgl7w" Jan 21 11:15:51 crc kubenswrapper[4824]: I0121 11:15:51.061798 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Jan 21 11:15:51 crc kubenswrapper[4824]: I0121 11:15:51.067029 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-rgl7w"] Jan 21 11:15:51 crc kubenswrapper[4824]: I0121 11:15:51.088499 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f32d32e5-b286-475a-bda8-5d79b746e576-config\") pod \"f32d32e5-b286-475a-bda8-5d79b746e576\" (UID: \"f32d32e5-b286-475a-bda8-5d79b746e576\") " Jan 21 11:15:51 crc kubenswrapper[4824]: I0121 11:15:51.088563 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f32d32e5-b286-475a-bda8-5d79b746e576-serving-cert\") pod \"f32d32e5-b286-475a-bda8-5d79b746e576\" (UID: \"f32d32e5-b286-475a-bda8-5d79b746e576\") " Jan 21 11:15:51 crc kubenswrapper[4824]: I0121 11:15:51.088621 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/f32d32e5-b286-475a-bda8-5d79b746e576-proxy-ca-bundles\") pod \"f32d32e5-b286-475a-bda8-5d79b746e576\" (UID: \"f32d32e5-b286-475a-bda8-5d79b746e576\") " Jan 21 11:15:51 crc kubenswrapper[4824]: I0121 11:15:51.088638 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g6d2t\" (UniqueName: \"kubernetes.io/projected/f32d32e5-b286-475a-bda8-5d79b746e576-kube-api-access-g6d2t\") pod \"f32d32e5-b286-475a-bda8-5d79b746e576\" (UID: \"f32d32e5-b286-475a-bda8-5d79b746e576\") " Jan 21 11:15:51 crc kubenswrapper[4824]: I0121 11:15:51.088680 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f32d32e5-b286-475a-bda8-5d79b746e576-client-ca\") pod \"f32d32e5-b286-475a-bda8-5d79b746e576\" (UID: \"f32d32e5-b286-475a-bda8-5d79b746e576\") " Jan 21 11:15:51 crc kubenswrapper[4824]: I0121 11:15:51.088838 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/3d23b1ac-adfc-4f40-8c75-710a71784b55-proxy-ca-bundles\") pod \"controller-manager-76bbb855b9-ct4ct\" (UID: \"3d23b1ac-adfc-4f40-8c75-710a71784b55\") " pod="openshift-controller-manager/controller-manager-76bbb855b9-ct4ct" Jan 21 11:15:51 crc kubenswrapper[4824]: I0121 11:15:51.088865 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jq6rh\" (UniqueName: \"kubernetes.io/projected/3d23b1ac-adfc-4f40-8c75-710a71784b55-kube-api-access-jq6rh\") pod \"controller-manager-76bbb855b9-ct4ct\" (UID: \"3d23b1ac-adfc-4f40-8c75-710a71784b55\") " pod="openshift-controller-manager/controller-manager-76bbb855b9-ct4ct" Jan 21 11:15:51 crc kubenswrapper[4824]: I0121 11:15:51.088916 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3d23b1ac-adfc-4f40-8c75-710a71784b55-client-ca\") pod \"controller-manager-76bbb855b9-ct4ct\" (UID: \"3d23b1ac-adfc-4f40-8c75-710a71784b55\") " pod="openshift-controller-manager/controller-manager-76bbb855b9-ct4ct" Jan 21 11:15:51 crc kubenswrapper[4824]: I0121 11:15:51.088953 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3d23b1ac-adfc-4f40-8c75-710a71784b55-config\") pod \"controller-manager-76bbb855b9-ct4ct\" (UID: \"3d23b1ac-adfc-4f40-8c75-710a71784b55\") " pod="openshift-controller-manager/controller-manager-76bbb855b9-ct4ct" Jan 21 11:15:51 crc kubenswrapper[4824]: I0121 11:15:51.089311 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f32d32e5-b286-475a-bda8-5d79b746e576-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "f32d32e5-b286-475a-bda8-5d79b746e576" (UID: "f32d32e5-b286-475a-bda8-5d79b746e576"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:15:51 crc kubenswrapper[4824]: I0121 11:15:51.089364 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f32d32e5-b286-475a-bda8-5d79b746e576-config" (OuterVolumeSpecName: "config") pod "f32d32e5-b286-475a-bda8-5d79b746e576" (UID: "f32d32e5-b286-475a-bda8-5d79b746e576"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:15:51 crc kubenswrapper[4824]: I0121 11:15:51.089477 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3d23b1ac-adfc-4f40-8c75-710a71784b55-serving-cert\") pod \"controller-manager-76bbb855b9-ct4ct\" (UID: \"3d23b1ac-adfc-4f40-8c75-710a71784b55\") " pod="openshift-controller-manager/controller-manager-76bbb855b9-ct4ct" Jan 21 11:15:51 crc kubenswrapper[4824]: I0121 11:15:51.089591 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f32d32e5-b286-475a-bda8-5d79b746e576-config\") on node \"crc\" DevicePath \"\"" Jan 21 11:15:51 crc kubenswrapper[4824]: I0121 11:15:51.089609 4824 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/f32d32e5-b286-475a-bda8-5d79b746e576-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Jan 21 11:15:51 crc kubenswrapper[4824]: I0121 11:15:51.090152 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f32d32e5-b286-475a-bda8-5d79b746e576-client-ca" (OuterVolumeSpecName: "client-ca") pod "f32d32e5-b286-475a-bda8-5d79b746e576" (UID: "f32d32e5-b286-475a-bda8-5d79b746e576"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:15:51 crc kubenswrapper[4824]: I0121 11:15:51.093485 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f32d32e5-b286-475a-bda8-5d79b746e576-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "f32d32e5-b286-475a-bda8-5d79b746e576" (UID: "f32d32e5-b286-475a-bda8-5d79b746e576"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:15:51 crc kubenswrapper[4824]: I0121 11:15:51.093646 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f32d32e5-b286-475a-bda8-5d79b746e576-kube-api-access-g6d2t" (OuterVolumeSpecName: "kube-api-access-g6d2t") pod "f32d32e5-b286-475a-bda8-5d79b746e576" (UID: "f32d32e5-b286-475a-bda8-5d79b746e576"). InnerVolumeSpecName "kube-api-access-g6d2t". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:15:51 crc kubenswrapper[4824]: I0121 11:15:51.190657 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/3d23b1ac-adfc-4f40-8c75-710a71784b55-proxy-ca-bundles\") pod \"controller-manager-76bbb855b9-ct4ct\" (UID: \"3d23b1ac-adfc-4f40-8c75-710a71784b55\") " pod="openshift-controller-manager/controller-manager-76bbb855b9-ct4ct" Jan 21 11:15:51 crc kubenswrapper[4824]: I0121 11:15:51.190829 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jq6rh\" (UniqueName: \"kubernetes.io/projected/3d23b1ac-adfc-4f40-8c75-710a71784b55-kube-api-access-jq6rh\") pod \"controller-manager-76bbb855b9-ct4ct\" (UID: \"3d23b1ac-adfc-4f40-8c75-710a71784b55\") " pod="openshift-controller-manager/controller-manager-76bbb855b9-ct4ct" Jan 21 11:15:51 crc kubenswrapper[4824]: I0121 11:15:51.190926 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c321fe0f-f966-4865-a619-6f5f26e9bb50-catalog-content\") pod \"certified-operators-rgl7w\" (UID: \"c321fe0f-f966-4865-a619-6f5f26e9bb50\") " pod="openshift-marketplace/certified-operators-rgl7w" Jan 21 11:15:51 crc kubenswrapper[4824]: I0121 11:15:51.191000 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c321fe0f-f966-4865-a619-6f5f26e9bb50-utilities\") pod \"certified-operators-rgl7w\" (UID: \"c321fe0f-f966-4865-a619-6f5f26e9bb50\") " pod="openshift-marketplace/certified-operators-rgl7w" Jan 21 11:15:51 crc kubenswrapper[4824]: I0121 11:15:51.191069 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3d23b1ac-adfc-4f40-8c75-710a71784b55-client-ca\") pod \"controller-manager-76bbb855b9-ct4ct\" (UID: \"3d23b1ac-adfc-4f40-8c75-710a71784b55\") " pod="openshift-controller-manager/controller-manager-76bbb855b9-ct4ct" Jan 21 11:15:51 crc kubenswrapper[4824]: I0121 11:15:51.191129 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3d23b1ac-adfc-4f40-8c75-710a71784b55-config\") pod \"controller-manager-76bbb855b9-ct4ct\" (UID: \"3d23b1ac-adfc-4f40-8c75-710a71784b55\") " pod="openshift-controller-manager/controller-manager-76bbb855b9-ct4ct" Jan 21 11:15:51 crc kubenswrapper[4824]: I0121 11:15:51.191165 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ms274\" (UniqueName: \"kubernetes.io/projected/c321fe0f-f966-4865-a619-6f5f26e9bb50-kube-api-access-ms274\") pod \"certified-operators-rgl7w\" (UID: \"c321fe0f-f966-4865-a619-6f5f26e9bb50\") " pod="openshift-marketplace/certified-operators-rgl7w" Jan 21 11:15:51 crc kubenswrapper[4824]: I0121 11:15:51.191201 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3d23b1ac-adfc-4f40-8c75-710a71784b55-serving-cert\") pod \"controller-manager-76bbb855b9-ct4ct\" (UID: \"3d23b1ac-adfc-4f40-8c75-710a71784b55\") " pod="openshift-controller-manager/controller-manager-76bbb855b9-ct4ct" Jan 21 11:15:51 crc kubenswrapper[4824]: I0121 11:15:51.191355 4824 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f32d32e5-b286-475a-bda8-5d79b746e576-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 21 11:15:51 crc kubenswrapper[4824]: I0121 11:15:51.191372 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g6d2t\" (UniqueName: \"kubernetes.io/projected/f32d32e5-b286-475a-bda8-5d79b746e576-kube-api-access-g6d2t\") on node \"crc\" DevicePath \"\"" Jan 21 11:15:51 crc kubenswrapper[4824]: I0121 11:15:51.191384 4824 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f32d32e5-b286-475a-bda8-5d79b746e576-client-ca\") on node \"crc\" DevicePath \"\"" Jan 21 11:15:51 crc kubenswrapper[4824]: I0121 11:15:51.191626 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/3d23b1ac-adfc-4f40-8c75-710a71784b55-proxy-ca-bundles\") pod \"controller-manager-76bbb855b9-ct4ct\" (UID: \"3d23b1ac-adfc-4f40-8c75-710a71784b55\") " pod="openshift-controller-manager/controller-manager-76bbb855b9-ct4ct" Jan 21 11:15:51 crc kubenswrapper[4824]: I0121 11:15:51.191711 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3d23b1ac-adfc-4f40-8c75-710a71784b55-client-ca\") pod \"controller-manager-76bbb855b9-ct4ct\" (UID: \"3d23b1ac-adfc-4f40-8c75-710a71784b55\") " pod="openshift-controller-manager/controller-manager-76bbb855b9-ct4ct" Jan 21 11:15:51 crc kubenswrapper[4824]: I0121 11:15:51.192136 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3d23b1ac-adfc-4f40-8c75-710a71784b55-config\") pod \"controller-manager-76bbb855b9-ct4ct\" (UID: \"3d23b1ac-adfc-4f40-8c75-710a71784b55\") " pod="openshift-controller-manager/controller-manager-76bbb855b9-ct4ct" Jan 21 11:15:51 crc kubenswrapper[4824]: I0121 11:15:51.194162 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3d23b1ac-adfc-4f40-8c75-710a71784b55-serving-cert\") pod \"controller-manager-76bbb855b9-ct4ct\" (UID: \"3d23b1ac-adfc-4f40-8c75-710a71784b55\") " pod="openshift-controller-manager/controller-manager-76bbb855b9-ct4ct" Jan 21 11:15:51 crc kubenswrapper[4824]: I0121 11:15:51.204019 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jq6rh\" (UniqueName: \"kubernetes.io/projected/3d23b1ac-adfc-4f40-8c75-710a71784b55-kube-api-access-jq6rh\") pod \"controller-manager-76bbb855b9-ct4ct\" (UID: \"3d23b1ac-adfc-4f40-8c75-710a71784b55\") " pod="openshift-controller-manager/controller-manager-76bbb855b9-ct4ct" Jan 21 11:15:51 crc kubenswrapper[4824]: I0121 11:15:51.258076 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-k2t4f"] Jan 21 11:15:51 crc kubenswrapper[4824]: I0121 11:15:51.258866 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-k2t4f" Jan 21 11:15:51 crc kubenswrapper[4824]: I0121 11:15:51.260532 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Jan 21 11:15:51 crc kubenswrapper[4824]: I0121 11:15:51.265600 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-k2t4f"] Jan 21 11:15:51 crc kubenswrapper[4824]: I0121 11:15:51.281707 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-76bbb855b9-ct4ct" Jan 21 11:15:51 crc kubenswrapper[4824]: I0121 11:15:51.291883 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ms274\" (UniqueName: \"kubernetes.io/projected/c321fe0f-f966-4865-a619-6f5f26e9bb50-kube-api-access-ms274\") pod \"certified-operators-rgl7w\" (UID: \"c321fe0f-f966-4865-a619-6f5f26e9bb50\") " pod="openshift-marketplace/certified-operators-rgl7w" Jan 21 11:15:51 crc kubenswrapper[4824]: I0121 11:15:51.291987 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c321fe0f-f966-4865-a619-6f5f26e9bb50-catalog-content\") pod \"certified-operators-rgl7w\" (UID: \"c321fe0f-f966-4865-a619-6f5f26e9bb50\") " pod="openshift-marketplace/certified-operators-rgl7w" Jan 21 11:15:51 crc kubenswrapper[4824]: I0121 11:15:51.292018 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c321fe0f-f966-4865-a619-6f5f26e9bb50-utilities\") pod \"certified-operators-rgl7w\" (UID: \"c321fe0f-f966-4865-a619-6f5f26e9bb50\") " pod="openshift-marketplace/certified-operators-rgl7w" Jan 21 11:15:51 crc kubenswrapper[4824]: I0121 11:15:51.292382 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c321fe0f-f966-4865-a619-6f5f26e9bb50-utilities\") pod \"certified-operators-rgl7w\" (UID: \"c321fe0f-f966-4865-a619-6f5f26e9bb50\") " pod="openshift-marketplace/certified-operators-rgl7w" Jan 21 11:15:51 crc kubenswrapper[4824]: I0121 11:15:51.292433 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c321fe0f-f966-4865-a619-6f5f26e9bb50-catalog-content\") pod \"certified-operators-rgl7w\" (UID: \"c321fe0f-f966-4865-a619-6f5f26e9bb50\") " pod="openshift-marketplace/certified-operators-rgl7w" Jan 21 11:15:51 crc kubenswrapper[4824]: I0121 11:15:51.304888 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ms274\" (UniqueName: \"kubernetes.io/projected/c321fe0f-f966-4865-a619-6f5f26e9bb50-kube-api-access-ms274\") pod \"certified-operators-rgl7w\" (UID: \"c321fe0f-f966-4865-a619-6f5f26e9bb50\") " pod="openshift-marketplace/certified-operators-rgl7w" Jan 21 11:15:51 crc kubenswrapper[4824]: I0121 11:15:51.371073 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rgl7w" Jan 21 11:15:51 crc kubenswrapper[4824]: I0121 11:15:51.393212 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-925vc\" (UniqueName: \"kubernetes.io/projected/46423d25-afc7-4a50-ba21-5f7c7b7012f2-kube-api-access-925vc\") pod \"redhat-marketplace-k2t4f\" (UID: \"46423d25-afc7-4a50-ba21-5f7c7b7012f2\") " pod="openshift-marketplace/redhat-marketplace-k2t4f" Jan 21 11:15:51 crc kubenswrapper[4824]: I0121 11:15:51.393308 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/46423d25-afc7-4a50-ba21-5f7c7b7012f2-catalog-content\") pod \"redhat-marketplace-k2t4f\" (UID: \"46423d25-afc7-4a50-ba21-5f7c7b7012f2\") " pod="openshift-marketplace/redhat-marketplace-k2t4f" Jan 21 11:15:51 crc kubenswrapper[4824]: I0121 11:15:51.393329 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/46423d25-afc7-4a50-ba21-5f7c7b7012f2-utilities\") pod \"redhat-marketplace-k2t4f\" (UID: \"46423d25-afc7-4a50-ba21-5f7c7b7012f2\") " pod="openshift-marketplace/redhat-marketplace-k2t4f" Jan 21 11:15:51 crc kubenswrapper[4824]: I0121 11:15:51.413151 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-76bbb855b9-ct4ct"] Jan 21 11:15:51 crc kubenswrapper[4824]: I0121 11:15:51.497362 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/46423d25-afc7-4a50-ba21-5f7c7b7012f2-catalog-content\") pod \"redhat-marketplace-k2t4f\" (UID: \"46423d25-afc7-4a50-ba21-5f7c7b7012f2\") " pod="openshift-marketplace/redhat-marketplace-k2t4f" Jan 21 11:15:51 crc kubenswrapper[4824]: I0121 11:15:51.497393 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/46423d25-afc7-4a50-ba21-5f7c7b7012f2-utilities\") pod \"redhat-marketplace-k2t4f\" (UID: \"46423d25-afc7-4a50-ba21-5f7c7b7012f2\") " pod="openshift-marketplace/redhat-marketplace-k2t4f" Jan 21 11:15:51 crc kubenswrapper[4824]: I0121 11:15:51.497424 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-925vc\" (UniqueName: \"kubernetes.io/projected/46423d25-afc7-4a50-ba21-5f7c7b7012f2-kube-api-access-925vc\") pod \"redhat-marketplace-k2t4f\" (UID: \"46423d25-afc7-4a50-ba21-5f7c7b7012f2\") " pod="openshift-marketplace/redhat-marketplace-k2t4f" Jan 21 11:15:51 crc kubenswrapper[4824]: I0121 11:15:51.498043 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/46423d25-afc7-4a50-ba21-5f7c7b7012f2-catalog-content\") pod \"redhat-marketplace-k2t4f\" (UID: \"46423d25-afc7-4a50-ba21-5f7c7b7012f2\") " pod="openshift-marketplace/redhat-marketplace-k2t4f" Jan 21 11:15:51 crc kubenswrapper[4824]: I0121 11:15:51.498066 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/46423d25-afc7-4a50-ba21-5f7c7b7012f2-utilities\") pod \"redhat-marketplace-k2t4f\" (UID: \"46423d25-afc7-4a50-ba21-5f7c7b7012f2\") " pod="openshift-marketplace/redhat-marketplace-k2t4f" Jan 21 11:15:51 crc kubenswrapper[4824]: I0121 11:15:51.510340 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-925vc\" (UniqueName: \"kubernetes.io/projected/46423d25-afc7-4a50-ba21-5f7c7b7012f2-kube-api-access-925vc\") pod \"redhat-marketplace-k2t4f\" (UID: \"46423d25-afc7-4a50-ba21-5f7c7b7012f2\") " pod="openshift-marketplace/redhat-marketplace-k2t4f" Jan 21 11:15:51 crc kubenswrapper[4824]: I0121 11:15:51.571708 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-k2t4f" Jan 21 11:15:51 crc kubenswrapper[4824]: I0121 11:15:51.678991 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-76bbb855b9-ct4ct" event={"ID":"3d23b1ac-adfc-4f40-8c75-710a71784b55","Type":"ContainerStarted","Data":"6e73e604b5615aef146f42a69da534899aab9da194d71bcb826e78193bc7ac1d"} Jan 21 11:15:51 crc kubenswrapper[4824]: I0121 11:15:51.679045 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-76bbb855b9-ct4ct" event={"ID":"3d23b1ac-adfc-4f40-8c75-710a71784b55","Type":"ContainerStarted","Data":"3d835dd9a1485af0ae9262121062825ff2601613bd318ffc3b5c44fd8e891dae"} Jan 21 11:15:51 crc kubenswrapper[4824]: I0121 11:15:51.679158 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-76bbb855b9-ct4ct" Jan 21 11:15:51 crc kubenswrapper[4824]: I0121 11:15:51.680635 4824 generic.go:334] "Generic (PLEG): container finished" podID="f32d32e5-b286-475a-bda8-5d79b746e576" containerID="7ca072731bf39a1cadc5fb56bba2c11e121f2b68f324f36479a7c2338c662422" exitCode=0 Jan 21 11:15:51 crc kubenswrapper[4824]: I0121 11:15:51.680680 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7c9b544d8-b9cfc" Jan 21 11:15:51 crc kubenswrapper[4824]: I0121 11:15:51.680705 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7c9b544d8-b9cfc" event={"ID":"f32d32e5-b286-475a-bda8-5d79b746e576","Type":"ContainerDied","Data":"7ca072731bf39a1cadc5fb56bba2c11e121f2b68f324f36479a7c2338c662422"} Jan 21 11:15:51 crc kubenswrapper[4824]: I0121 11:15:51.680729 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7c9b544d8-b9cfc" event={"ID":"f32d32e5-b286-475a-bda8-5d79b746e576","Type":"ContainerDied","Data":"41f5c0a541981a8fe31ce891bdb4387c75d383bf8c055b1b0de1ae7af35b3bcd"} Jan 21 11:15:51 crc kubenswrapper[4824]: I0121 11:15:51.680757 4824 scope.go:117] "RemoveContainer" containerID="7ca072731bf39a1cadc5fb56bba2c11e121f2b68f324f36479a7c2338c662422" Jan 21 11:15:51 crc kubenswrapper[4824]: I0121 11:15:51.695655 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-76bbb855b9-ct4ct" Jan 21 11:15:51 crc kubenswrapper[4824]: I0121 11:15:51.697122 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-m4srg" event={"ID":"53e1053f-f40b-4afc-ad9c-b07d46fc1f94","Type":"ContainerStarted","Data":"a68964f2e1a5aa3db059a08321155ec91cdb51762a74e4dfe8cdb3b1ebb163a3"} Jan 21 11:15:51 crc kubenswrapper[4824]: I0121 11:15:51.700711 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-76bbb855b9-ct4ct" podStartSLOduration=2.700699273 podStartE2EDuration="2.700699273s" podCreationTimestamp="2026-01-21 11:15:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:15:51.700125475 +0000 UTC m=+293.993154767" watchObservedRunningTime="2026-01-21 11:15:51.700699273 +0000 UTC m=+293.993728566" Jan 21 11:15:51 crc kubenswrapper[4824]: I0121 11:15:51.711125 4824 generic.go:334] "Generic (PLEG): container finished" podID="3447df50-807b-487c-9acf-2a2250780bdc" containerID="affad804ab6c5ef16e0c73a329f54a91b3bdbf2150e0045b7e2a45788bbf14f0" exitCode=0 Jan 21 11:15:51 crc kubenswrapper[4824]: I0121 11:15:51.711195 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5fdb5cfd56-pqprn" Jan 21 11:15:51 crc kubenswrapper[4824]: I0121 11:15:51.712216 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cmbbh" event={"ID":"3447df50-807b-487c-9acf-2a2250780bdc","Type":"ContainerDied","Data":"affad804ab6c5ef16e0c73a329f54a91b3bdbf2150e0045b7e2a45788bbf14f0"} Jan 21 11:15:51 crc kubenswrapper[4824]: I0121 11:15:51.723414 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-m4srg" podStartSLOduration=2.231438049 podStartE2EDuration="3.723398941s" podCreationTimestamp="2026-01-21 11:15:48 +0000 UTC" firstStartedPulling="2026-01-21 11:15:49.635056416 +0000 UTC m=+291.928085708" lastFinishedPulling="2026-01-21 11:15:51.127017308 +0000 UTC m=+293.420046600" observedRunningTime="2026-01-21 11:15:51.720483986 +0000 UTC m=+294.013513288" watchObservedRunningTime="2026-01-21 11:15:51.723398941 +0000 UTC m=+294.016428232" Jan 21 11:15:51 crc kubenswrapper[4824]: I0121 11:15:51.738174 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-7c9b544d8-b9cfc"] Jan 21 11:15:51 crc kubenswrapper[4824]: I0121 11:15:51.746543 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-7c9b544d8-b9cfc"] Jan 21 11:15:51 crc kubenswrapper[4824]: I0121 11:15:51.746874 4824 scope.go:117] "RemoveContainer" containerID="7ca072731bf39a1cadc5fb56bba2c11e121f2b68f324f36479a7c2338c662422" Jan 21 11:15:51 crc kubenswrapper[4824]: E0121 11:15:51.749703 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7ca072731bf39a1cadc5fb56bba2c11e121f2b68f324f36479a7c2338c662422\": container with ID starting with 7ca072731bf39a1cadc5fb56bba2c11e121f2b68f324f36479a7c2338c662422 not found: ID does not exist" containerID="7ca072731bf39a1cadc5fb56bba2c11e121f2b68f324f36479a7c2338c662422" Jan 21 11:15:51 crc kubenswrapper[4824]: I0121 11:15:51.749733 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7ca072731bf39a1cadc5fb56bba2c11e121f2b68f324f36479a7c2338c662422"} err="failed to get container status \"7ca072731bf39a1cadc5fb56bba2c11e121f2b68f324f36479a7c2338c662422\": rpc error: code = NotFound desc = could not find container \"7ca072731bf39a1cadc5fb56bba2c11e121f2b68f324f36479a7c2338c662422\": container with ID starting with 7ca072731bf39a1cadc5fb56bba2c11e121f2b68f324f36479a7c2338c662422 not found: ID does not exist" Jan 21 11:15:51 crc kubenswrapper[4824]: I0121 11:15:51.765436 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-rgl7w"] Jan 21 11:15:51 crc kubenswrapper[4824]: I0121 11:15:51.787541 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5fdb5cfd56-pqprn"] Jan 21 11:15:51 crc kubenswrapper[4824]: I0121 11:15:51.790357 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5fdb5cfd56-pqprn"] Jan 21 11:15:51 crc kubenswrapper[4824]: I0121 11:15:51.818560 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-k2t4f"] Jan 21 11:15:51 crc kubenswrapper[4824]: W0121 11:15:51.825571 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod46423d25_afc7_4a50_ba21_5f7c7b7012f2.slice/crio-552118ccb0cef413941998045e3b97b625892a5c504a290797b4bb8b3028c183 WatchSource:0}: Error finding container 552118ccb0cef413941998045e3b97b625892a5c504a290797b4bb8b3028c183: Status 404 returned error can't find the container with id 552118ccb0cef413941998045e3b97b625892a5c504a290797b4bb8b3028c183 Jan 21 11:15:52 crc kubenswrapper[4824]: I0121 11:15:52.057587 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="66af07e9-d9e9-4584-b458-afd995ac1eb3" path="/var/lib/kubelet/pods/66af07e9-d9e9-4584-b458-afd995ac1eb3/volumes" Jan 21 11:15:52 crc kubenswrapper[4824]: I0121 11:15:52.058640 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f32d32e5-b286-475a-bda8-5d79b746e576" path="/var/lib/kubelet/pods/f32d32e5-b286-475a-bda8-5d79b746e576/volumes" Jan 21 11:15:52 crc kubenswrapper[4824]: I0121 11:15:52.718187 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cmbbh" event={"ID":"3447df50-807b-487c-9acf-2a2250780bdc","Type":"ContainerStarted","Data":"89eb084b536fbfaab5d4e0ef427f60e1acf890ff15c52bf33a6e6997548d197c"} Jan 21 11:15:52 crc kubenswrapper[4824]: I0121 11:15:52.719934 4824 generic.go:334] "Generic (PLEG): container finished" podID="c321fe0f-f966-4865-a619-6f5f26e9bb50" containerID="0fc010d3d5d3a46de0f9ce4eaa81396f02f3a8e9ecaa5cd0de536f8051a3c583" exitCode=0 Jan 21 11:15:52 crc kubenswrapper[4824]: I0121 11:15:52.719999 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rgl7w" event={"ID":"c321fe0f-f966-4865-a619-6f5f26e9bb50","Type":"ContainerDied","Data":"0fc010d3d5d3a46de0f9ce4eaa81396f02f3a8e9ecaa5cd0de536f8051a3c583"} Jan 21 11:15:52 crc kubenswrapper[4824]: I0121 11:15:52.720049 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rgl7w" event={"ID":"c321fe0f-f966-4865-a619-6f5f26e9bb50","Type":"ContainerStarted","Data":"6f5d4e99055d52c4bf5250c538705a65a0a9ac83c8e6a13826190f5232d586a2"} Jan 21 11:15:52 crc kubenswrapper[4824]: I0121 11:15:52.723138 4824 generic.go:334] "Generic (PLEG): container finished" podID="46423d25-afc7-4a50-ba21-5f7c7b7012f2" containerID="726235af80b56c27b7242f3bf79d007ef44d2287ccaa0e576724a345d8d479bf" exitCode=0 Jan 21 11:15:52 crc kubenswrapper[4824]: I0121 11:15:52.723243 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-k2t4f" event={"ID":"46423d25-afc7-4a50-ba21-5f7c7b7012f2","Type":"ContainerDied","Data":"726235af80b56c27b7242f3bf79d007ef44d2287ccaa0e576724a345d8d479bf"} Jan 21 11:15:52 crc kubenswrapper[4824]: I0121 11:15:52.723334 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-k2t4f" event={"ID":"46423d25-afc7-4a50-ba21-5f7c7b7012f2","Type":"ContainerStarted","Data":"552118ccb0cef413941998045e3b97b625892a5c504a290797b4bb8b3028c183"} Jan 21 11:15:52 crc kubenswrapper[4824]: I0121 11:15:52.733325 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-cmbbh" podStartSLOduration=2.170493884 podStartE2EDuration="4.733308733s" podCreationTimestamp="2026-01-21 11:15:48 +0000 UTC" firstStartedPulling="2026-01-21 11:15:49.642003415 +0000 UTC m=+291.935032707" lastFinishedPulling="2026-01-21 11:15:52.204818264 +0000 UTC m=+294.497847556" observedRunningTime="2026-01-21 11:15:52.730278732 +0000 UTC m=+295.023308023" watchObservedRunningTime="2026-01-21 11:15:52.733308733 +0000 UTC m=+295.026338025" Jan 21 11:15:53 crc kubenswrapper[4824]: I0121 11:15:53.728126 4824 generic.go:334] "Generic (PLEG): container finished" podID="c321fe0f-f966-4865-a619-6f5f26e9bb50" containerID="df8c5cff86de5a10e577959873f18fe345273d2ec855f9afa2fc62cd1e0ec39c" exitCode=0 Jan 21 11:15:53 crc kubenswrapper[4824]: I0121 11:15:53.728196 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rgl7w" event={"ID":"c321fe0f-f966-4865-a619-6f5f26e9bb50","Type":"ContainerDied","Data":"df8c5cff86de5a10e577959873f18fe345273d2ec855f9afa2fc62cd1e0ec39c"} Jan 21 11:15:53 crc kubenswrapper[4824]: I0121 11:15:53.729568 4824 generic.go:334] "Generic (PLEG): container finished" podID="46423d25-afc7-4a50-ba21-5f7c7b7012f2" containerID="f168a7e84d19eb96f1cb78ac7a1a837363cfeb22a2806201cc947b1c314eda7f" exitCode=0 Jan 21 11:15:53 crc kubenswrapper[4824]: I0121 11:15:53.729636 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-k2t4f" event={"ID":"46423d25-afc7-4a50-ba21-5f7c7b7012f2","Type":"ContainerDied","Data":"f168a7e84d19eb96f1cb78ac7a1a837363cfeb22a2806201cc947b1c314eda7f"} Jan 21 11:15:53 crc kubenswrapper[4824]: I0121 11:15:53.920295 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-75ddb6d6df-pz2wn"] Jan 21 11:15:53 crc kubenswrapper[4824]: I0121 11:15:53.920852 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-75ddb6d6df-pz2wn" Jan 21 11:15:53 crc kubenswrapper[4824]: I0121 11:15:53.923170 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Jan 21 11:15:53 crc kubenswrapper[4824]: I0121 11:15:53.923348 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Jan 21 11:15:53 crc kubenswrapper[4824]: I0121 11:15:53.923402 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Jan 21 11:15:53 crc kubenswrapper[4824]: I0121 11:15:53.923438 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Jan 21 11:15:53 crc kubenswrapper[4824]: I0121 11:15:53.923592 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Jan 21 11:15:53 crc kubenswrapper[4824]: I0121 11:15:53.923648 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Jan 21 11:15:53 crc kubenswrapper[4824]: I0121 11:15:53.930089 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-75ddb6d6df-pz2wn"] Jan 21 11:15:54 crc kubenswrapper[4824]: I0121 11:15:54.029433 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/65ad4810-1d65-4965-b8c6-273069e2020f-config\") pod \"route-controller-manager-75ddb6d6df-pz2wn\" (UID: \"65ad4810-1d65-4965-b8c6-273069e2020f\") " pod="openshift-route-controller-manager/route-controller-manager-75ddb6d6df-pz2wn" Jan 21 11:15:54 crc kubenswrapper[4824]: I0121 11:15:54.029649 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7dxsx\" (UniqueName: \"kubernetes.io/projected/65ad4810-1d65-4965-b8c6-273069e2020f-kube-api-access-7dxsx\") pod \"route-controller-manager-75ddb6d6df-pz2wn\" (UID: \"65ad4810-1d65-4965-b8c6-273069e2020f\") " pod="openshift-route-controller-manager/route-controller-manager-75ddb6d6df-pz2wn" Jan 21 11:15:54 crc kubenswrapper[4824]: I0121 11:15:54.029712 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/65ad4810-1d65-4965-b8c6-273069e2020f-client-ca\") pod \"route-controller-manager-75ddb6d6df-pz2wn\" (UID: \"65ad4810-1d65-4965-b8c6-273069e2020f\") " pod="openshift-route-controller-manager/route-controller-manager-75ddb6d6df-pz2wn" Jan 21 11:15:54 crc kubenswrapper[4824]: I0121 11:15:54.029771 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/65ad4810-1d65-4965-b8c6-273069e2020f-serving-cert\") pod \"route-controller-manager-75ddb6d6df-pz2wn\" (UID: \"65ad4810-1d65-4965-b8c6-273069e2020f\") " pod="openshift-route-controller-manager/route-controller-manager-75ddb6d6df-pz2wn" Jan 21 11:15:54 crc kubenswrapper[4824]: I0121 11:15:54.130713 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7dxsx\" (UniqueName: \"kubernetes.io/projected/65ad4810-1d65-4965-b8c6-273069e2020f-kube-api-access-7dxsx\") pod \"route-controller-manager-75ddb6d6df-pz2wn\" (UID: \"65ad4810-1d65-4965-b8c6-273069e2020f\") " pod="openshift-route-controller-manager/route-controller-manager-75ddb6d6df-pz2wn" Jan 21 11:15:54 crc kubenswrapper[4824]: I0121 11:15:54.130766 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/65ad4810-1d65-4965-b8c6-273069e2020f-client-ca\") pod \"route-controller-manager-75ddb6d6df-pz2wn\" (UID: \"65ad4810-1d65-4965-b8c6-273069e2020f\") " pod="openshift-route-controller-manager/route-controller-manager-75ddb6d6df-pz2wn" Jan 21 11:15:54 crc kubenswrapper[4824]: I0121 11:15:54.130807 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/65ad4810-1d65-4965-b8c6-273069e2020f-serving-cert\") pod \"route-controller-manager-75ddb6d6df-pz2wn\" (UID: \"65ad4810-1d65-4965-b8c6-273069e2020f\") " pod="openshift-route-controller-manager/route-controller-manager-75ddb6d6df-pz2wn" Jan 21 11:15:54 crc kubenswrapper[4824]: I0121 11:15:54.130838 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/65ad4810-1d65-4965-b8c6-273069e2020f-config\") pod \"route-controller-manager-75ddb6d6df-pz2wn\" (UID: \"65ad4810-1d65-4965-b8c6-273069e2020f\") " pod="openshift-route-controller-manager/route-controller-manager-75ddb6d6df-pz2wn" Jan 21 11:15:54 crc kubenswrapper[4824]: I0121 11:15:54.131580 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/65ad4810-1d65-4965-b8c6-273069e2020f-client-ca\") pod \"route-controller-manager-75ddb6d6df-pz2wn\" (UID: \"65ad4810-1d65-4965-b8c6-273069e2020f\") " pod="openshift-route-controller-manager/route-controller-manager-75ddb6d6df-pz2wn" Jan 21 11:15:54 crc kubenswrapper[4824]: I0121 11:15:54.131761 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/65ad4810-1d65-4965-b8c6-273069e2020f-config\") pod \"route-controller-manager-75ddb6d6df-pz2wn\" (UID: \"65ad4810-1d65-4965-b8c6-273069e2020f\") " pod="openshift-route-controller-manager/route-controller-manager-75ddb6d6df-pz2wn" Jan 21 11:15:54 crc kubenswrapper[4824]: I0121 11:15:54.135673 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/65ad4810-1d65-4965-b8c6-273069e2020f-serving-cert\") pod \"route-controller-manager-75ddb6d6df-pz2wn\" (UID: \"65ad4810-1d65-4965-b8c6-273069e2020f\") " pod="openshift-route-controller-manager/route-controller-manager-75ddb6d6df-pz2wn" Jan 21 11:15:54 crc kubenswrapper[4824]: I0121 11:15:54.143900 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7dxsx\" (UniqueName: \"kubernetes.io/projected/65ad4810-1d65-4965-b8c6-273069e2020f-kube-api-access-7dxsx\") pod \"route-controller-manager-75ddb6d6df-pz2wn\" (UID: \"65ad4810-1d65-4965-b8c6-273069e2020f\") " pod="openshift-route-controller-manager/route-controller-manager-75ddb6d6df-pz2wn" Jan 21 11:15:54 crc kubenswrapper[4824]: I0121 11:15:54.231555 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-75ddb6d6df-pz2wn" Jan 21 11:15:54 crc kubenswrapper[4824]: I0121 11:15:54.591157 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-75ddb6d6df-pz2wn"] Jan 21 11:15:54 crc kubenswrapper[4824]: I0121 11:15:54.736179 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rgl7w" event={"ID":"c321fe0f-f966-4865-a619-6f5f26e9bb50","Type":"ContainerStarted","Data":"2475e62f3fd1d88e660f6d84396426f9e55aa1f97c8d45572e9660583aaa75a9"} Jan 21 11:15:54 crc kubenswrapper[4824]: I0121 11:15:54.737815 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-75ddb6d6df-pz2wn" event={"ID":"65ad4810-1d65-4965-b8c6-273069e2020f","Type":"ContainerStarted","Data":"a7982b5bdad04a3dc642d1b963cd9c799010cdd1364fdb4841f49984b61f2f3a"} Jan 21 11:15:54 crc kubenswrapper[4824]: I0121 11:15:54.737860 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-75ddb6d6df-pz2wn" Jan 21 11:15:54 crc kubenswrapper[4824]: I0121 11:15:54.738340 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-75ddb6d6df-pz2wn" event={"ID":"65ad4810-1d65-4965-b8c6-273069e2020f","Type":"ContainerStarted","Data":"50a6e7e59a40fc7ef484103f69dec6c22086ee8fe5ecb855e025964076f2c1cf"} Jan 21 11:15:54 crc kubenswrapper[4824]: I0121 11:15:54.739809 4824 patch_prober.go:28] interesting pod/route-controller-manager-75ddb6d6df-pz2wn container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.67:8443/healthz\": dial tcp 10.217.0.67:8443: connect: connection refused" start-of-body= Jan 21 11:15:54 crc kubenswrapper[4824]: I0121 11:15:54.739846 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-75ddb6d6df-pz2wn" podUID="65ad4810-1d65-4965-b8c6-273069e2020f" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.67:8443/healthz\": dial tcp 10.217.0.67:8443: connect: connection refused" Jan 21 11:15:54 crc kubenswrapper[4824]: I0121 11:15:54.740107 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-k2t4f" event={"ID":"46423d25-afc7-4a50-ba21-5f7c7b7012f2","Type":"ContainerStarted","Data":"d2fc14ea25cdf27fd2552309d072f2f5fbad183d12dd8087283f643af1014aac"} Jan 21 11:15:54 crc kubenswrapper[4824]: I0121 11:15:54.766374 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-rgl7w" podStartSLOduration=2.280903155 podStartE2EDuration="3.76635991s" podCreationTimestamp="2026-01-21 11:15:51 +0000 UTC" firstStartedPulling="2026-01-21 11:15:52.721153431 +0000 UTC m=+295.014182723" lastFinishedPulling="2026-01-21 11:15:54.206610186 +0000 UTC m=+296.499639478" observedRunningTime="2026-01-21 11:15:54.752151092 +0000 UTC m=+297.045180384" watchObservedRunningTime="2026-01-21 11:15:54.76635991 +0000 UTC m=+297.059389202" Jan 21 11:15:54 crc kubenswrapper[4824]: I0121 11:15:54.767588 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-75ddb6d6df-pz2wn" podStartSLOduration=5.767581987 podStartE2EDuration="5.767581987s" podCreationTimestamp="2026-01-21 11:15:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:15:54.765658022 +0000 UTC m=+297.058687314" watchObservedRunningTime="2026-01-21 11:15:54.767581987 +0000 UTC m=+297.060611279" Jan 21 11:15:54 crc kubenswrapper[4824]: I0121 11:15:54.779488 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-k2t4f" podStartSLOduration=2.234450271 podStartE2EDuration="3.77946183s" podCreationTimestamp="2026-01-21 11:15:51 +0000 UTC" firstStartedPulling="2026-01-21 11:15:52.72438443 +0000 UTC m=+295.017413722" lastFinishedPulling="2026-01-21 11:15:54.269395989 +0000 UTC m=+296.562425281" observedRunningTime="2026-01-21 11:15:54.777376152 +0000 UTC m=+297.070405445" watchObservedRunningTime="2026-01-21 11:15:54.77946183 +0000 UTC m=+297.072491122" Jan 21 11:15:55 crc kubenswrapper[4824]: I0121 11:15:55.747416 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-75ddb6d6df-pz2wn" Jan 21 11:15:57 crc kubenswrapper[4824]: I0121 11:15:57.910646 4824 cert_rotation.go:91] certificate rotation detected, shutting down client connections to start using new credentials Jan 21 11:15:58 crc kubenswrapper[4824]: I0121 11:15:58.775136 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-cmbbh" Jan 21 11:15:58 crc kubenswrapper[4824]: I0121 11:15:58.775174 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-cmbbh" Jan 21 11:15:58 crc kubenswrapper[4824]: I0121 11:15:58.801903 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-cmbbh" Jan 21 11:15:58 crc kubenswrapper[4824]: I0121 11:15:58.975734 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-m4srg" Jan 21 11:15:58 crc kubenswrapper[4824]: I0121 11:15:58.975770 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-m4srg" Jan 21 11:15:59 crc kubenswrapper[4824]: I0121 11:15:59.000245 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-m4srg" Jan 21 11:15:59 crc kubenswrapper[4824]: I0121 11:15:59.783678 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-m4srg" Jan 21 11:15:59 crc kubenswrapper[4824]: I0121 11:15:59.788155 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-cmbbh" Jan 21 11:16:01 crc kubenswrapper[4824]: I0121 11:16:01.371948 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-rgl7w" Jan 21 11:16:01 crc kubenswrapper[4824]: I0121 11:16:01.372222 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-rgl7w" Jan 21 11:16:01 crc kubenswrapper[4824]: I0121 11:16:01.398955 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-rgl7w" Jan 21 11:16:01 crc kubenswrapper[4824]: I0121 11:16:01.572772 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-k2t4f" Jan 21 11:16:01 crc kubenswrapper[4824]: I0121 11:16:01.572816 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-k2t4f" Jan 21 11:16:01 crc kubenswrapper[4824]: I0121 11:16:01.599356 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-k2t4f" Jan 21 11:16:01 crc kubenswrapper[4824]: I0121 11:16:01.795551 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-k2t4f" Jan 21 11:16:01 crc kubenswrapper[4824]: I0121 11:16:01.798079 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-rgl7w" Jan 21 11:16:07 crc kubenswrapper[4824]: I0121 11:16:07.827345 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-76bbb855b9-ct4ct"] Jan 21 11:16:07 crc kubenswrapper[4824]: I0121 11:16:07.827687 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-76bbb855b9-ct4ct" podUID="3d23b1ac-adfc-4f40-8c75-710a71784b55" containerName="controller-manager" containerID="cri-o://6e73e604b5615aef146f42a69da534899aab9da194d71bcb826e78193bc7ac1d" gracePeriod=30 Jan 21 11:16:07 crc kubenswrapper[4824]: I0121 11:16:07.838590 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-75ddb6d6df-pz2wn"] Jan 21 11:16:07 crc kubenswrapper[4824]: I0121 11:16:07.838761 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-75ddb6d6df-pz2wn" podUID="65ad4810-1d65-4965-b8c6-273069e2020f" containerName="route-controller-manager" containerID="cri-o://a7982b5bdad04a3dc642d1b963cd9c799010cdd1364fdb4841f49984b61f2f3a" gracePeriod=30 Jan 21 11:16:08 crc kubenswrapper[4824]: I0121 11:16:08.243614 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-75ddb6d6df-pz2wn" Jan 21 11:16:08 crc kubenswrapper[4824]: I0121 11:16:08.303380 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-76bbb855b9-ct4ct" Jan 21 11:16:08 crc kubenswrapper[4824]: I0121 11:16:08.374106 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/65ad4810-1d65-4965-b8c6-273069e2020f-serving-cert\") pod \"65ad4810-1d65-4965-b8c6-273069e2020f\" (UID: \"65ad4810-1d65-4965-b8c6-273069e2020f\") " Jan 21 11:16:08 crc kubenswrapper[4824]: I0121 11:16:08.374190 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/65ad4810-1d65-4965-b8c6-273069e2020f-config\") pod \"65ad4810-1d65-4965-b8c6-273069e2020f\" (UID: \"65ad4810-1d65-4965-b8c6-273069e2020f\") " Jan 21 11:16:08 crc kubenswrapper[4824]: I0121 11:16:08.374239 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/65ad4810-1d65-4965-b8c6-273069e2020f-client-ca\") pod \"65ad4810-1d65-4965-b8c6-273069e2020f\" (UID: \"65ad4810-1d65-4965-b8c6-273069e2020f\") " Jan 21 11:16:08 crc kubenswrapper[4824]: I0121 11:16:08.374317 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7dxsx\" (UniqueName: \"kubernetes.io/projected/65ad4810-1d65-4965-b8c6-273069e2020f-kube-api-access-7dxsx\") pod \"65ad4810-1d65-4965-b8c6-273069e2020f\" (UID: \"65ad4810-1d65-4965-b8c6-273069e2020f\") " Jan 21 11:16:08 crc kubenswrapper[4824]: I0121 11:16:08.374832 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/65ad4810-1d65-4965-b8c6-273069e2020f-config" (OuterVolumeSpecName: "config") pod "65ad4810-1d65-4965-b8c6-273069e2020f" (UID: "65ad4810-1d65-4965-b8c6-273069e2020f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:16:08 crc kubenswrapper[4824]: I0121 11:16:08.374973 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/65ad4810-1d65-4965-b8c6-273069e2020f-client-ca" (OuterVolumeSpecName: "client-ca") pod "65ad4810-1d65-4965-b8c6-273069e2020f" (UID: "65ad4810-1d65-4965-b8c6-273069e2020f"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:16:08 crc kubenswrapper[4824]: I0121 11:16:08.378633 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/65ad4810-1d65-4965-b8c6-273069e2020f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "65ad4810-1d65-4965-b8c6-273069e2020f" (UID: "65ad4810-1d65-4965-b8c6-273069e2020f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:16:08 crc kubenswrapper[4824]: I0121 11:16:08.378766 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/65ad4810-1d65-4965-b8c6-273069e2020f-kube-api-access-7dxsx" (OuterVolumeSpecName: "kube-api-access-7dxsx") pod "65ad4810-1d65-4965-b8c6-273069e2020f" (UID: "65ad4810-1d65-4965-b8c6-273069e2020f"). InnerVolumeSpecName "kube-api-access-7dxsx". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:16:08 crc kubenswrapper[4824]: I0121 11:16:08.475173 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3d23b1ac-adfc-4f40-8c75-710a71784b55-config\") pod \"3d23b1ac-adfc-4f40-8c75-710a71784b55\" (UID: \"3d23b1ac-adfc-4f40-8c75-710a71784b55\") " Jan 21 11:16:08 crc kubenswrapper[4824]: I0121 11:16:08.475213 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jq6rh\" (UniqueName: \"kubernetes.io/projected/3d23b1ac-adfc-4f40-8c75-710a71784b55-kube-api-access-jq6rh\") pod \"3d23b1ac-adfc-4f40-8c75-710a71784b55\" (UID: \"3d23b1ac-adfc-4f40-8c75-710a71784b55\") " Jan 21 11:16:08 crc kubenswrapper[4824]: I0121 11:16:08.475249 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3d23b1ac-adfc-4f40-8c75-710a71784b55-client-ca\") pod \"3d23b1ac-adfc-4f40-8c75-710a71784b55\" (UID: \"3d23b1ac-adfc-4f40-8c75-710a71784b55\") " Jan 21 11:16:08 crc kubenswrapper[4824]: I0121 11:16:08.475264 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/3d23b1ac-adfc-4f40-8c75-710a71784b55-proxy-ca-bundles\") pod \"3d23b1ac-adfc-4f40-8c75-710a71784b55\" (UID: \"3d23b1ac-adfc-4f40-8c75-710a71784b55\") " Jan 21 11:16:08 crc kubenswrapper[4824]: I0121 11:16:08.475280 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3d23b1ac-adfc-4f40-8c75-710a71784b55-serving-cert\") pod \"3d23b1ac-adfc-4f40-8c75-710a71784b55\" (UID: \"3d23b1ac-adfc-4f40-8c75-710a71784b55\") " Jan 21 11:16:08 crc kubenswrapper[4824]: I0121 11:16:08.475476 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/65ad4810-1d65-4965-b8c6-273069e2020f-config\") on node \"crc\" DevicePath \"\"" Jan 21 11:16:08 crc kubenswrapper[4824]: I0121 11:16:08.475487 4824 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/65ad4810-1d65-4965-b8c6-273069e2020f-client-ca\") on node \"crc\" DevicePath \"\"" Jan 21 11:16:08 crc kubenswrapper[4824]: I0121 11:16:08.475497 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7dxsx\" (UniqueName: \"kubernetes.io/projected/65ad4810-1d65-4965-b8c6-273069e2020f-kube-api-access-7dxsx\") on node \"crc\" DevicePath \"\"" Jan 21 11:16:08 crc kubenswrapper[4824]: I0121 11:16:08.475506 4824 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/65ad4810-1d65-4965-b8c6-273069e2020f-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 21 11:16:08 crc kubenswrapper[4824]: I0121 11:16:08.475822 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3d23b1ac-adfc-4f40-8c75-710a71784b55-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "3d23b1ac-adfc-4f40-8c75-710a71784b55" (UID: "3d23b1ac-adfc-4f40-8c75-710a71784b55"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:16:08 crc kubenswrapper[4824]: I0121 11:16:08.475882 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3d23b1ac-adfc-4f40-8c75-710a71784b55-config" (OuterVolumeSpecName: "config") pod "3d23b1ac-adfc-4f40-8c75-710a71784b55" (UID: "3d23b1ac-adfc-4f40-8c75-710a71784b55"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:16:08 crc kubenswrapper[4824]: I0121 11:16:08.475895 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3d23b1ac-adfc-4f40-8c75-710a71784b55-client-ca" (OuterVolumeSpecName: "client-ca") pod "3d23b1ac-adfc-4f40-8c75-710a71784b55" (UID: "3d23b1ac-adfc-4f40-8c75-710a71784b55"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:16:08 crc kubenswrapper[4824]: I0121 11:16:08.477882 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3d23b1ac-adfc-4f40-8c75-710a71784b55-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "3d23b1ac-adfc-4f40-8c75-710a71784b55" (UID: "3d23b1ac-adfc-4f40-8c75-710a71784b55"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:16:08 crc kubenswrapper[4824]: I0121 11:16:08.477910 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3d23b1ac-adfc-4f40-8c75-710a71784b55-kube-api-access-jq6rh" (OuterVolumeSpecName: "kube-api-access-jq6rh") pod "3d23b1ac-adfc-4f40-8c75-710a71784b55" (UID: "3d23b1ac-adfc-4f40-8c75-710a71784b55"). InnerVolumeSpecName "kube-api-access-jq6rh". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:16:08 crc kubenswrapper[4824]: I0121 11:16:08.576299 4824 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3d23b1ac-adfc-4f40-8c75-710a71784b55-client-ca\") on node \"crc\" DevicePath \"\"" Jan 21 11:16:08 crc kubenswrapper[4824]: I0121 11:16:08.576327 4824 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/3d23b1ac-adfc-4f40-8c75-710a71784b55-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Jan 21 11:16:08 crc kubenswrapper[4824]: I0121 11:16:08.576336 4824 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3d23b1ac-adfc-4f40-8c75-710a71784b55-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 21 11:16:08 crc kubenswrapper[4824]: I0121 11:16:08.576345 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3d23b1ac-adfc-4f40-8c75-710a71784b55-config\") on node \"crc\" DevicePath \"\"" Jan 21 11:16:08 crc kubenswrapper[4824]: I0121 11:16:08.576353 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jq6rh\" (UniqueName: \"kubernetes.io/projected/3d23b1ac-adfc-4f40-8c75-710a71784b55-kube-api-access-jq6rh\") on node \"crc\" DevicePath \"\"" Jan 21 11:16:08 crc kubenswrapper[4824]: I0121 11:16:08.796462 4824 generic.go:334] "Generic (PLEG): container finished" podID="3d23b1ac-adfc-4f40-8c75-710a71784b55" containerID="6e73e604b5615aef146f42a69da534899aab9da194d71bcb826e78193bc7ac1d" exitCode=0 Jan 21 11:16:08 crc kubenswrapper[4824]: I0121 11:16:08.796500 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-76bbb855b9-ct4ct" Jan 21 11:16:08 crc kubenswrapper[4824]: I0121 11:16:08.796514 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-76bbb855b9-ct4ct" event={"ID":"3d23b1ac-adfc-4f40-8c75-710a71784b55","Type":"ContainerDied","Data":"6e73e604b5615aef146f42a69da534899aab9da194d71bcb826e78193bc7ac1d"} Jan 21 11:16:08 crc kubenswrapper[4824]: I0121 11:16:08.796758 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-76bbb855b9-ct4ct" event={"ID":"3d23b1ac-adfc-4f40-8c75-710a71784b55","Type":"ContainerDied","Data":"3d835dd9a1485af0ae9262121062825ff2601613bd318ffc3b5c44fd8e891dae"} Jan 21 11:16:08 crc kubenswrapper[4824]: I0121 11:16:08.796776 4824 scope.go:117] "RemoveContainer" containerID="6e73e604b5615aef146f42a69da534899aab9da194d71bcb826e78193bc7ac1d" Jan 21 11:16:08 crc kubenswrapper[4824]: I0121 11:16:08.798874 4824 generic.go:334] "Generic (PLEG): container finished" podID="65ad4810-1d65-4965-b8c6-273069e2020f" containerID="a7982b5bdad04a3dc642d1b963cd9c799010cdd1364fdb4841f49984b61f2f3a" exitCode=0 Jan 21 11:16:08 crc kubenswrapper[4824]: I0121 11:16:08.798902 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-75ddb6d6df-pz2wn" event={"ID":"65ad4810-1d65-4965-b8c6-273069e2020f","Type":"ContainerDied","Data":"a7982b5bdad04a3dc642d1b963cd9c799010cdd1364fdb4841f49984b61f2f3a"} Jan 21 11:16:08 crc kubenswrapper[4824]: I0121 11:16:08.798921 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-75ddb6d6df-pz2wn" event={"ID":"65ad4810-1d65-4965-b8c6-273069e2020f","Type":"ContainerDied","Data":"50a6e7e59a40fc7ef484103f69dec6c22086ee8fe5ecb855e025964076f2c1cf"} Jan 21 11:16:08 crc kubenswrapper[4824]: I0121 11:16:08.798980 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-75ddb6d6df-pz2wn" Jan 21 11:16:08 crc kubenswrapper[4824]: I0121 11:16:08.808435 4824 scope.go:117] "RemoveContainer" containerID="6e73e604b5615aef146f42a69da534899aab9da194d71bcb826e78193bc7ac1d" Jan 21 11:16:08 crc kubenswrapper[4824]: E0121 11:16:08.809473 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6e73e604b5615aef146f42a69da534899aab9da194d71bcb826e78193bc7ac1d\": container with ID starting with 6e73e604b5615aef146f42a69da534899aab9da194d71bcb826e78193bc7ac1d not found: ID does not exist" containerID="6e73e604b5615aef146f42a69da534899aab9da194d71bcb826e78193bc7ac1d" Jan 21 11:16:08 crc kubenswrapper[4824]: I0121 11:16:08.809502 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6e73e604b5615aef146f42a69da534899aab9da194d71bcb826e78193bc7ac1d"} err="failed to get container status \"6e73e604b5615aef146f42a69da534899aab9da194d71bcb826e78193bc7ac1d\": rpc error: code = NotFound desc = could not find container \"6e73e604b5615aef146f42a69da534899aab9da194d71bcb826e78193bc7ac1d\": container with ID starting with 6e73e604b5615aef146f42a69da534899aab9da194d71bcb826e78193bc7ac1d not found: ID does not exist" Jan 21 11:16:08 crc kubenswrapper[4824]: I0121 11:16:08.809519 4824 scope.go:117] "RemoveContainer" containerID="a7982b5bdad04a3dc642d1b963cd9c799010cdd1364fdb4841f49984b61f2f3a" Jan 21 11:16:08 crc kubenswrapper[4824]: I0121 11:16:08.817210 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-76bbb855b9-ct4ct"] Jan 21 11:16:08 crc kubenswrapper[4824]: I0121 11:16:08.822952 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-76bbb855b9-ct4ct"] Jan 21 11:16:08 crc kubenswrapper[4824]: I0121 11:16:08.827918 4824 scope.go:117] "RemoveContainer" containerID="a7982b5bdad04a3dc642d1b963cd9c799010cdd1364fdb4841f49984b61f2f3a" Jan 21 11:16:08 crc kubenswrapper[4824]: E0121 11:16:08.828489 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a7982b5bdad04a3dc642d1b963cd9c799010cdd1364fdb4841f49984b61f2f3a\": container with ID starting with a7982b5bdad04a3dc642d1b963cd9c799010cdd1364fdb4841f49984b61f2f3a not found: ID does not exist" containerID="a7982b5bdad04a3dc642d1b963cd9c799010cdd1364fdb4841f49984b61f2f3a" Jan 21 11:16:08 crc kubenswrapper[4824]: I0121 11:16:08.828526 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a7982b5bdad04a3dc642d1b963cd9c799010cdd1364fdb4841f49984b61f2f3a"} err="failed to get container status \"a7982b5bdad04a3dc642d1b963cd9c799010cdd1364fdb4841f49984b61f2f3a\": rpc error: code = NotFound desc = could not find container \"a7982b5bdad04a3dc642d1b963cd9c799010cdd1364fdb4841f49984b61f2f3a\": container with ID starting with a7982b5bdad04a3dc642d1b963cd9c799010cdd1364fdb4841f49984b61f2f3a not found: ID does not exist" Jan 21 11:16:08 crc kubenswrapper[4824]: I0121 11:16:08.829090 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-75ddb6d6df-pz2wn"] Jan 21 11:16:08 crc kubenswrapper[4824]: I0121 11:16:08.831778 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-75ddb6d6df-pz2wn"] Jan 21 11:16:08 crc kubenswrapper[4824]: I0121 11:16:08.930838 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5fdb5cfd56-mtp66"] Jan 21 11:16:08 crc kubenswrapper[4824]: E0121 11:16:08.931096 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="65ad4810-1d65-4965-b8c6-273069e2020f" containerName="route-controller-manager" Jan 21 11:16:08 crc kubenswrapper[4824]: I0121 11:16:08.931109 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="65ad4810-1d65-4965-b8c6-273069e2020f" containerName="route-controller-manager" Jan 21 11:16:08 crc kubenswrapper[4824]: E0121 11:16:08.931122 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d23b1ac-adfc-4f40-8c75-710a71784b55" containerName="controller-manager" Jan 21 11:16:08 crc kubenswrapper[4824]: I0121 11:16:08.931128 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d23b1ac-adfc-4f40-8c75-710a71784b55" containerName="controller-manager" Jan 21 11:16:08 crc kubenswrapper[4824]: I0121 11:16:08.931221 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="65ad4810-1d65-4965-b8c6-273069e2020f" containerName="route-controller-manager" Jan 21 11:16:08 crc kubenswrapper[4824]: I0121 11:16:08.931231 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="3d23b1ac-adfc-4f40-8c75-710a71784b55" containerName="controller-manager" Jan 21 11:16:08 crc kubenswrapper[4824]: I0121 11:16:08.931603 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5fdb5cfd56-mtp66" Jan 21 11:16:08 crc kubenswrapper[4824]: I0121 11:16:08.933264 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Jan 21 11:16:08 crc kubenswrapper[4824]: I0121 11:16:08.933577 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Jan 21 11:16:08 crc kubenswrapper[4824]: I0121 11:16:08.933687 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Jan 21 11:16:08 crc kubenswrapper[4824]: I0121 11:16:08.933886 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Jan 21 11:16:08 crc kubenswrapper[4824]: I0121 11:16:08.934014 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Jan 21 11:16:08 crc kubenswrapper[4824]: I0121 11:16:08.935141 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Jan 21 11:16:08 crc kubenswrapper[4824]: I0121 11:16:08.937445 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-5975f76c65-dnkhj"] Jan 21 11:16:08 crc kubenswrapper[4824]: I0121 11:16:08.939001 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5975f76c65-dnkhj" Jan 21 11:16:08 crc kubenswrapper[4824]: I0121 11:16:08.940223 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5fdb5cfd56-mtp66"] Jan 21 11:16:08 crc kubenswrapper[4824]: I0121 11:16:08.940720 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Jan 21 11:16:08 crc kubenswrapper[4824]: I0121 11:16:08.941621 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Jan 21 11:16:08 crc kubenswrapper[4824]: I0121 11:16:08.942064 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Jan 21 11:16:08 crc kubenswrapper[4824]: I0121 11:16:08.942321 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Jan 21 11:16:08 crc kubenswrapper[4824]: I0121 11:16:08.942534 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Jan 21 11:16:08 crc kubenswrapper[4824]: I0121 11:16:08.942942 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-5975f76c65-dnkhj"] Jan 21 11:16:08 crc kubenswrapper[4824]: I0121 11:16:08.945623 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Jan 21 11:16:08 crc kubenswrapper[4824]: I0121 11:16:08.945697 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Jan 21 11:16:08 crc kubenswrapper[4824]: I0121 11:16:08.981271 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g5w7m\" (UniqueName: \"kubernetes.io/projected/b3bae151-0d8c-4715-a57c-77a832079b43-kube-api-access-g5w7m\") pod \"route-controller-manager-5fdb5cfd56-mtp66\" (UID: \"b3bae151-0d8c-4715-a57c-77a832079b43\") " pod="openshift-route-controller-manager/route-controller-manager-5fdb5cfd56-mtp66" Jan 21 11:16:08 crc kubenswrapper[4824]: I0121 11:16:08.981332 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ee260316-013d-420c-8462-9919f6a10f12-client-ca\") pod \"controller-manager-5975f76c65-dnkhj\" (UID: \"ee260316-013d-420c-8462-9919f6a10f12\") " pod="openshift-controller-manager/controller-manager-5975f76c65-dnkhj" Jan 21 11:16:08 crc kubenswrapper[4824]: I0121 11:16:08.981360 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b3bae151-0d8c-4715-a57c-77a832079b43-client-ca\") pod \"route-controller-manager-5fdb5cfd56-mtp66\" (UID: \"b3bae151-0d8c-4715-a57c-77a832079b43\") " pod="openshift-route-controller-manager/route-controller-manager-5fdb5cfd56-mtp66" Jan 21 11:16:08 crc kubenswrapper[4824]: I0121 11:16:08.981401 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b3bae151-0d8c-4715-a57c-77a832079b43-config\") pod \"route-controller-manager-5fdb5cfd56-mtp66\" (UID: \"b3bae151-0d8c-4715-a57c-77a832079b43\") " pod="openshift-route-controller-manager/route-controller-manager-5fdb5cfd56-mtp66" Jan 21 11:16:08 crc kubenswrapper[4824]: I0121 11:16:08.981416 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/ee260316-013d-420c-8462-9919f6a10f12-proxy-ca-bundles\") pod \"controller-manager-5975f76c65-dnkhj\" (UID: \"ee260316-013d-420c-8462-9919f6a10f12\") " pod="openshift-controller-manager/controller-manager-5975f76c65-dnkhj" Jan 21 11:16:08 crc kubenswrapper[4824]: I0121 11:16:08.981432 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ee260316-013d-420c-8462-9919f6a10f12-config\") pod \"controller-manager-5975f76c65-dnkhj\" (UID: \"ee260316-013d-420c-8462-9919f6a10f12\") " pod="openshift-controller-manager/controller-manager-5975f76c65-dnkhj" Jan 21 11:16:08 crc kubenswrapper[4824]: I0121 11:16:08.981449 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ee260316-013d-420c-8462-9919f6a10f12-serving-cert\") pod \"controller-manager-5975f76c65-dnkhj\" (UID: \"ee260316-013d-420c-8462-9919f6a10f12\") " pod="openshift-controller-manager/controller-manager-5975f76c65-dnkhj" Jan 21 11:16:08 crc kubenswrapper[4824]: I0121 11:16:08.981488 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b3bae151-0d8c-4715-a57c-77a832079b43-serving-cert\") pod \"route-controller-manager-5fdb5cfd56-mtp66\" (UID: \"b3bae151-0d8c-4715-a57c-77a832079b43\") " pod="openshift-route-controller-manager/route-controller-manager-5fdb5cfd56-mtp66" Jan 21 11:16:08 crc kubenswrapper[4824]: I0121 11:16:08.981508 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fgcf4\" (UniqueName: \"kubernetes.io/projected/ee260316-013d-420c-8462-9919f6a10f12-kube-api-access-fgcf4\") pod \"controller-manager-5975f76c65-dnkhj\" (UID: \"ee260316-013d-420c-8462-9919f6a10f12\") " pod="openshift-controller-manager/controller-manager-5975f76c65-dnkhj" Jan 21 11:16:09 crc kubenswrapper[4824]: I0121 11:16:09.082247 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b3bae151-0d8c-4715-a57c-77a832079b43-client-ca\") pod \"route-controller-manager-5fdb5cfd56-mtp66\" (UID: \"b3bae151-0d8c-4715-a57c-77a832079b43\") " pod="openshift-route-controller-manager/route-controller-manager-5fdb5cfd56-mtp66" Jan 21 11:16:09 crc kubenswrapper[4824]: I0121 11:16:09.082295 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b3bae151-0d8c-4715-a57c-77a832079b43-config\") pod \"route-controller-manager-5fdb5cfd56-mtp66\" (UID: \"b3bae151-0d8c-4715-a57c-77a832079b43\") " pod="openshift-route-controller-manager/route-controller-manager-5fdb5cfd56-mtp66" Jan 21 11:16:09 crc kubenswrapper[4824]: I0121 11:16:09.082314 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/ee260316-013d-420c-8462-9919f6a10f12-proxy-ca-bundles\") pod \"controller-manager-5975f76c65-dnkhj\" (UID: \"ee260316-013d-420c-8462-9919f6a10f12\") " pod="openshift-controller-manager/controller-manager-5975f76c65-dnkhj" Jan 21 11:16:09 crc kubenswrapper[4824]: I0121 11:16:09.082346 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ee260316-013d-420c-8462-9919f6a10f12-config\") pod \"controller-manager-5975f76c65-dnkhj\" (UID: \"ee260316-013d-420c-8462-9919f6a10f12\") " pod="openshift-controller-manager/controller-manager-5975f76c65-dnkhj" Jan 21 11:16:09 crc kubenswrapper[4824]: I0121 11:16:09.082363 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ee260316-013d-420c-8462-9919f6a10f12-serving-cert\") pod \"controller-manager-5975f76c65-dnkhj\" (UID: \"ee260316-013d-420c-8462-9919f6a10f12\") " pod="openshift-controller-manager/controller-manager-5975f76c65-dnkhj" Jan 21 11:16:09 crc kubenswrapper[4824]: I0121 11:16:09.082388 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b3bae151-0d8c-4715-a57c-77a832079b43-serving-cert\") pod \"route-controller-manager-5fdb5cfd56-mtp66\" (UID: \"b3bae151-0d8c-4715-a57c-77a832079b43\") " pod="openshift-route-controller-manager/route-controller-manager-5fdb5cfd56-mtp66" Jan 21 11:16:09 crc kubenswrapper[4824]: I0121 11:16:09.082414 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fgcf4\" (UniqueName: \"kubernetes.io/projected/ee260316-013d-420c-8462-9919f6a10f12-kube-api-access-fgcf4\") pod \"controller-manager-5975f76c65-dnkhj\" (UID: \"ee260316-013d-420c-8462-9919f6a10f12\") " pod="openshift-controller-manager/controller-manager-5975f76c65-dnkhj" Jan 21 11:16:09 crc kubenswrapper[4824]: I0121 11:16:09.082449 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g5w7m\" (UniqueName: \"kubernetes.io/projected/b3bae151-0d8c-4715-a57c-77a832079b43-kube-api-access-g5w7m\") pod \"route-controller-manager-5fdb5cfd56-mtp66\" (UID: \"b3bae151-0d8c-4715-a57c-77a832079b43\") " pod="openshift-route-controller-manager/route-controller-manager-5fdb5cfd56-mtp66" Jan 21 11:16:09 crc kubenswrapper[4824]: I0121 11:16:09.082480 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ee260316-013d-420c-8462-9919f6a10f12-client-ca\") pod \"controller-manager-5975f76c65-dnkhj\" (UID: \"ee260316-013d-420c-8462-9919f6a10f12\") " pod="openshift-controller-manager/controller-manager-5975f76c65-dnkhj" Jan 21 11:16:09 crc kubenswrapper[4824]: I0121 11:16:09.083275 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b3bae151-0d8c-4715-a57c-77a832079b43-client-ca\") pod \"route-controller-manager-5fdb5cfd56-mtp66\" (UID: \"b3bae151-0d8c-4715-a57c-77a832079b43\") " pod="openshift-route-controller-manager/route-controller-manager-5fdb5cfd56-mtp66" Jan 21 11:16:09 crc kubenswrapper[4824]: I0121 11:16:09.083488 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ee260316-013d-420c-8462-9919f6a10f12-client-ca\") pod \"controller-manager-5975f76c65-dnkhj\" (UID: \"ee260316-013d-420c-8462-9919f6a10f12\") " pod="openshift-controller-manager/controller-manager-5975f76c65-dnkhj" Jan 21 11:16:09 crc kubenswrapper[4824]: I0121 11:16:09.083695 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ee260316-013d-420c-8462-9919f6a10f12-config\") pod \"controller-manager-5975f76c65-dnkhj\" (UID: \"ee260316-013d-420c-8462-9919f6a10f12\") " pod="openshift-controller-manager/controller-manager-5975f76c65-dnkhj" Jan 21 11:16:09 crc kubenswrapper[4824]: I0121 11:16:09.083738 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b3bae151-0d8c-4715-a57c-77a832079b43-config\") pod \"route-controller-manager-5fdb5cfd56-mtp66\" (UID: \"b3bae151-0d8c-4715-a57c-77a832079b43\") " pod="openshift-route-controller-manager/route-controller-manager-5fdb5cfd56-mtp66" Jan 21 11:16:09 crc kubenswrapper[4824]: I0121 11:16:09.084494 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/ee260316-013d-420c-8462-9919f6a10f12-proxy-ca-bundles\") pod \"controller-manager-5975f76c65-dnkhj\" (UID: \"ee260316-013d-420c-8462-9919f6a10f12\") " pod="openshift-controller-manager/controller-manager-5975f76c65-dnkhj" Jan 21 11:16:09 crc kubenswrapper[4824]: I0121 11:16:09.085994 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b3bae151-0d8c-4715-a57c-77a832079b43-serving-cert\") pod \"route-controller-manager-5fdb5cfd56-mtp66\" (UID: \"b3bae151-0d8c-4715-a57c-77a832079b43\") " pod="openshift-route-controller-manager/route-controller-manager-5fdb5cfd56-mtp66" Jan 21 11:16:09 crc kubenswrapper[4824]: I0121 11:16:09.086662 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ee260316-013d-420c-8462-9919f6a10f12-serving-cert\") pod \"controller-manager-5975f76c65-dnkhj\" (UID: \"ee260316-013d-420c-8462-9919f6a10f12\") " pod="openshift-controller-manager/controller-manager-5975f76c65-dnkhj" Jan 21 11:16:09 crc kubenswrapper[4824]: I0121 11:16:09.094483 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fgcf4\" (UniqueName: \"kubernetes.io/projected/ee260316-013d-420c-8462-9919f6a10f12-kube-api-access-fgcf4\") pod \"controller-manager-5975f76c65-dnkhj\" (UID: \"ee260316-013d-420c-8462-9919f6a10f12\") " pod="openshift-controller-manager/controller-manager-5975f76c65-dnkhj" Jan 21 11:16:09 crc kubenswrapper[4824]: I0121 11:16:09.094782 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g5w7m\" (UniqueName: \"kubernetes.io/projected/b3bae151-0d8c-4715-a57c-77a832079b43-kube-api-access-g5w7m\") pod \"route-controller-manager-5fdb5cfd56-mtp66\" (UID: \"b3bae151-0d8c-4715-a57c-77a832079b43\") " pod="openshift-route-controller-manager/route-controller-manager-5fdb5cfd56-mtp66" Jan 21 11:16:09 crc kubenswrapper[4824]: I0121 11:16:09.249324 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5fdb5cfd56-mtp66" Jan 21 11:16:09 crc kubenswrapper[4824]: I0121 11:16:09.255365 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5975f76c65-dnkhj" Jan 21 11:16:09 crc kubenswrapper[4824]: I0121 11:16:09.588410 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-5975f76c65-dnkhj"] Jan 21 11:16:09 crc kubenswrapper[4824]: W0121 11:16:09.591374 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podee260316_013d_420c_8462_9919f6a10f12.slice/crio-13bc2579bffa5fa9f518ccd75c1971d4a17538aa01ff7ca03fe991bc54a382bf WatchSource:0}: Error finding container 13bc2579bffa5fa9f518ccd75c1971d4a17538aa01ff7ca03fe991bc54a382bf: Status 404 returned error can't find the container with id 13bc2579bffa5fa9f518ccd75c1971d4a17538aa01ff7ca03fe991bc54a382bf Jan 21 11:16:09 crc kubenswrapper[4824]: I0121 11:16:09.622120 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5fdb5cfd56-mtp66"] Jan 21 11:16:09 crc kubenswrapper[4824]: W0121 11:16:09.631515 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb3bae151_0d8c_4715_a57c_77a832079b43.slice/crio-35f60ff8859e4ce43d1479a7cb73566d5c30cc65d97effa7365b89c1c015acd7 WatchSource:0}: Error finding container 35f60ff8859e4ce43d1479a7cb73566d5c30cc65d97effa7365b89c1c015acd7: Status 404 returned error can't find the container with id 35f60ff8859e4ce43d1479a7cb73566d5c30cc65d97effa7365b89c1c015acd7 Jan 21 11:16:09 crc kubenswrapper[4824]: I0121 11:16:09.805354 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5fdb5cfd56-mtp66" event={"ID":"b3bae151-0d8c-4715-a57c-77a832079b43","Type":"ContainerStarted","Data":"9e977ab912f17990088ac141a2495425021fa0c75aea01ae0f65cd9dcdca4cec"} Jan 21 11:16:09 crc kubenswrapper[4824]: I0121 11:16:09.805389 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5fdb5cfd56-mtp66" event={"ID":"b3bae151-0d8c-4715-a57c-77a832079b43","Type":"ContainerStarted","Data":"35f60ff8859e4ce43d1479a7cb73566d5c30cc65d97effa7365b89c1c015acd7"} Jan 21 11:16:09 crc kubenswrapper[4824]: I0121 11:16:09.805497 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-5fdb5cfd56-mtp66" Jan 21 11:16:09 crc kubenswrapper[4824]: I0121 11:16:09.807389 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5975f76c65-dnkhj" event={"ID":"ee260316-013d-420c-8462-9919f6a10f12","Type":"ContainerStarted","Data":"d33c304d998b36c519097b8d550522dcb828ab9bce9778ca7b1d54534b00db6f"} Jan 21 11:16:09 crc kubenswrapper[4824]: I0121 11:16:09.807419 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5975f76c65-dnkhj" event={"ID":"ee260316-013d-420c-8462-9919f6a10f12","Type":"ContainerStarted","Data":"13bc2579bffa5fa9f518ccd75c1971d4a17538aa01ff7ca03fe991bc54a382bf"} Jan 21 11:16:09 crc kubenswrapper[4824]: I0121 11:16:09.807611 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-5975f76c65-dnkhj" Jan 21 11:16:09 crc kubenswrapper[4824]: I0121 11:16:09.821262 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-5975f76c65-dnkhj" Jan 21 11:16:09 crc kubenswrapper[4824]: I0121 11:16:09.831985 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-5975f76c65-dnkhj" podStartSLOduration=2.831970823 podStartE2EDuration="2.831970823s" podCreationTimestamp="2026-01-21 11:16:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:16:09.830791758 +0000 UTC m=+312.123821050" watchObservedRunningTime="2026-01-21 11:16:09.831970823 +0000 UTC m=+312.125000114" Jan 21 11:16:09 crc kubenswrapper[4824]: I0121 11:16:09.833192 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-5fdb5cfd56-mtp66" podStartSLOduration=2.833182288 podStartE2EDuration="2.833182288s" podCreationTimestamp="2026-01-21 11:16:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:16:09.821263922 +0000 UTC m=+312.114293214" watchObservedRunningTime="2026-01-21 11:16:09.833182288 +0000 UTC m=+312.126211580" Jan 21 11:16:10 crc kubenswrapper[4824]: I0121 11:16:10.054325 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3d23b1ac-adfc-4f40-8c75-710a71784b55" path="/var/lib/kubelet/pods/3d23b1ac-adfc-4f40-8c75-710a71784b55/volumes" Jan 21 11:16:10 crc kubenswrapper[4824]: I0121 11:16:10.054971 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="65ad4810-1d65-4965-b8c6-273069e2020f" path="/var/lib/kubelet/pods/65ad4810-1d65-4965-b8c6-273069e2020f/volumes" Jan 21 11:16:10 crc kubenswrapper[4824]: I0121 11:16:10.084428 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-5fdb5cfd56-mtp66" Jan 21 11:16:14 crc kubenswrapper[4824]: I0121 11:16:14.801586 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-r2kbm"] Jan 21 11:16:14 crc kubenswrapper[4824]: I0121 11:16:14.802517 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-r2kbm" Jan 21 11:16:14 crc kubenswrapper[4824]: I0121 11:16:14.818693 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-r2kbm"] Jan 21 11:16:14 crc kubenswrapper[4824]: I0121 11:16:14.945817 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mfj4n\" (UniqueName: \"kubernetes.io/projected/9cd26320-635a-4df5-ba9e-3a8e45d80c8b-kube-api-access-mfj4n\") pod \"image-registry-66df7c8f76-r2kbm\" (UID: \"9cd26320-635a-4df5-ba9e-3a8e45d80c8b\") " pod="openshift-image-registry/image-registry-66df7c8f76-r2kbm" Jan 21 11:16:14 crc kubenswrapper[4824]: I0121 11:16:14.945876 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-r2kbm\" (UID: \"9cd26320-635a-4df5-ba9e-3a8e45d80c8b\") " pod="openshift-image-registry/image-registry-66df7c8f76-r2kbm" Jan 21 11:16:14 crc kubenswrapper[4824]: I0121 11:16:14.945907 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/9cd26320-635a-4df5-ba9e-3a8e45d80c8b-registry-certificates\") pod \"image-registry-66df7c8f76-r2kbm\" (UID: \"9cd26320-635a-4df5-ba9e-3a8e45d80c8b\") " pod="openshift-image-registry/image-registry-66df7c8f76-r2kbm" Jan 21 11:16:14 crc kubenswrapper[4824]: I0121 11:16:14.945933 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9cd26320-635a-4df5-ba9e-3a8e45d80c8b-trusted-ca\") pod \"image-registry-66df7c8f76-r2kbm\" (UID: \"9cd26320-635a-4df5-ba9e-3a8e45d80c8b\") " pod="openshift-image-registry/image-registry-66df7c8f76-r2kbm" Jan 21 11:16:14 crc kubenswrapper[4824]: I0121 11:16:14.946008 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/9cd26320-635a-4df5-ba9e-3a8e45d80c8b-bound-sa-token\") pod \"image-registry-66df7c8f76-r2kbm\" (UID: \"9cd26320-635a-4df5-ba9e-3a8e45d80c8b\") " pod="openshift-image-registry/image-registry-66df7c8f76-r2kbm" Jan 21 11:16:14 crc kubenswrapper[4824]: I0121 11:16:14.946148 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/9cd26320-635a-4df5-ba9e-3a8e45d80c8b-installation-pull-secrets\") pod \"image-registry-66df7c8f76-r2kbm\" (UID: \"9cd26320-635a-4df5-ba9e-3a8e45d80c8b\") " pod="openshift-image-registry/image-registry-66df7c8f76-r2kbm" Jan 21 11:16:14 crc kubenswrapper[4824]: I0121 11:16:14.946173 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/9cd26320-635a-4df5-ba9e-3a8e45d80c8b-registry-tls\") pod \"image-registry-66df7c8f76-r2kbm\" (UID: \"9cd26320-635a-4df5-ba9e-3a8e45d80c8b\") " pod="openshift-image-registry/image-registry-66df7c8f76-r2kbm" Jan 21 11:16:14 crc kubenswrapper[4824]: I0121 11:16:14.946195 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/9cd26320-635a-4df5-ba9e-3a8e45d80c8b-ca-trust-extracted\") pod \"image-registry-66df7c8f76-r2kbm\" (UID: \"9cd26320-635a-4df5-ba9e-3a8e45d80c8b\") " pod="openshift-image-registry/image-registry-66df7c8f76-r2kbm" Jan 21 11:16:14 crc kubenswrapper[4824]: I0121 11:16:14.964239 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-r2kbm\" (UID: \"9cd26320-635a-4df5-ba9e-3a8e45d80c8b\") " pod="openshift-image-registry/image-registry-66df7c8f76-r2kbm" Jan 21 11:16:15 crc kubenswrapper[4824]: I0121 11:16:15.047484 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mfj4n\" (UniqueName: \"kubernetes.io/projected/9cd26320-635a-4df5-ba9e-3a8e45d80c8b-kube-api-access-mfj4n\") pod \"image-registry-66df7c8f76-r2kbm\" (UID: \"9cd26320-635a-4df5-ba9e-3a8e45d80c8b\") " pod="openshift-image-registry/image-registry-66df7c8f76-r2kbm" Jan 21 11:16:15 crc kubenswrapper[4824]: I0121 11:16:15.047552 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/9cd26320-635a-4df5-ba9e-3a8e45d80c8b-registry-certificates\") pod \"image-registry-66df7c8f76-r2kbm\" (UID: \"9cd26320-635a-4df5-ba9e-3a8e45d80c8b\") " pod="openshift-image-registry/image-registry-66df7c8f76-r2kbm" Jan 21 11:16:15 crc kubenswrapper[4824]: I0121 11:16:15.047580 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9cd26320-635a-4df5-ba9e-3a8e45d80c8b-trusted-ca\") pod \"image-registry-66df7c8f76-r2kbm\" (UID: \"9cd26320-635a-4df5-ba9e-3a8e45d80c8b\") " pod="openshift-image-registry/image-registry-66df7c8f76-r2kbm" Jan 21 11:16:15 crc kubenswrapper[4824]: I0121 11:16:15.047596 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/9cd26320-635a-4df5-ba9e-3a8e45d80c8b-bound-sa-token\") pod \"image-registry-66df7c8f76-r2kbm\" (UID: \"9cd26320-635a-4df5-ba9e-3a8e45d80c8b\") " pod="openshift-image-registry/image-registry-66df7c8f76-r2kbm" Jan 21 11:16:15 crc kubenswrapper[4824]: I0121 11:16:15.047665 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/9cd26320-635a-4df5-ba9e-3a8e45d80c8b-installation-pull-secrets\") pod \"image-registry-66df7c8f76-r2kbm\" (UID: \"9cd26320-635a-4df5-ba9e-3a8e45d80c8b\") " pod="openshift-image-registry/image-registry-66df7c8f76-r2kbm" Jan 21 11:16:15 crc kubenswrapper[4824]: I0121 11:16:15.047682 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/9cd26320-635a-4df5-ba9e-3a8e45d80c8b-registry-tls\") pod \"image-registry-66df7c8f76-r2kbm\" (UID: \"9cd26320-635a-4df5-ba9e-3a8e45d80c8b\") " pod="openshift-image-registry/image-registry-66df7c8f76-r2kbm" Jan 21 11:16:15 crc kubenswrapper[4824]: I0121 11:16:15.047705 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/9cd26320-635a-4df5-ba9e-3a8e45d80c8b-ca-trust-extracted\") pod \"image-registry-66df7c8f76-r2kbm\" (UID: \"9cd26320-635a-4df5-ba9e-3a8e45d80c8b\") " pod="openshift-image-registry/image-registry-66df7c8f76-r2kbm" Jan 21 11:16:15 crc kubenswrapper[4824]: I0121 11:16:15.048196 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/9cd26320-635a-4df5-ba9e-3a8e45d80c8b-ca-trust-extracted\") pod \"image-registry-66df7c8f76-r2kbm\" (UID: \"9cd26320-635a-4df5-ba9e-3a8e45d80c8b\") " pod="openshift-image-registry/image-registry-66df7c8f76-r2kbm" Jan 21 11:16:15 crc kubenswrapper[4824]: I0121 11:16:15.048650 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/9cd26320-635a-4df5-ba9e-3a8e45d80c8b-registry-certificates\") pod \"image-registry-66df7c8f76-r2kbm\" (UID: \"9cd26320-635a-4df5-ba9e-3a8e45d80c8b\") " pod="openshift-image-registry/image-registry-66df7c8f76-r2kbm" Jan 21 11:16:15 crc kubenswrapper[4824]: I0121 11:16:15.048649 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9cd26320-635a-4df5-ba9e-3a8e45d80c8b-trusted-ca\") pod \"image-registry-66df7c8f76-r2kbm\" (UID: \"9cd26320-635a-4df5-ba9e-3a8e45d80c8b\") " pod="openshift-image-registry/image-registry-66df7c8f76-r2kbm" Jan 21 11:16:15 crc kubenswrapper[4824]: I0121 11:16:15.052576 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/9cd26320-635a-4df5-ba9e-3a8e45d80c8b-installation-pull-secrets\") pod \"image-registry-66df7c8f76-r2kbm\" (UID: \"9cd26320-635a-4df5-ba9e-3a8e45d80c8b\") " pod="openshift-image-registry/image-registry-66df7c8f76-r2kbm" Jan 21 11:16:15 crc kubenswrapper[4824]: I0121 11:16:15.052602 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/9cd26320-635a-4df5-ba9e-3a8e45d80c8b-registry-tls\") pod \"image-registry-66df7c8f76-r2kbm\" (UID: \"9cd26320-635a-4df5-ba9e-3a8e45d80c8b\") " pod="openshift-image-registry/image-registry-66df7c8f76-r2kbm" Jan 21 11:16:15 crc kubenswrapper[4824]: I0121 11:16:15.066639 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mfj4n\" (UniqueName: \"kubernetes.io/projected/9cd26320-635a-4df5-ba9e-3a8e45d80c8b-kube-api-access-mfj4n\") pod \"image-registry-66df7c8f76-r2kbm\" (UID: \"9cd26320-635a-4df5-ba9e-3a8e45d80c8b\") " pod="openshift-image-registry/image-registry-66df7c8f76-r2kbm" Jan 21 11:16:15 crc kubenswrapper[4824]: I0121 11:16:15.072224 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/9cd26320-635a-4df5-ba9e-3a8e45d80c8b-bound-sa-token\") pod \"image-registry-66df7c8f76-r2kbm\" (UID: \"9cd26320-635a-4df5-ba9e-3a8e45d80c8b\") " pod="openshift-image-registry/image-registry-66df7c8f76-r2kbm" Jan 21 11:16:15 crc kubenswrapper[4824]: I0121 11:16:15.116204 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-r2kbm" Jan 21 11:16:15 crc kubenswrapper[4824]: I0121 11:16:15.460776 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-r2kbm"] Jan 21 11:16:15 crc kubenswrapper[4824]: I0121 11:16:15.829372 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-r2kbm" event={"ID":"9cd26320-635a-4df5-ba9e-3a8e45d80c8b","Type":"ContainerStarted","Data":"7f93526973f872e8f6481d2dcb95cd785ca23649c3ced89b7fb423c2be014890"} Jan 21 11:16:15 crc kubenswrapper[4824]: I0121 11:16:15.829547 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-r2kbm" event={"ID":"9cd26320-635a-4df5-ba9e-3a8e45d80c8b","Type":"ContainerStarted","Data":"4ff9af531299c805524fe70e5d120d9b9fd04dd2983e7f0d6803f2117eb65f26"} Jan 21 11:16:15 crc kubenswrapper[4824]: I0121 11:16:15.829561 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-r2kbm" Jan 21 11:16:15 crc kubenswrapper[4824]: I0121 11:16:15.842569 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-r2kbm" podStartSLOduration=1.84255385 podStartE2EDuration="1.84255385s" podCreationTimestamp="2026-01-21 11:16:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:16:15.84085029 +0000 UTC m=+318.133879582" watchObservedRunningTime="2026-01-21 11:16:15.84255385 +0000 UTC m=+318.135583142" Jan 21 11:16:35 crc kubenswrapper[4824]: I0121 11:16:35.120131 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-r2kbm" Jan 21 11:16:35 crc kubenswrapper[4824]: I0121 11:16:35.151163 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-wxlxx"] Jan 21 11:16:46 crc kubenswrapper[4824]: I0121 11:16:46.065219 4824 patch_prober.go:28] interesting pod/machine-config-daemon-zqjg9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 21 11:16:46 crc kubenswrapper[4824]: I0121 11:16:46.065592 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 21 11:17:00 crc kubenswrapper[4824]: I0121 11:17:00.174775 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-wxlxx" podUID="94e71514-0af6-4254-8815-c933ade6c9da" containerName="registry" containerID="cri-o://6c3387d6526e3b087ac4042dcb88df75f294f6b26a6635786027622ac70f54bf" gracePeriod=30 Jan 21 11:17:00 crc kubenswrapper[4824]: I0121 11:17:00.472688 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-wxlxx" Jan 21 11:17:00 crc kubenswrapper[4824]: I0121 11:17:00.552763 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/94e71514-0af6-4254-8815-c933ade6c9da-registry-tls\") pod \"94e71514-0af6-4254-8815-c933ade6c9da\" (UID: \"94e71514-0af6-4254-8815-c933ade6c9da\") " Jan 21 11:17:00 crc kubenswrapper[4824]: I0121 11:17:00.552810 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/94e71514-0af6-4254-8815-c933ade6c9da-trusted-ca\") pod \"94e71514-0af6-4254-8815-c933ade6c9da\" (UID: \"94e71514-0af6-4254-8815-c933ade6c9da\") " Jan 21 11:17:00 crc kubenswrapper[4824]: I0121 11:17:00.552831 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/94e71514-0af6-4254-8815-c933ade6c9da-ca-trust-extracted\") pod \"94e71514-0af6-4254-8815-c933ade6c9da\" (UID: \"94e71514-0af6-4254-8815-c933ade6c9da\") " Jan 21 11:17:00 crc kubenswrapper[4824]: I0121 11:17:00.552858 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/94e71514-0af6-4254-8815-c933ade6c9da-bound-sa-token\") pod \"94e71514-0af6-4254-8815-c933ade6c9da\" (UID: \"94e71514-0af6-4254-8815-c933ade6c9da\") " Jan 21 11:17:00 crc kubenswrapper[4824]: I0121 11:17:00.552885 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/94e71514-0af6-4254-8815-c933ade6c9da-installation-pull-secrets\") pod \"94e71514-0af6-4254-8815-c933ade6c9da\" (UID: \"94e71514-0af6-4254-8815-c933ade6c9da\") " Jan 21 11:17:00 crc kubenswrapper[4824]: I0121 11:17:00.552905 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/94e71514-0af6-4254-8815-c933ade6c9da-registry-certificates\") pod \"94e71514-0af6-4254-8815-c933ade6c9da\" (UID: \"94e71514-0af6-4254-8815-c933ade6c9da\") " Jan 21 11:17:00 crc kubenswrapper[4824]: I0121 11:17:00.553740 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"94e71514-0af6-4254-8815-c933ade6c9da\" (UID: \"94e71514-0af6-4254-8815-c933ade6c9da\") " Jan 21 11:17:00 crc kubenswrapper[4824]: I0121 11:17:00.554091 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zcznm\" (UniqueName: \"kubernetes.io/projected/94e71514-0af6-4254-8815-c933ade6c9da-kube-api-access-zcznm\") pod \"94e71514-0af6-4254-8815-c933ade6c9da\" (UID: \"94e71514-0af6-4254-8815-c933ade6c9da\") " Jan 21 11:17:00 crc kubenswrapper[4824]: I0121 11:17:00.553768 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/94e71514-0af6-4254-8815-c933ade6c9da-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "94e71514-0af6-4254-8815-c933ade6c9da" (UID: "94e71514-0af6-4254-8815-c933ade6c9da"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:17:00 crc kubenswrapper[4824]: I0121 11:17:00.553809 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/94e71514-0af6-4254-8815-c933ade6c9da-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "94e71514-0af6-4254-8815-c933ade6c9da" (UID: "94e71514-0af6-4254-8815-c933ade6c9da"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:17:00 crc kubenswrapper[4824]: I0121 11:17:00.554419 4824 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/94e71514-0af6-4254-8815-c933ade6c9da-trusted-ca\") on node \"crc\" DevicePath \"\"" Jan 21 11:17:00 crc kubenswrapper[4824]: I0121 11:17:00.554436 4824 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/94e71514-0af6-4254-8815-c933ade6c9da-registry-certificates\") on node \"crc\" DevicePath \"\"" Jan 21 11:17:00 crc kubenswrapper[4824]: I0121 11:17:00.563248 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "94e71514-0af6-4254-8815-c933ade6c9da" (UID: "94e71514-0af6-4254-8815-c933ade6c9da"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Jan 21 11:17:00 crc kubenswrapper[4824]: I0121 11:17:00.563264 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/94e71514-0af6-4254-8815-c933ade6c9da-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "94e71514-0af6-4254-8815-c933ade6c9da" (UID: "94e71514-0af6-4254-8815-c933ade6c9da"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:17:00 crc kubenswrapper[4824]: I0121 11:17:00.563534 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/94e71514-0af6-4254-8815-c933ade6c9da-kube-api-access-zcznm" (OuterVolumeSpecName: "kube-api-access-zcznm") pod "94e71514-0af6-4254-8815-c933ade6c9da" (UID: "94e71514-0af6-4254-8815-c933ade6c9da"). InnerVolumeSpecName "kube-api-access-zcznm". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:17:00 crc kubenswrapper[4824]: I0121 11:17:00.563930 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/94e71514-0af6-4254-8815-c933ade6c9da-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "94e71514-0af6-4254-8815-c933ade6c9da" (UID: "94e71514-0af6-4254-8815-c933ade6c9da"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:17:00 crc kubenswrapper[4824]: I0121 11:17:00.564432 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/94e71514-0af6-4254-8815-c933ade6c9da-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "94e71514-0af6-4254-8815-c933ade6c9da" (UID: "94e71514-0af6-4254-8815-c933ade6c9da"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:17:00 crc kubenswrapper[4824]: I0121 11:17:00.565977 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/94e71514-0af6-4254-8815-c933ade6c9da-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "94e71514-0af6-4254-8815-c933ade6c9da" (UID: "94e71514-0af6-4254-8815-c933ade6c9da"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:17:00 crc kubenswrapper[4824]: I0121 11:17:00.655810 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zcznm\" (UniqueName: \"kubernetes.io/projected/94e71514-0af6-4254-8815-c933ade6c9da-kube-api-access-zcznm\") on node \"crc\" DevicePath \"\"" Jan 21 11:17:00 crc kubenswrapper[4824]: I0121 11:17:00.655837 4824 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/94e71514-0af6-4254-8815-c933ade6c9da-registry-tls\") on node \"crc\" DevicePath \"\"" Jan 21 11:17:00 crc kubenswrapper[4824]: I0121 11:17:00.655847 4824 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/94e71514-0af6-4254-8815-c933ade6c9da-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Jan 21 11:17:00 crc kubenswrapper[4824]: I0121 11:17:00.655855 4824 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/94e71514-0af6-4254-8815-c933ade6c9da-bound-sa-token\") on node \"crc\" DevicePath \"\"" Jan 21 11:17:00 crc kubenswrapper[4824]: I0121 11:17:00.655863 4824 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/94e71514-0af6-4254-8815-c933ade6c9da-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Jan 21 11:17:01 crc kubenswrapper[4824]: I0121 11:17:01.000931 4824 generic.go:334] "Generic (PLEG): container finished" podID="94e71514-0af6-4254-8815-c933ade6c9da" containerID="6c3387d6526e3b087ac4042dcb88df75f294f6b26a6635786027622ac70f54bf" exitCode=0 Jan 21 11:17:01 crc kubenswrapper[4824]: I0121 11:17:01.000989 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-wxlxx" event={"ID":"94e71514-0af6-4254-8815-c933ade6c9da","Type":"ContainerDied","Data":"6c3387d6526e3b087ac4042dcb88df75f294f6b26a6635786027622ac70f54bf"} Jan 21 11:17:01 crc kubenswrapper[4824]: I0121 11:17:01.001040 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-wxlxx" event={"ID":"94e71514-0af6-4254-8815-c933ade6c9da","Type":"ContainerDied","Data":"7d8d535dae1a2ab198548905f9a9571895c25a2b895cfb327e93cc9833852054"} Jan 21 11:17:01 crc kubenswrapper[4824]: I0121 11:17:01.001060 4824 scope.go:117] "RemoveContainer" containerID="6c3387d6526e3b087ac4042dcb88df75f294f6b26a6635786027622ac70f54bf" Jan 21 11:17:01 crc kubenswrapper[4824]: I0121 11:17:01.002496 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-wxlxx" Jan 21 11:17:01 crc kubenswrapper[4824]: I0121 11:17:01.012376 4824 scope.go:117] "RemoveContainer" containerID="6c3387d6526e3b087ac4042dcb88df75f294f6b26a6635786027622ac70f54bf" Jan 21 11:17:01 crc kubenswrapper[4824]: E0121 11:17:01.012641 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6c3387d6526e3b087ac4042dcb88df75f294f6b26a6635786027622ac70f54bf\": container with ID starting with 6c3387d6526e3b087ac4042dcb88df75f294f6b26a6635786027622ac70f54bf not found: ID does not exist" containerID="6c3387d6526e3b087ac4042dcb88df75f294f6b26a6635786027622ac70f54bf" Jan 21 11:17:01 crc kubenswrapper[4824]: I0121 11:17:01.012669 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6c3387d6526e3b087ac4042dcb88df75f294f6b26a6635786027622ac70f54bf"} err="failed to get container status \"6c3387d6526e3b087ac4042dcb88df75f294f6b26a6635786027622ac70f54bf\": rpc error: code = NotFound desc = could not find container \"6c3387d6526e3b087ac4042dcb88df75f294f6b26a6635786027622ac70f54bf\": container with ID starting with 6c3387d6526e3b087ac4042dcb88df75f294f6b26a6635786027622ac70f54bf not found: ID does not exist" Jan 21 11:17:01 crc kubenswrapper[4824]: I0121 11:17:01.021267 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-wxlxx"] Jan 21 11:17:01 crc kubenswrapper[4824]: I0121 11:17:01.023240 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-wxlxx"] Jan 21 11:17:02 crc kubenswrapper[4824]: I0121 11:17:02.053934 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="94e71514-0af6-4254-8815-c933ade6c9da" path="/var/lib/kubelet/pods/94e71514-0af6-4254-8815-c933ade6c9da/volumes" Jan 21 11:17:16 crc kubenswrapper[4824]: I0121 11:17:16.065302 4824 patch_prober.go:28] interesting pod/machine-config-daemon-zqjg9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 21 11:17:16 crc kubenswrapper[4824]: I0121 11:17:16.065673 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 21 11:17:46 crc kubenswrapper[4824]: I0121 11:17:46.065178 4824 patch_prober.go:28] interesting pod/machine-config-daemon-zqjg9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 21 11:17:46 crc kubenswrapper[4824]: I0121 11:17:46.065480 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 21 11:17:46 crc kubenswrapper[4824]: I0121 11:17:46.065528 4824 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" Jan 21 11:17:46 crc kubenswrapper[4824]: I0121 11:17:46.066001 4824 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"04e919af77a770ce30313c5a8770baf390f15a9d134ed04fa80518ff38656c8a"} pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 21 11:17:46 crc kubenswrapper[4824]: I0121 11:17:46.066065 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" containerName="machine-config-daemon" containerID="cri-o://04e919af77a770ce30313c5a8770baf390f15a9d134ed04fa80518ff38656c8a" gracePeriod=600 Jan 21 11:17:47 crc kubenswrapper[4824]: I0121 11:17:47.159525 4824 generic.go:334] "Generic (PLEG): container finished" podID="33f3d922-4ffe-409b-a49a-d88c85898260" containerID="04e919af77a770ce30313c5a8770baf390f15a9d134ed04fa80518ff38656c8a" exitCode=0 Jan 21 11:17:47 crc kubenswrapper[4824]: I0121 11:17:47.159599 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" event={"ID":"33f3d922-4ffe-409b-a49a-d88c85898260","Type":"ContainerDied","Data":"04e919af77a770ce30313c5a8770baf390f15a9d134ed04fa80518ff38656c8a"} Jan 21 11:17:47 crc kubenswrapper[4824]: I0121 11:17:47.159756 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" event={"ID":"33f3d922-4ffe-409b-a49a-d88c85898260","Type":"ContainerStarted","Data":"0494d41399c844b692156a76450e15e5791154f06b98545910f6ac8899bc8509"} Jan 21 11:17:47 crc kubenswrapper[4824]: I0121 11:17:47.159777 4824 scope.go:117] "RemoveContainer" containerID="2b1b37c6d2994021cfc30f85e92c40ae35677113b17aa066d52bc7a76ae5c153" Jan 21 11:19:46 crc kubenswrapper[4824]: I0121 11:19:46.064988 4824 patch_prober.go:28] interesting pod/machine-config-daemon-zqjg9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 21 11:19:46 crc kubenswrapper[4824]: I0121 11:19:46.065331 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 21 11:19:58 crc kubenswrapper[4824]: I0121 11:19:58.162712 4824 scope.go:117] "RemoveContainer" containerID="e964ec9dc2f3dad14f23b05ef15becda9a72c91eb5ec8bd728789277a97681be" Jan 21 11:19:58 crc kubenswrapper[4824]: I0121 11:19:58.177695 4824 scope.go:117] "RemoveContainer" containerID="812b638a67ec88d3b10766fb23536e55485cc458e7ca4b815bcac4f9331b7868" Jan 21 11:19:58 crc kubenswrapper[4824]: I0121 11:19:58.189515 4824 scope.go:117] "RemoveContainer" containerID="75d2b0ba70b6aa340dd2f71ac90a2f4234cfa6ec869faff5f999306922a53203" Jan 21 11:20:15 crc kubenswrapper[4824]: I0121 11:20:15.146348 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-cf98fcc89-xjwkp"] Jan 21 11:20:15 crc kubenswrapper[4824]: E0121 11:20:15.146929 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="94e71514-0af6-4254-8815-c933ade6c9da" containerName="registry" Jan 21 11:20:15 crc kubenswrapper[4824]: I0121 11:20:15.146943 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="94e71514-0af6-4254-8815-c933ade6c9da" containerName="registry" Jan 21 11:20:15 crc kubenswrapper[4824]: I0121 11:20:15.147101 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="94e71514-0af6-4254-8815-c933ade6c9da" containerName="registry" Jan 21 11:20:15 crc kubenswrapper[4824]: I0121 11:20:15.147454 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-cf98fcc89-xjwkp" Jan 21 11:20:15 crc kubenswrapper[4824]: I0121 11:20:15.148860 4824 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-qds56" Jan 21 11:20:15 crc kubenswrapper[4824]: I0121 11:20:15.149052 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Jan 21 11:20:15 crc kubenswrapper[4824]: I0121 11:20:15.149542 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Jan 21 11:20:15 crc kubenswrapper[4824]: I0121 11:20:15.151731 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-858654f9db-qhlzr"] Jan 21 11:20:15 crc kubenswrapper[4824]: I0121 11:20:15.152315 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-858654f9db-qhlzr" Jan 21 11:20:15 crc kubenswrapper[4824]: I0121 11:20:15.153392 4824 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-mwbvc" Jan 21 11:20:15 crc kubenswrapper[4824]: I0121 11:20:15.158893 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-cf98fcc89-xjwkp"] Jan 21 11:20:15 crc kubenswrapper[4824]: I0121 11:20:15.162347 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-687f57d79b-r6qj7"] Jan 21 11:20:15 crc kubenswrapper[4824]: I0121 11:20:15.162923 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-687f57d79b-r6qj7" Jan 21 11:20:15 crc kubenswrapper[4824]: I0121 11:20:15.164219 4824 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-gn9gr" Jan 21 11:20:15 crc kubenswrapper[4824]: I0121 11:20:15.165131 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-858654f9db-qhlzr"] Jan 21 11:20:15 crc kubenswrapper[4824]: I0121 11:20:15.199718 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-687f57d79b-r6qj7"] Jan 21 11:20:15 crc kubenswrapper[4824]: I0121 11:20:15.341172 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qp54q\" (UniqueName: \"kubernetes.io/projected/b12640f9-bf75-4056-9c20-a75d63a0e177-kube-api-access-qp54q\") pod \"cert-manager-webhook-687f57d79b-r6qj7\" (UID: \"b12640f9-bf75-4056-9c20-a75d63a0e177\") " pod="cert-manager/cert-manager-webhook-687f57d79b-r6qj7" Jan 21 11:20:15 crc kubenswrapper[4824]: I0121 11:20:15.341236 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6c7pg\" (UniqueName: \"kubernetes.io/projected/b9405067-2dfa-46f9-9471-fe456b3ca5aa-kube-api-access-6c7pg\") pod \"cert-manager-858654f9db-qhlzr\" (UID: \"b9405067-2dfa-46f9-9471-fe456b3ca5aa\") " pod="cert-manager/cert-manager-858654f9db-qhlzr" Jan 21 11:20:15 crc kubenswrapper[4824]: I0121 11:20:15.341336 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2kvbn\" (UniqueName: \"kubernetes.io/projected/d783bb77-baab-4e1c-99e9-ddba7ef78fc2-kube-api-access-2kvbn\") pod \"cert-manager-cainjector-cf98fcc89-xjwkp\" (UID: \"d783bb77-baab-4e1c-99e9-ddba7ef78fc2\") " pod="cert-manager/cert-manager-cainjector-cf98fcc89-xjwkp" Jan 21 11:20:15 crc kubenswrapper[4824]: I0121 11:20:15.442220 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2kvbn\" (UniqueName: \"kubernetes.io/projected/d783bb77-baab-4e1c-99e9-ddba7ef78fc2-kube-api-access-2kvbn\") pod \"cert-manager-cainjector-cf98fcc89-xjwkp\" (UID: \"d783bb77-baab-4e1c-99e9-ddba7ef78fc2\") " pod="cert-manager/cert-manager-cainjector-cf98fcc89-xjwkp" Jan 21 11:20:15 crc kubenswrapper[4824]: I0121 11:20:15.442334 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qp54q\" (UniqueName: \"kubernetes.io/projected/b12640f9-bf75-4056-9c20-a75d63a0e177-kube-api-access-qp54q\") pod \"cert-manager-webhook-687f57d79b-r6qj7\" (UID: \"b12640f9-bf75-4056-9c20-a75d63a0e177\") " pod="cert-manager/cert-manager-webhook-687f57d79b-r6qj7" Jan 21 11:20:15 crc kubenswrapper[4824]: I0121 11:20:15.442383 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6c7pg\" (UniqueName: \"kubernetes.io/projected/b9405067-2dfa-46f9-9471-fe456b3ca5aa-kube-api-access-6c7pg\") pod \"cert-manager-858654f9db-qhlzr\" (UID: \"b9405067-2dfa-46f9-9471-fe456b3ca5aa\") " pod="cert-manager/cert-manager-858654f9db-qhlzr" Jan 21 11:20:15 crc kubenswrapper[4824]: I0121 11:20:15.458654 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6c7pg\" (UniqueName: \"kubernetes.io/projected/b9405067-2dfa-46f9-9471-fe456b3ca5aa-kube-api-access-6c7pg\") pod \"cert-manager-858654f9db-qhlzr\" (UID: \"b9405067-2dfa-46f9-9471-fe456b3ca5aa\") " pod="cert-manager/cert-manager-858654f9db-qhlzr" Jan 21 11:20:15 crc kubenswrapper[4824]: I0121 11:20:15.459666 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qp54q\" (UniqueName: \"kubernetes.io/projected/b12640f9-bf75-4056-9c20-a75d63a0e177-kube-api-access-qp54q\") pod \"cert-manager-webhook-687f57d79b-r6qj7\" (UID: \"b12640f9-bf75-4056-9c20-a75d63a0e177\") " pod="cert-manager/cert-manager-webhook-687f57d79b-r6qj7" Jan 21 11:20:15 crc kubenswrapper[4824]: I0121 11:20:15.459680 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2kvbn\" (UniqueName: \"kubernetes.io/projected/d783bb77-baab-4e1c-99e9-ddba7ef78fc2-kube-api-access-2kvbn\") pod \"cert-manager-cainjector-cf98fcc89-xjwkp\" (UID: \"d783bb77-baab-4e1c-99e9-ddba7ef78fc2\") " pod="cert-manager/cert-manager-cainjector-cf98fcc89-xjwkp" Jan 21 11:20:15 crc kubenswrapper[4824]: I0121 11:20:15.461505 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-cf98fcc89-xjwkp" Jan 21 11:20:15 crc kubenswrapper[4824]: I0121 11:20:15.470551 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-858654f9db-qhlzr" Jan 21 11:20:15 crc kubenswrapper[4824]: I0121 11:20:15.476349 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-687f57d79b-r6qj7" Jan 21 11:20:15 crc kubenswrapper[4824]: I0121 11:20:15.813643 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-cf98fcc89-xjwkp"] Jan 21 11:20:15 crc kubenswrapper[4824]: I0121 11:20:15.820250 4824 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 21 11:20:15 crc kubenswrapper[4824]: I0121 11:20:15.852998 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-858654f9db-qhlzr"] Jan 21 11:20:15 crc kubenswrapper[4824]: W0121 11:20:15.858018 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb9405067_2dfa_46f9_9471_fe456b3ca5aa.slice/crio-28e66bc3cbeb909b4fb3ac4e4f61ab4e329846378d4543ec2b96cd9fdfa8db70 WatchSource:0}: Error finding container 28e66bc3cbeb909b4fb3ac4e4f61ab4e329846378d4543ec2b96cd9fdfa8db70: Status 404 returned error can't find the container with id 28e66bc3cbeb909b4fb3ac4e4f61ab4e329846378d4543ec2b96cd9fdfa8db70 Jan 21 11:20:15 crc kubenswrapper[4824]: I0121 11:20:15.875187 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-687f57d79b-r6qj7"] Jan 21 11:20:15 crc kubenswrapper[4824]: W0121 11:20:15.896334 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb12640f9_bf75_4056_9c20_a75d63a0e177.slice/crio-b1517ed0a3cfaa749f6d4f7dc9db45c31ec848c0133cd19151049d63d2d45be3 WatchSource:0}: Error finding container b1517ed0a3cfaa749f6d4f7dc9db45c31ec848c0133cd19151049d63d2d45be3: Status 404 returned error can't find the container with id b1517ed0a3cfaa749f6d4f7dc9db45c31ec848c0133cd19151049d63d2d45be3 Jan 21 11:20:16 crc kubenswrapper[4824]: I0121 11:20:16.065084 4824 patch_prober.go:28] interesting pod/machine-config-daemon-zqjg9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 21 11:20:16 crc kubenswrapper[4824]: I0121 11:20:16.065128 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 21 11:20:16 crc kubenswrapper[4824]: I0121 11:20:16.736034 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-cf98fcc89-xjwkp" event={"ID":"d783bb77-baab-4e1c-99e9-ddba7ef78fc2","Type":"ContainerStarted","Data":"75d0a83e78ed8aa21e24436721746b82415add7b87f03e4bebcd002f4587d885"} Jan 21 11:20:16 crc kubenswrapper[4824]: I0121 11:20:16.736977 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-858654f9db-qhlzr" event={"ID":"b9405067-2dfa-46f9-9471-fe456b3ca5aa","Type":"ContainerStarted","Data":"28e66bc3cbeb909b4fb3ac4e4f61ab4e329846378d4543ec2b96cd9fdfa8db70"} Jan 21 11:20:16 crc kubenswrapper[4824]: I0121 11:20:16.738077 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-687f57d79b-r6qj7" event={"ID":"b12640f9-bf75-4056-9c20-a75d63a0e177","Type":"ContainerStarted","Data":"b1517ed0a3cfaa749f6d4f7dc9db45c31ec848c0133cd19151049d63d2d45be3"} Jan 21 11:20:18 crc kubenswrapper[4824]: I0121 11:20:18.749231 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-cf98fcc89-xjwkp" event={"ID":"d783bb77-baab-4e1c-99e9-ddba7ef78fc2","Type":"ContainerStarted","Data":"4d041b2f34851fd9a7b84d0a992dcaddfda503a570a95e13ab3cd028a73af7b5"} Jan 21 11:20:18 crc kubenswrapper[4824]: I0121 11:20:18.750605 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-858654f9db-qhlzr" event={"ID":"b9405067-2dfa-46f9-9471-fe456b3ca5aa","Type":"ContainerStarted","Data":"c22e6735006a7cf615e2c6c3f3835939e153c7d76b78ede8f850ddea1e043bd6"} Jan 21 11:20:18 crc kubenswrapper[4824]: I0121 11:20:18.752154 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-687f57d79b-r6qj7" event={"ID":"b12640f9-bf75-4056-9c20-a75d63a0e177","Type":"ContainerStarted","Data":"35dc60dcf46360dc9d6bc1df7c69987e9a40e749194976d90d26ec7dfa44a288"} Jan 21 11:20:18 crc kubenswrapper[4824]: I0121 11:20:18.752277 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-687f57d79b-r6qj7" Jan 21 11:20:18 crc kubenswrapper[4824]: I0121 11:20:18.775329 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-cf98fcc89-xjwkp" podStartSLOduration=1.00580889 podStartE2EDuration="3.775312004s" podCreationTimestamp="2026-01-21 11:20:15 +0000 UTC" firstStartedPulling="2026-01-21 11:20:15.820053051 +0000 UTC m=+558.113082344" lastFinishedPulling="2026-01-21 11:20:18.589556166 +0000 UTC m=+560.882585458" observedRunningTime="2026-01-21 11:20:18.771884718 +0000 UTC m=+561.064914010" watchObservedRunningTime="2026-01-21 11:20:18.775312004 +0000 UTC m=+561.068341296" Jan 21 11:20:18 crc kubenswrapper[4824]: I0121 11:20:18.795491 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-687f57d79b-r6qj7" podStartSLOduration=1.08249887 podStartE2EDuration="3.795474777s" podCreationTimestamp="2026-01-21 11:20:15 +0000 UTC" firstStartedPulling="2026-01-21 11:20:15.902882567 +0000 UTC m=+558.195911859" lastFinishedPulling="2026-01-21 11:20:18.615858464 +0000 UTC m=+560.908887766" observedRunningTime="2026-01-21 11:20:18.794617602 +0000 UTC m=+561.087646894" watchObservedRunningTime="2026-01-21 11:20:18.795474777 +0000 UTC m=+561.088504069" Jan 21 11:20:18 crc kubenswrapper[4824]: I0121 11:20:18.816240 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-858654f9db-qhlzr" podStartSLOduration=1.085658702 podStartE2EDuration="3.816218413s" podCreationTimestamp="2026-01-21 11:20:15 +0000 UTC" firstStartedPulling="2026-01-21 11:20:15.859708225 +0000 UTC m=+558.152737517" lastFinishedPulling="2026-01-21 11:20:18.590267937 +0000 UTC m=+560.883297228" observedRunningTime="2026-01-21 11:20:18.81577043 +0000 UTC m=+561.108799723" watchObservedRunningTime="2026-01-21 11:20:18.816218413 +0000 UTC m=+561.109247706" Jan 21 11:20:25 crc kubenswrapper[4824]: I0121 11:20:25.479363 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-687f57d79b-r6qj7" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.194441 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-htkvk"] Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.194929 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" podUID="26b1108a-4b98-4e39-a3cd-e0d055089fd8" containerName="ovn-controller" containerID="cri-o://5e94ef8a8f47dc75ecfbab2e3fb8e7097c1b73eed32d7ca947f11ea198cea880" gracePeriod=30 Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.194999 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" podUID="26b1108a-4b98-4e39-a3cd-e0d055089fd8" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://cee9cc9eec3fc815f1621a22a37805691288dbb60dd3ff64194a84be62abcddc" gracePeriod=30 Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.195035 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" podUID="26b1108a-4b98-4e39-a3cd-e0d055089fd8" containerName="sbdb" containerID="cri-o://948e0497eaffa7b0d84b4085c9ed8d987eb2c6527c9a82c10e576ed9c9c1e21b" gracePeriod=30 Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.195110 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" podUID="26b1108a-4b98-4e39-a3cd-e0d055089fd8" containerName="kube-rbac-proxy-node" containerID="cri-o://fc319bbc06568cc811ded2b418b04b986a31f321fed492e8be50104af79bb30b" gracePeriod=30 Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.194986 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" podUID="26b1108a-4b98-4e39-a3cd-e0d055089fd8" containerName="nbdb" containerID="cri-o://fc4c96e33e069f7d08d33caf92cd0987668506f931fd134b246d5ce5e934f95c" gracePeriod=30 Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.195089 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" podUID="26b1108a-4b98-4e39-a3cd-e0d055089fd8" containerName="ovn-acl-logging" containerID="cri-o://64a9065254f889b1c43d63759283335bf78fc4fa1531a4baae472c06c075dae3" gracePeriod=30 Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.195068 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" podUID="26b1108a-4b98-4e39-a3cd-e0d055089fd8" containerName="northd" containerID="cri-o://ddbb0e5ae02e9d5d82a80ead7efc91f1b8e557a9fb6f39940e5530fae8de2d38" gracePeriod=30 Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.219889 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" podUID="26b1108a-4b98-4e39-a3cd-e0d055089fd8" containerName="ovnkube-controller" containerID="cri-o://99d1fdcb6aaacd8685fdad0e8e6e667a311f27767270fdc0f22419bc3b8af38a" gracePeriod=30 Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.455485 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-htkvk_26b1108a-4b98-4e39-a3cd-e0d055089fd8/ovnkube-controller/3.log" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.457717 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-htkvk_26b1108a-4b98-4e39-a3cd-e0d055089fd8/ovn-acl-logging/0.log" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.458072 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-htkvk_26b1108a-4b98-4e39-a3cd-e0d055089fd8/ovn-controller/0.log" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.458361 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.504329 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-r8p75"] Jan 21 11:20:26 crc kubenswrapper[4824]: E0121 11:20:26.504634 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26b1108a-4b98-4e39-a3cd-e0d055089fd8" containerName="ovnkube-controller" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.504647 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="26b1108a-4b98-4e39-a3cd-e0d055089fd8" containerName="ovnkube-controller" Jan 21 11:20:26 crc kubenswrapper[4824]: E0121 11:20:26.504657 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26b1108a-4b98-4e39-a3cd-e0d055089fd8" containerName="ovnkube-controller" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.504663 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="26b1108a-4b98-4e39-a3cd-e0d055089fd8" containerName="ovnkube-controller" Jan 21 11:20:26 crc kubenswrapper[4824]: E0121 11:20:26.504675 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26b1108a-4b98-4e39-a3cd-e0d055089fd8" containerName="ovnkube-controller" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.504680 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="26b1108a-4b98-4e39-a3cd-e0d055089fd8" containerName="ovnkube-controller" Jan 21 11:20:26 crc kubenswrapper[4824]: E0121 11:20:26.504695 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26b1108a-4b98-4e39-a3cd-e0d055089fd8" containerName="nbdb" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.504702 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="26b1108a-4b98-4e39-a3cd-e0d055089fd8" containerName="nbdb" Jan 21 11:20:26 crc kubenswrapper[4824]: E0121 11:20:26.504726 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26b1108a-4b98-4e39-a3cd-e0d055089fd8" containerName="kubecfg-setup" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.504732 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="26b1108a-4b98-4e39-a3cd-e0d055089fd8" containerName="kubecfg-setup" Jan 21 11:20:26 crc kubenswrapper[4824]: E0121 11:20:26.504743 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26b1108a-4b98-4e39-a3cd-e0d055089fd8" containerName="ovn-acl-logging" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.504748 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="26b1108a-4b98-4e39-a3cd-e0d055089fd8" containerName="ovn-acl-logging" Jan 21 11:20:26 crc kubenswrapper[4824]: E0121 11:20:26.504755 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26b1108a-4b98-4e39-a3cd-e0d055089fd8" containerName="kube-rbac-proxy-node" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.504767 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="26b1108a-4b98-4e39-a3cd-e0d055089fd8" containerName="kube-rbac-proxy-node" Jan 21 11:20:26 crc kubenswrapper[4824]: E0121 11:20:26.504773 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26b1108a-4b98-4e39-a3cd-e0d055089fd8" containerName="sbdb" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.504779 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="26b1108a-4b98-4e39-a3cd-e0d055089fd8" containerName="sbdb" Jan 21 11:20:26 crc kubenswrapper[4824]: E0121 11:20:26.504790 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26b1108a-4b98-4e39-a3cd-e0d055089fd8" containerName="ovn-controller" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.504796 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="26b1108a-4b98-4e39-a3cd-e0d055089fd8" containerName="ovn-controller" Jan 21 11:20:26 crc kubenswrapper[4824]: E0121 11:20:26.504806 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26b1108a-4b98-4e39-a3cd-e0d055089fd8" containerName="kube-rbac-proxy-ovn-metrics" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.504812 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="26b1108a-4b98-4e39-a3cd-e0d055089fd8" containerName="kube-rbac-proxy-ovn-metrics" Jan 21 11:20:26 crc kubenswrapper[4824]: E0121 11:20:26.504818 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26b1108a-4b98-4e39-a3cd-e0d055089fd8" containerName="northd" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.504823 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="26b1108a-4b98-4e39-a3cd-e0d055089fd8" containerName="northd" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.505644 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="26b1108a-4b98-4e39-a3cd-e0d055089fd8" containerName="ovnkube-controller" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.505683 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="26b1108a-4b98-4e39-a3cd-e0d055089fd8" containerName="ovn-controller" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.505699 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="26b1108a-4b98-4e39-a3cd-e0d055089fd8" containerName="ovnkube-controller" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.505711 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="26b1108a-4b98-4e39-a3cd-e0d055089fd8" containerName="kube-rbac-proxy-ovn-metrics" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.506411 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="26b1108a-4b98-4e39-a3cd-e0d055089fd8" containerName="ovn-acl-logging" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.506438 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="26b1108a-4b98-4e39-a3cd-e0d055089fd8" containerName="northd" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.506457 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="26b1108a-4b98-4e39-a3cd-e0d055089fd8" containerName="sbdb" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.506465 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="26b1108a-4b98-4e39-a3cd-e0d055089fd8" containerName="kube-rbac-proxy-node" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.506473 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="26b1108a-4b98-4e39-a3cd-e0d055089fd8" containerName="ovnkube-controller" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.506486 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="26b1108a-4b98-4e39-a3cd-e0d055089fd8" containerName="nbdb" Jan 21 11:20:26 crc kubenswrapper[4824]: E0121 11:20:26.506892 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26b1108a-4b98-4e39-a3cd-e0d055089fd8" containerName="ovnkube-controller" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.506911 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="26b1108a-4b98-4e39-a3cd-e0d055089fd8" containerName="ovnkube-controller" Jan 21 11:20:26 crc kubenswrapper[4824]: E0121 11:20:26.506920 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26b1108a-4b98-4e39-a3cd-e0d055089fd8" containerName="ovnkube-controller" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.506925 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="26b1108a-4b98-4e39-a3cd-e0d055089fd8" containerName="ovnkube-controller" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.507045 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="26b1108a-4b98-4e39-a3cd-e0d055089fd8" containerName="ovnkube-controller" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.507236 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="26b1108a-4b98-4e39-a3cd-e0d055089fd8" containerName="ovnkube-controller" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.508675 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-r8p75" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.556821 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-host-kubelet\") pod \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\" (UID: \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\") " Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.556856 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-run-systemd\") pod \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\" (UID: \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\") " Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.556872 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-var-lib-openvswitch\") pod \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\" (UID: \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\") " Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.556892 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-log-socket\") pod \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\" (UID: \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\") " Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.556887 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "26b1108a-4b98-4e39-a3cd-e0d055089fd8" (UID: "26b1108a-4b98-4e39-a3cd-e0d055089fd8"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.556909 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-host-slash\") pod \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\" (UID: \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\") " Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.556943 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-host-var-lib-cni-networks-ovn-kubernetes\") pod \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\" (UID: \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\") " Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.556948 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-host-slash" (OuterVolumeSpecName: "host-slash") pod "26b1108a-4b98-4e39-a3cd-e0d055089fd8" (UID: "26b1108a-4b98-4e39-a3cd-e0d055089fd8"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.556996 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-log-socket" (OuterVolumeSpecName: "log-socket") pod "26b1108a-4b98-4e39-a3cd-e0d055089fd8" (UID: "26b1108a-4b98-4e39-a3cd-e0d055089fd8"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.556993 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "26b1108a-4b98-4e39-a3cd-e0d055089fd8" (UID: "26b1108a-4b98-4e39-a3cd-e0d055089fd8"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.557007 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/26b1108a-4b98-4e39-a3cd-e0d055089fd8-env-overrides\") pod \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\" (UID: \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\") " Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.557068 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-host-run-ovn-kubernetes\") pod \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\" (UID: \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\") " Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.557092 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/26b1108a-4b98-4e39-a3cd-e0d055089fd8-ovnkube-config\") pod \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\" (UID: \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\") " Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.557111 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-etc-openvswitch\") pod \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\" (UID: \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\") " Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.557129 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-run-ovn\") pod \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\" (UID: \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\") " Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.557142 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-host-cni-bin\") pod \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\" (UID: \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\") " Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.557136 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "26b1108a-4b98-4e39-a3cd-e0d055089fd8" (UID: "26b1108a-4b98-4e39-a3cd-e0d055089fd8"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.557162 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "26b1108a-4b98-4e39-a3cd-e0d055089fd8" (UID: "26b1108a-4b98-4e39-a3cd-e0d055089fd8"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.557169 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "26b1108a-4b98-4e39-a3cd-e0d055089fd8" (UID: "26b1108a-4b98-4e39-a3cd-e0d055089fd8"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.557212 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "26b1108a-4b98-4e39-a3cd-e0d055089fd8" (UID: "26b1108a-4b98-4e39-a3cd-e0d055089fd8"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.557184 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "26b1108a-4b98-4e39-a3cd-e0d055089fd8" (UID: "26b1108a-4b98-4e39-a3cd-e0d055089fd8"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.557193 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-host-run-netns\") pod \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\" (UID: \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\") " Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.557276 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-run-openvswitch\") pod \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\" (UID: \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\") " Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.557210 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "26b1108a-4b98-4e39-a3cd-e0d055089fd8" (UID: "26b1108a-4b98-4e39-a3cd-e0d055089fd8"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.557302 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-node-log\") pod \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\" (UID: \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\") " Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.557316 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "26b1108a-4b98-4e39-a3cd-e0d055089fd8" (UID: "26b1108a-4b98-4e39-a3cd-e0d055089fd8"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.557327 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/26b1108a-4b98-4e39-a3cd-e0d055089fd8-ovn-node-metrics-cert\") pod \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\" (UID: \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\") " Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.557341 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-node-log" (OuterVolumeSpecName: "node-log") pod "26b1108a-4b98-4e39-a3cd-e0d055089fd8" (UID: "26b1108a-4b98-4e39-a3cd-e0d055089fd8"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.557346 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zcf8q\" (UniqueName: \"kubernetes.io/projected/26b1108a-4b98-4e39-a3cd-e0d055089fd8-kube-api-access-zcf8q\") pod \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\" (UID: \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\") " Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.557377 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-systemd-units\") pod \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\" (UID: \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\") " Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.557407 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/26b1108a-4b98-4e39-a3cd-e0d055089fd8-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "26b1108a-4b98-4e39-a3cd-e0d055089fd8" (UID: "26b1108a-4b98-4e39-a3cd-e0d055089fd8"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.557411 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/26b1108a-4b98-4e39-a3cd-e0d055089fd8-ovnkube-script-lib\") pod \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\" (UID: \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\") " Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.557452 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-host-cni-netd\") pod \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\" (UID: \"26b1108a-4b98-4e39-a3cd-e0d055089fd8\") " Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.557484 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "26b1108a-4b98-4e39-a3cd-e0d055089fd8" (UID: "26b1108a-4b98-4e39-a3cd-e0d055089fd8"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.557526 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/26b1108a-4b98-4e39-a3cd-e0d055089fd8-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "26b1108a-4b98-4e39-a3cd-e0d055089fd8" (UID: "26b1108a-4b98-4e39-a3cd-e0d055089fd8"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.557570 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "26b1108a-4b98-4e39-a3cd-e0d055089fd8" (UID: "26b1108a-4b98-4e39-a3cd-e0d055089fd8"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.557813 4824 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-host-cni-netd\") on node \"crc\" DevicePath \"\"" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.557835 4824 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-host-kubelet\") on node \"crc\" DevicePath \"\"" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.557846 4824 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.557858 4824 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-log-socket\") on node \"crc\" DevicePath \"\"" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.557867 4824 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-host-slash\") on node \"crc\" DevicePath \"\"" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.557867 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/26b1108a-4b98-4e39-a3cd-e0d055089fd8-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "26b1108a-4b98-4e39-a3cd-e0d055089fd8" (UID: "26b1108a-4b98-4e39-a3cd-e0d055089fd8"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.557877 4824 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.557888 4824 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/26b1108a-4b98-4e39-a3cd-e0d055089fd8-env-overrides\") on node \"crc\" DevicePath \"\"" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.557899 4824 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.557907 4824 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/26b1108a-4b98-4e39-a3cd-e0d055089fd8-ovnkube-config\") on node \"crc\" DevicePath \"\"" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.557916 4824 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.557925 4824 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-run-ovn\") on node \"crc\" DevicePath \"\"" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.557933 4824 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-host-cni-bin\") on node \"crc\" DevicePath \"\"" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.557941 4824 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-host-run-netns\") on node \"crc\" DevicePath \"\"" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.557950 4824 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-run-openvswitch\") on node \"crc\" DevicePath \"\"" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.557974 4824 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-node-log\") on node \"crc\" DevicePath \"\"" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.557983 4824 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-systemd-units\") on node \"crc\" DevicePath \"\"" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.561303 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/26b1108a-4b98-4e39-a3cd-e0d055089fd8-kube-api-access-zcf8q" (OuterVolumeSpecName: "kube-api-access-zcf8q") pod "26b1108a-4b98-4e39-a3cd-e0d055089fd8" (UID: "26b1108a-4b98-4e39-a3cd-e0d055089fd8"). InnerVolumeSpecName "kube-api-access-zcf8q". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.561313 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/26b1108a-4b98-4e39-a3cd-e0d055089fd8-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "26b1108a-4b98-4e39-a3cd-e0d055089fd8" (UID: "26b1108a-4b98-4e39-a3cd-e0d055089fd8"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.567339 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "26b1108a-4b98-4e39-a3cd-e0d055089fd8" (UID: "26b1108a-4b98-4e39-a3cd-e0d055089fd8"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.659336 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/177dd5b5-7311-483f-8c21-0dae78cd3d09-env-overrides\") pod \"ovnkube-node-r8p75\" (UID: \"177dd5b5-7311-483f-8c21-0dae78cd3d09\") " pod="openshift-ovn-kubernetes/ovnkube-node-r8p75" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.659396 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/177dd5b5-7311-483f-8c21-0dae78cd3d09-run-systemd\") pod \"ovnkube-node-r8p75\" (UID: \"177dd5b5-7311-483f-8c21-0dae78cd3d09\") " pod="openshift-ovn-kubernetes/ovnkube-node-r8p75" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.659440 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/177dd5b5-7311-483f-8c21-0dae78cd3d09-systemd-units\") pod \"ovnkube-node-r8p75\" (UID: \"177dd5b5-7311-483f-8c21-0dae78cd3d09\") " pod="openshift-ovn-kubernetes/ovnkube-node-r8p75" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.659468 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/177dd5b5-7311-483f-8c21-0dae78cd3d09-node-log\") pod \"ovnkube-node-r8p75\" (UID: \"177dd5b5-7311-483f-8c21-0dae78cd3d09\") " pod="openshift-ovn-kubernetes/ovnkube-node-r8p75" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.659491 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/177dd5b5-7311-483f-8c21-0dae78cd3d09-var-lib-openvswitch\") pod \"ovnkube-node-r8p75\" (UID: \"177dd5b5-7311-483f-8c21-0dae78cd3d09\") " pod="openshift-ovn-kubernetes/ovnkube-node-r8p75" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.659529 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/177dd5b5-7311-483f-8c21-0dae78cd3d09-log-socket\") pod \"ovnkube-node-r8p75\" (UID: \"177dd5b5-7311-483f-8c21-0dae78cd3d09\") " pod="openshift-ovn-kubernetes/ovnkube-node-r8p75" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.659544 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/177dd5b5-7311-483f-8c21-0dae78cd3d09-host-cni-netd\") pod \"ovnkube-node-r8p75\" (UID: \"177dd5b5-7311-483f-8c21-0dae78cd3d09\") " pod="openshift-ovn-kubernetes/ovnkube-node-r8p75" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.659560 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/177dd5b5-7311-483f-8c21-0dae78cd3d09-run-ovn\") pod \"ovnkube-node-r8p75\" (UID: \"177dd5b5-7311-483f-8c21-0dae78cd3d09\") " pod="openshift-ovn-kubernetes/ovnkube-node-r8p75" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.659575 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/177dd5b5-7311-483f-8c21-0dae78cd3d09-ovnkube-script-lib\") pod \"ovnkube-node-r8p75\" (UID: \"177dd5b5-7311-483f-8c21-0dae78cd3d09\") " pod="openshift-ovn-kubernetes/ovnkube-node-r8p75" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.659626 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cf7mn\" (UniqueName: \"kubernetes.io/projected/177dd5b5-7311-483f-8c21-0dae78cd3d09-kube-api-access-cf7mn\") pod \"ovnkube-node-r8p75\" (UID: \"177dd5b5-7311-483f-8c21-0dae78cd3d09\") " pod="openshift-ovn-kubernetes/ovnkube-node-r8p75" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.659642 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/177dd5b5-7311-483f-8c21-0dae78cd3d09-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-r8p75\" (UID: \"177dd5b5-7311-483f-8c21-0dae78cd3d09\") " pod="openshift-ovn-kubernetes/ovnkube-node-r8p75" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.659658 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/177dd5b5-7311-483f-8c21-0dae78cd3d09-host-kubelet\") pod \"ovnkube-node-r8p75\" (UID: \"177dd5b5-7311-483f-8c21-0dae78cd3d09\") " pod="openshift-ovn-kubernetes/ovnkube-node-r8p75" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.659671 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/177dd5b5-7311-483f-8c21-0dae78cd3d09-host-slash\") pod \"ovnkube-node-r8p75\" (UID: \"177dd5b5-7311-483f-8c21-0dae78cd3d09\") " pod="openshift-ovn-kubernetes/ovnkube-node-r8p75" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.659712 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/177dd5b5-7311-483f-8c21-0dae78cd3d09-ovn-node-metrics-cert\") pod \"ovnkube-node-r8p75\" (UID: \"177dd5b5-7311-483f-8c21-0dae78cd3d09\") " pod="openshift-ovn-kubernetes/ovnkube-node-r8p75" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.659735 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/177dd5b5-7311-483f-8c21-0dae78cd3d09-etc-openvswitch\") pod \"ovnkube-node-r8p75\" (UID: \"177dd5b5-7311-483f-8c21-0dae78cd3d09\") " pod="openshift-ovn-kubernetes/ovnkube-node-r8p75" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.659751 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/177dd5b5-7311-483f-8c21-0dae78cd3d09-host-run-netns\") pod \"ovnkube-node-r8p75\" (UID: \"177dd5b5-7311-483f-8c21-0dae78cd3d09\") " pod="openshift-ovn-kubernetes/ovnkube-node-r8p75" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.659764 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/177dd5b5-7311-483f-8c21-0dae78cd3d09-ovnkube-config\") pod \"ovnkube-node-r8p75\" (UID: \"177dd5b5-7311-483f-8c21-0dae78cd3d09\") " pod="openshift-ovn-kubernetes/ovnkube-node-r8p75" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.659782 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/177dd5b5-7311-483f-8c21-0dae78cd3d09-host-run-ovn-kubernetes\") pod \"ovnkube-node-r8p75\" (UID: \"177dd5b5-7311-483f-8c21-0dae78cd3d09\") " pod="openshift-ovn-kubernetes/ovnkube-node-r8p75" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.659866 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/177dd5b5-7311-483f-8c21-0dae78cd3d09-run-openvswitch\") pod \"ovnkube-node-r8p75\" (UID: \"177dd5b5-7311-483f-8c21-0dae78cd3d09\") " pod="openshift-ovn-kubernetes/ovnkube-node-r8p75" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.659939 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/177dd5b5-7311-483f-8c21-0dae78cd3d09-host-cni-bin\") pod \"ovnkube-node-r8p75\" (UID: \"177dd5b5-7311-483f-8c21-0dae78cd3d09\") " pod="openshift-ovn-kubernetes/ovnkube-node-r8p75" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.660047 4824 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/26b1108a-4b98-4e39-a3cd-e0d055089fd8-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.660063 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zcf8q\" (UniqueName: \"kubernetes.io/projected/26b1108a-4b98-4e39-a3cd-e0d055089fd8-kube-api-access-zcf8q\") on node \"crc\" DevicePath \"\"" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.660074 4824 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/26b1108a-4b98-4e39-a3cd-e0d055089fd8-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.660084 4824 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/26b1108a-4b98-4e39-a3cd-e0d055089fd8-run-systemd\") on node \"crc\" DevicePath \"\"" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.761048 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/177dd5b5-7311-483f-8c21-0dae78cd3d09-run-openvswitch\") pod \"ovnkube-node-r8p75\" (UID: \"177dd5b5-7311-483f-8c21-0dae78cd3d09\") " pod="openshift-ovn-kubernetes/ovnkube-node-r8p75" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.761096 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/177dd5b5-7311-483f-8c21-0dae78cd3d09-host-cni-bin\") pod \"ovnkube-node-r8p75\" (UID: \"177dd5b5-7311-483f-8c21-0dae78cd3d09\") " pod="openshift-ovn-kubernetes/ovnkube-node-r8p75" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.761126 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/177dd5b5-7311-483f-8c21-0dae78cd3d09-env-overrides\") pod \"ovnkube-node-r8p75\" (UID: \"177dd5b5-7311-483f-8c21-0dae78cd3d09\") " pod="openshift-ovn-kubernetes/ovnkube-node-r8p75" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.761150 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/177dd5b5-7311-483f-8c21-0dae78cd3d09-run-systemd\") pod \"ovnkube-node-r8p75\" (UID: \"177dd5b5-7311-483f-8c21-0dae78cd3d09\") " pod="openshift-ovn-kubernetes/ovnkube-node-r8p75" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.761162 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/177dd5b5-7311-483f-8c21-0dae78cd3d09-run-openvswitch\") pod \"ovnkube-node-r8p75\" (UID: \"177dd5b5-7311-483f-8c21-0dae78cd3d09\") " pod="openshift-ovn-kubernetes/ovnkube-node-r8p75" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.761176 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/177dd5b5-7311-483f-8c21-0dae78cd3d09-systemd-units\") pod \"ovnkube-node-r8p75\" (UID: \"177dd5b5-7311-483f-8c21-0dae78cd3d09\") " pod="openshift-ovn-kubernetes/ovnkube-node-r8p75" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.761213 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/177dd5b5-7311-483f-8c21-0dae78cd3d09-node-log\") pod \"ovnkube-node-r8p75\" (UID: \"177dd5b5-7311-483f-8c21-0dae78cd3d09\") " pod="openshift-ovn-kubernetes/ovnkube-node-r8p75" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.761167 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/177dd5b5-7311-483f-8c21-0dae78cd3d09-host-cni-bin\") pod \"ovnkube-node-r8p75\" (UID: \"177dd5b5-7311-483f-8c21-0dae78cd3d09\") " pod="openshift-ovn-kubernetes/ovnkube-node-r8p75" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.761240 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/177dd5b5-7311-483f-8c21-0dae78cd3d09-var-lib-openvswitch\") pod \"ovnkube-node-r8p75\" (UID: \"177dd5b5-7311-483f-8c21-0dae78cd3d09\") " pod="openshift-ovn-kubernetes/ovnkube-node-r8p75" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.761263 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/177dd5b5-7311-483f-8c21-0dae78cd3d09-var-lib-openvswitch\") pod \"ovnkube-node-r8p75\" (UID: \"177dd5b5-7311-483f-8c21-0dae78cd3d09\") " pod="openshift-ovn-kubernetes/ovnkube-node-r8p75" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.761216 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/177dd5b5-7311-483f-8c21-0dae78cd3d09-systemd-units\") pod \"ovnkube-node-r8p75\" (UID: \"177dd5b5-7311-483f-8c21-0dae78cd3d09\") " pod="openshift-ovn-kubernetes/ovnkube-node-r8p75" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.761263 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/177dd5b5-7311-483f-8c21-0dae78cd3d09-log-socket\") pod \"ovnkube-node-r8p75\" (UID: \"177dd5b5-7311-483f-8c21-0dae78cd3d09\") " pod="openshift-ovn-kubernetes/ovnkube-node-r8p75" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.761284 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/177dd5b5-7311-483f-8c21-0dae78cd3d09-log-socket\") pod \"ovnkube-node-r8p75\" (UID: \"177dd5b5-7311-483f-8c21-0dae78cd3d09\") " pod="openshift-ovn-kubernetes/ovnkube-node-r8p75" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.761217 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/177dd5b5-7311-483f-8c21-0dae78cd3d09-run-systemd\") pod \"ovnkube-node-r8p75\" (UID: \"177dd5b5-7311-483f-8c21-0dae78cd3d09\") " pod="openshift-ovn-kubernetes/ovnkube-node-r8p75" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.761296 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/177dd5b5-7311-483f-8c21-0dae78cd3d09-host-cni-netd\") pod \"ovnkube-node-r8p75\" (UID: \"177dd5b5-7311-483f-8c21-0dae78cd3d09\") " pod="openshift-ovn-kubernetes/ovnkube-node-r8p75" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.761314 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/177dd5b5-7311-483f-8c21-0dae78cd3d09-host-cni-netd\") pod \"ovnkube-node-r8p75\" (UID: \"177dd5b5-7311-483f-8c21-0dae78cd3d09\") " pod="openshift-ovn-kubernetes/ovnkube-node-r8p75" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.761322 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/177dd5b5-7311-483f-8c21-0dae78cd3d09-run-ovn\") pod \"ovnkube-node-r8p75\" (UID: \"177dd5b5-7311-483f-8c21-0dae78cd3d09\") " pod="openshift-ovn-kubernetes/ovnkube-node-r8p75" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.761338 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/177dd5b5-7311-483f-8c21-0dae78cd3d09-ovnkube-script-lib\") pod \"ovnkube-node-r8p75\" (UID: \"177dd5b5-7311-483f-8c21-0dae78cd3d09\") " pod="openshift-ovn-kubernetes/ovnkube-node-r8p75" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.761358 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/177dd5b5-7311-483f-8c21-0dae78cd3d09-run-ovn\") pod \"ovnkube-node-r8p75\" (UID: \"177dd5b5-7311-483f-8c21-0dae78cd3d09\") " pod="openshift-ovn-kubernetes/ovnkube-node-r8p75" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.761387 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cf7mn\" (UniqueName: \"kubernetes.io/projected/177dd5b5-7311-483f-8c21-0dae78cd3d09-kube-api-access-cf7mn\") pod \"ovnkube-node-r8p75\" (UID: \"177dd5b5-7311-483f-8c21-0dae78cd3d09\") " pod="openshift-ovn-kubernetes/ovnkube-node-r8p75" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.761406 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/177dd5b5-7311-483f-8c21-0dae78cd3d09-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-r8p75\" (UID: \"177dd5b5-7311-483f-8c21-0dae78cd3d09\") " pod="openshift-ovn-kubernetes/ovnkube-node-r8p75" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.761427 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/177dd5b5-7311-483f-8c21-0dae78cd3d09-host-kubelet\") pod \"ovnkube-node-r8p75\" (UID: \"177dd5b5-7311-483f-8c21-0dae78cd3d09\") " pod="openshift-ovn-kubernetes/ovnkube-node-r8p75" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.761444 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/177dd5b5-7311-483f-8c21-0dae78cd3d09-host-slash\") pod \"ovnkube-node-r8p75\" (UID: \"177dd5b5-7311-483f-8c21-0dae78cd3d09\") " pod="openshift-ovn-kubernetes/ovnkube-node-r8p75" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.761470 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/177dd5b5-7311-483f-8c21-0dae78cd3d09-ovn-node-metrics-cert\") pod \"ovnkube-node-r8p75\" (UID: \"177dd5b5-7311-483f-8c21-0dae78cd3d09\") " pod="openshift-ovn-kubernetes/ovnkube-node-r8p75" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.761489 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/177dd5b5-7311-483f-8c21-0dae78cd3d09-etc-openvswitch\") pod \"ovnkube-node-r8p75\" (UID: \"177dd5b5-7311-483f-8c21-0dae78cd3d09\") " pod="openshift-ovn-kubernetes/ovnkube-node-r8p75" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.761522 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/177dd5b5-7311-483f-8c21-0dae78cd3d09-host-run-netns\") pod \"ovnkube-node-r8p75\" (UID: \"177dd5b5-7311-483f-8c21-0dae78cd3d09\") " pod="openshift-ovn-kubernetes/ovnkube-node-r8p75" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.761538 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/177dd5b5-7311-483f-8c21-0dae78cd3d09-ovnkube-config\") pod \"ovnkube-node-r8p75\" (UID: \"177dd5b5-7311-483f-8c21-0dae78cd3d09\") " pod="openshift-ovn-kubernetes/ovnkube-node-r8p75" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.761544 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/177dd5b5-7311-483f-8c21-0dae78cd3d09-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-r8p75\" (UID: \"177dd5b5-7311-483f-8c21-0dae78cd3d09\") " pod="openshift-ovn-kubernetes/ovnkube-node-r8p75" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.761580 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/177dd5b5-7311-483f-8c21-0dae78cd3d09-host-run-ovn-kubernetes\") pod \"ovnkube-node-r8p75\" (UID: \"177dd5b5-7311-483f-8c21-0dae78cd3d09\") " pod="openshift-ovn-kubernetes/ovnkube-node-r8p75" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.761579 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/177dd5b5-7311-483f-8c21-0dae78cd3d09-host-kubelet\") pod \"ovnkube-node-r8p75\" (UID: \"177dd5b5-7311-483f-8c21-0dae78cd3d09\") " pod="openshift-ovn-kubernetes/ovnkube-node-r8p75" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.761561 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/177dd5b5-7311-483f-8c21-0dae78cd3d09-host-run-ovn-kubernetes\") pod \"ovnkube-node-r8p75\" (UID: \"177dd5b5-7311-483f-8c21-0dae78cd3d09\") " pod="openshift-ovn-kubernetes/ovnkube-node-r8p75" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.761590 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/177dd5b5-7311-483f-8c21-0dae78cd3d09-host-run-netns\") pod \"ovnkube-node-r8p75\" (UID: \"177dd5b5-7311-483f-8c21-0dae78cd3d09\") " pod="openshift-ovn-kubernetes/ovnkube-node-r8p75" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.761608 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/177dd5b5-7311-483f-8c21-0dae78cd3d09-etc-openvswitch\") pod \"ovnkube-node-r8p75\" (UID: \"177dd5b5-7311-483f-8c21-0dae78cd3d09\") " pod="openshift-ovn-kubernetes/ovnkube-node-r8p75" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.761609 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/177dd5b5-7311-483f-8c21-0dae78cd3d09-env-overrides\") pod \"ovnkube-node-r8p75\" (UID: \"177dd5b5-7311-483f-8c21-0dae78cd3d09\") " pod="openshift-ovn-kubernetes/ovnkube-node-r8p75" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.761549 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/177dd5b5-7311-483f-8c21-0dae78cd3d09-host-slash\") pod \"ovnkube-node-r8p75\" (UID: \"177dd5b5-7311-483f-8c21-0dae78cd3d09\") " pod="openshift-ovn-kubernetes/ovnkube-node-r8p75" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.761763 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/177dd5b5-7311-483f-8c21-0dae78cd3d09-node-log\") pod \"ovnkube-node-r8p75\" (UID: \"177dd5b5-7311-483f-8c21-0dae78cd3d09\") " pod="openshift-ovn-kubernetes/ovnkube-node-r8p75" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.761941 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/177dd5b5-7311-483f-8c21-0dae78cd3d09-ovnkube-script-lib\") pod \"ovnkube-node-r8p75\" (UID: \"177dd5b5-7311-483f-8c21-0dae78cd3d09\") " pod="openshift-ovn-kubernetes/ovnkube-node-r8p75" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.762235 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/177dd5b5-7311-483f-8c21-0dae78cd3d09-ovnkube-config\") pod \"ovnkube-node-r8p75\" (UID: \"177dd5b5-7311-483f-8c21-0dae78cd3d09\") " pod="openshift-ovn-kubernetes/ovnkube-node-r8p75" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.764558 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/177dd5b5-7311-483f-8c21-0dae78cd3d09-ovn-node-metrics-cert\") pod \"ovnkube-node-r8p75\" (UID: \"177dd5b5-7311-483f-8c21-0dae78cd3d09\") " pod="openshift-ovn-kubernetes/ovnkube-node-r8p75" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.775943 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cf7mn\" (UniqueName: \"kubernetes.io/projected/177dd5b5-7311-483f-8c21-0dae78cd3d09-kube-api-access-cf7mn\") pod \"ovnkube-node-r8p75\" (UID: \"177dd5b5-7311-483f-8c21-0dae78cd3d09\") " pod="openshift-ovn-kubernetes/ovnkube-node-r8p75" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.783319 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-htkvk_26b1108a-4b98-4e39-a3cd-e0d055089fd8/ovnkube-controller/3.log" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.784983 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-htkvk_26b1108a-4b98-4e39-a3cd-e0d055089fd8/ovn-acl-logging/0.log" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.785358 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-htkvk_26b1108a-4b98-4e39-a3cd-e0d055089fd8/ovn-controller/0.log" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.785679 4824 generic.go:334] "Generic (PLEG): container finished" podID="26b1108a-4b98-4e39-a3cd-e0d055089fd8" containerID="99d1fdcb6aaacd8685fdad0e8e6e667a311f27767270fdc0f22419bc3b8af38a" exitCode=0 Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.785699 4824 generic.go:334] "Generic (PLEG): container finished" podID="26b1108a-4b98-4e39-a3cd-e0d055089fd8" containerID="948e0497eaffa7b0d84b4085c9ed8d987eb2c6527c9a82c10e576ed9c9c1e21b" exitCode=0 Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.785707 4824 generic.go:334] "Generic (PLEG): container finished" podID="26b1108a-4b98-4e39-a3cd-e0d055089fd8" containerID="fc4c96e33e069f7d08d33caf92cd0987668506f931fd134b246d5ce5e934f95c" exitCode=0 Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.785713 4824 generic.go:334] "Generic (PLEG): container finished" podID="26b1108a-4b98-4e39-a3cd-e0d055089fd8" containerID="ddbb0e5ae02e9d5d82a80ead7efc91f1b8e557a9fb6f39940e5530fae8de2d38" exitCode=0 Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.785721 4824 generic.go:334] "Generic (PLEG): container finished" podID="26b1108a-4b98-4e39-a3cd-e0d055089fd8" containerID="cee9cc9eec3fc815f1621a22a37805691288dbb60dd3ff64194a84be62abcddc" exitCode=0 Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.785727 4824 generic.go:334] "Generic (PLEG): container finished" podID="26b1108a-4b98-4e39-a3cd-e0d055089fd8" containerID="fc319bbc06568cc811ded2b418b04b986a31f321fed492e8be50104af79bb30b" exitCode=0 Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.785732 4824 generic.go:334] "Generic (PLEG): container finished" podID="26b1108a-4b98-4e39-a3cd-e0d055089fd8" containerID="64a9065254f889b1c43d63759283335bf78fc4fa1531a4baae472c06c075dae3" exitCode=143 Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.785737 4824 generic.go:334] "Generic (PLEG): container finished" podID="26b1108a-4b98-4e39-a3cd-e0d055089fd8" containerID="5e94ef8a8f47dc75ecfbab2e3fb8e7097c1b73eed32d7ca947f11ea198cea880" exitCode=143 Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.785746 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.785751 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" event={"ID":"26b1108a-4b98-4e39-a3cd-e0d055089fd8","Type":"ContainerDied","Data":"99d1fdcb6aaacd8685fdad0e8e6e667a311f27767270fdc0f22419bc3b8af38a"} Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.785783 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" event={"ID":"26b1108a-4b98-4e39-a3cd-e0d055089fd8","Type":"ContainerDied","Data":"948e0497eaffa7b0d84b4085c9ed8d987eb2c6527c9a82c10e576ed9c9c1e21b"} Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.785794 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" event={"ID":"26b1108a-4b98-4e39-a3cd-e0d055089fd8","Type":"ContainerDied","Data":"fc4c96e33e069f7d08d33caf92cd0987668506f931fd134b246d5ce5e934f95c"} Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.785804 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" event={"ID":"26b1108a-4b98-4e39-a3cd-e0d055089fd8","Type":"ContainerDied","Data":"ddbb0e5ae02e9d5d82a80ead7efc91f1b8e557a9fb6f39940e5530fae8de2d38"} Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.785813 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" event={"ID":"26b1108a-4b98-4e39-a3cd-e0d055089fd8","Type":"ContainerDied","Data":"cee9cc9eec3fc815f1621a22a37805691288dbb60dd3ff64194a84be62abcddc"} Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.785821 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" event={"ID":"26b1108a-4b98-4e39-a3cd-e0d055089fd8","Type":"ContainerDied","Data":"fc319bbc06568cc811ded2b418b04b986a31f321fed492e8be50104af79bb30b"} Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.785830 4824 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"86a6881f109dab1f910fe505a21c7c800e587de4dded2635147997c51c05c483"} Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.785832 4824 scope.go:117] "RemoveContainer" containerID="99d1fdcb6aaacd8685fdad0e8e6e667a311f27767270fdc0f22419bc3b8af38a" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.785840 4824 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"948e0497eaffa7b0d84b4085c9ed8d987eb2c6527c9a82c10e576ed9c9c1e21b"} Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.785933 4824 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"fc4c96e33e069f7d08d33caf92cd0987668506f931fd134b246d5ce5e934f95c"} Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.785946 4824 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ddbb0e5ae02e9d5d82a80ead7efc91f1b8e557a9fb6f39940e5530fae8de2d38"} Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.785951 4824 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"cee9cc9eec3fc815f1621a22a37805691288dbb60dd3ff64194a84be62abcddc"} Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.785971 4824 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"fc319bbc06568cc811ded2b418b04b986a31f321fed492e8be50104af79bb30b"} Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.785976 4824 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"64a9065254f889b1c43d63759283335bf78fc4fa1531a4baae472c06c075dae3"} Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.785981 4824 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"5e94ef8a8f47dc75ecfbab2e3fb8e7097c1b73eed32d7ca947f11ea198cea880"} Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.785986 4824 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3"} Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.786000 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" event={"ID":"26b1108a-4b98-4e39-a3cd-e0d055089fd8","Type":"ContainerDied","Data":"64a9065254f889b1c43d63759283335bf78fc4fa1531a4baae472c06c075dae3"} Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.786016 4824 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"99d1fdcb6aaacd8685fdad0e8e6e667a311f27767270fdc0f22419bc3b8af38a"} Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.786021 4824 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"86a6881f109dab1f910fe505a21c7c800e587de4dded2635147997c51c05c483"} Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.786027 4824 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"948e0497eaffa7b0d84b4085c9ed8d987eb2c6527c9a82c10e576ed9c9c1e21b"} Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.786031 4824 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"fc4c96e33e069f7d08d33caf92cd0987668506f931fd134b246d5ce5e934f95c"} Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.786036 4824 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ddbb0e5ae02e9d5d82a80ead7efc91f1b8e557a9fb6f39940e5530fae8de2d38"} Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.786040 4824 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"cee9cc9eec3fc815f1621a22a37805691288dbb60dd3ff64194a84be62abcddc"} Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.786045 4824 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"fc319bbc06568cc811ded2b418b04b986a31f321fed492e8be50104af79bb30b"} Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.786049 4824 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"64a9065254f889b1c43d63759283335bf78fc4fa1531a4baae472c06c075dae3"} Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.786053 4824 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"5e94ef8a8f47dc75ecfbab2e3fb8e7097c1b73eed32d7ca947f11ea198cea880"} Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.786058 4824 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3"} Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.786064 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" event={"ID":"26b1108a-4b98-4e39-a3cd-e0d055089fd8","Type":"ContainerDied","Data":"5e94ef8a8f47dc75ecfbab2e3fb8e7097c1b73eed32d7ca947f11ea198cea880"} Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.786072 4824 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"99d1fdcb6aaacd8685fdad0e8e6e667a311f27767270fdc0f22419bc3b8af38a"} Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.786079 4824 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"86a6881f109dab1f910fe505a21c7c800e587de4dded2635147997c51c05c483"} Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.786084 4824 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"948e0497eaffa7b0d84b4085c9ed8d987eb2c6527c9a82c10e576ed9c9c1e21b"} Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.786089 4824 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"fc4c96e33e069f7d08d33caf92cd0987668506f931fd134b246d5ce5e934f95c"} Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.786093 4824 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ddbb0e5ae02e9d5d82a80ead7efc91f1b8e557a9fb6f39940e5530fae8de2d38"} Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.786098 4824 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"cee9cc9eec3fc815f1621a22a37805691288dbb60dd3ff64194a84be62abcddc"} Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.786102 4824 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"fc319bbc06568cc811ded2b418b04b986a31f321fed492e8be50104af79bb30b"} Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.786107 4824 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"64a9065254f889b1c43d63759283335bf78fc4fa1531a4baae472c06c075dae3"} Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.786114 4824 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"5e94ef8a8f47dc75ecfbab2e3fb8e7097c1b73eed32d7ca947f11ea198cea880"} Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.786119 4824 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3"} Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.786125 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-htkvk" event={"ID":"26b1108a-4b98-4e39-a3cd-e0d055089fd8","Type":"ContainerDied","Data":"b233a1680204a0985feed02a396e93f06104160dd5b2e0d82128d39c672a1129"} Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.786132 4824 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"99d1fdcb6aaacd8685fdad0e8e6e667a311f27767270fdc0f22419bc3b8af38a"} Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.786137 4824 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"86a6881f109dab1f910fe505a21c7c800e587de4dded2635147997c51c05c483"} Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.786142 4824 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"948e0497eaffa7b0d84b4085c9ed8d987eb2c6527c9a82c10e576ed9c9c1e21b"} Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.786146 4824 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"fc4c96e33e069f7d08d33caf92cd0987668506f931fd134b246d5ce5e934f95c"} Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.786151 4824 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ddbb0e5ae02e9d5d82a80ead7efc91f1b8e557a9fb6f39940e5530fae8de2d38"} Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.786155 4824 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"cee9cc9eec3fc815f1621a22a37805691288dbb60dd3ff64194a84be62abcddc"} Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.786160 4824 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"fc319bbc06568cc811ded2b418b04b986a31f321fed492e8be50104af79bb30b"} Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.786164 4824 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"64a9065254f889b1c43d63759283335bf78fc4fa1531a4baae472c06c075dae3"} Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.786168 4824 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"5e94ef8a8f47dc75ecfbab2e3fb8e7097c1b73eed32d7ca947f11ea198cea880"} Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.786172 4824 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3"} Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.787143 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-x94nb_6b9588d1-5876-4c2c-8de5-e5c0ad46f04e/kube-multus/2.log" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.787488 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-x94nb_6b9588d1-5876-4c2c-8de5-e5c0ad46f04e/kube-multus/1.log" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.787530 4824 generic.go:334] "Generic (PLEG): container finished" podID="6b9588d1-5876-4c2c-8de5-e5c0ad46f04e" containerID="80258959697dd4e8883d646587a5ac979419b23c4e94eace61a8b23e0885b470" exitCode=2 Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.787549 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-x94nb" event={"ID":"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e","Type":"ContainerDied","Data":"80258959697dd4e8883d646587a5ac979419b23c4e94eace61a8b23e0885b470"} Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.787563 4824 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"49ce055ec8090d6957cb7e23fe75edb82c0c2d7f87e7203d280606368dd3179f"} Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.787930 4824 scope.go:117] "RemoveContainer" containerID="80258959697dd4e8883d646587a5ac979419b23c4e94eace61a8b23e0885b470" Jan 21 11:20:26 crc kubenswrapper[4824]: E0121 11:20:26.788169 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-x94nb_openshift-multus(6b9588d1-5876-4c2c-8de5-e5c0ad46f04e)\"" pod="openshift-multus/multus-x94nb" podUID="6b9588d1-5876-4c2c-8de5-e5c0ad46f04e" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.797952 4824 scope.go:117] "RemoveContainer" containerID="86a6881f109dab1f910fe505a21c7c800e587de4dded2635147997c51c05c483" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.813619 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-htkvk"] Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.816022 4824 scope.go:117] "RemoveContainer" containerID="948e0497eaffa7b0d84b4085c9ed8d987eb2c6527c9a82c10e576ed9c9c1e21b" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.816492 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-htkvk"] Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.819766 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-r8p75" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.824981 4824 scope.go:117] "RemoveContainer" containerID="fc4c96e33e069f7d08d33caf92cd0987668506f931fd134b246d5ce5e934f95c" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.836071 4824 scope.go:117] "RemoveContainer" containerID="ddbb0e5ae02e9d5d82a80ead7efc91f1b8e557a9fb6f39940e5530fae8de2d38" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.844881 4824 scope.go:117] "RemoveContainer" containerID="cee9cc9eec3fc815f1621a22a37805691288dbb60dd3ff64194a84be62abcddc" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.856076 4824 scope.go:117] "RemoveContainer" containerID="fc319bbc06568cc811ded2b418b04b986a31f321fed492e8be50104af79bb30b" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.866356 4824 scope.go:117] "RemoveContainer" containerID="64a9065254f889b1c43d63759283335bf78fc4fa1531a4baae472c06c075dae3" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.886791 4824 scope.go:117] "RemoveContainer" containerID="5e94ef8a8f47dc75ecfbab2e3fb8e7097c1b73eed32d7ca947f11ea198cea880" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.896911 4824 scope.go:117] "RemoveContainer" containerID="82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.906210 4824 scope.go:117] "RemoveContainer" containerID="99d1fdcb6aaacd8685fdad0e8e6e667a311f27767270fdc0f22419bc3b8af38a" Jan 21 11:20:26 crc kubenswrapper[4824]: E0121 11:20:26.906519 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"99d1fdcb6aaacd8685fdad0e8e6e667a311f27767270fdc0f22419bc3b8af38a\": container with ID starting with 99d1fdcb6aaacd8685fdad0e8e6e667a311f27767270fdc0f22419bc3b8af38a not found: ID does not exist" containerID="99d1fdcb6aaacd8685fdad0e8e6e667a311f27767270fdc0f22419bc3b8af38a" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.906552 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"99d1fdcb6aaacd8685fdad0e8e6e667a311f27767270fdc0f22419bc3b8af38a"} err="failed to get container status \"99d1fdcb6aaacd8685fdad0e8e6e667a311f27767270fdc0f22419bc3b8af38a\": rpc error: code = NotFound desc = could not find container \"99d1fdcb6aaacd8685fdad0e8e6e667a311f27767270fdc0f22419bc3b8af38a\": container with ID starting with 99d1fdcb6aaacd8685fdad0e8e6e667a311f27767270fdc0f22419bc3b8af38a not found: ID does not exist" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.906573 4824 scope.go:117] "RemoveContainer" containerID="86a6881f109dab1f910fe505a21c7c800e587de4dded2635147997c51c05c483" Jan 21 11:20:26 crc kubenswrapper[4824]: E0121 11:20:26.906779 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"86a6881f109dab1f910fe505a21c7c800e587de4dded2635147997c51c05c483\": container with ID starting with 86a6881f109dab1f910fe505a21c7c800e587de4dded2635147997c51c05c483 not found: ID does not exist" containerID="86a6881f109dab1f910fe505a21c7c800e587de4dded2635147997c51c05c483" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.906804 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"86a6881f109dab1f910fe505a21c7c800e587de4dded2635147997c51c05c483"} err="failed to get container status \"86a6881f109dab1f910fe505a21c7c800e587de4dded2635147997c51c05c483\": rpc error: code = NotFound desc = could not find container \"86a6881f109dab1f910fe505a21c7c800e587de4dded2635147997c51c05c483\": container with ID starting with 86a6881f109dab1f910fe505a21c7c800e587de4dded2635147997c51c05c483 not found: ID does not exist" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.906824 4824 scope.go:117] "RemoveContainer" containerID="948e0497eaffa7b0d84b4085c9ed8d987eb2c6527c9a82c10e576ed9c9c1e21b" Jan 21 11:20:26 crc kubenswrapper[4824]: E0121 11:20:26.907044 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"948e0497eaffa7b0d84b4085c9ed8d987eb2c6527c9a82c10e576ed9c9c1e21b\": container with ID starting with 948e0497eaffa7b0d84b4085c9ed8d987eb2c6527c9a82c10e576ed9c9c1e21b not found: ID does not exist" containerID="948e0497eaffa7b0d84b4085c9ed8d987eb2c6527c9a82c10e576ed9c9c1e21b" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.907089 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"948e0497eaffa7b0d84b4085c9ed8d987eb2c6527c9a82c10e576ed9c9c1e21b"} err="failed to get container status \"948e0497eaffa7b0d84b4085c9ed8d987eb2c6527c9a82c10e576ed9c9c1e21b\": rpc error: code = NotFound desc = could not find container \"948e0497eaffa7b0d84b4085c9ed8d987eb2c6527c9a82c10e576ed9c9c1e21b\": container with ID starting with 948e0497eaffa7b0d84b4085c9ed8d987eb2c6527c9a82c10e576ed9c9c1e21b not found: ID does not exist" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.907105 4824 scope.go:117] "RemoveContainer" containerID="fc4c96e33e069f7d08d33caf92cd0987668506f931fd134b246d5ce5e934f95c" Jan 21 11:20:26 crc kubenswrapper[4824]: E0121 11:20:26.907330 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fc4c96e33e069f7d08d33caf92cd0987668506f931fd134b246d5ce5e934f95c\": container with ID starting with fc4c96e33e069f7d08d33caf92cd0987668506f931fd134b246d5ce5e934f95c not found: ID does not exist" containerID="fc4c96e33e069f7d08d33caf92cd0987668506f931fd134b246d5ce5e934f95c" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.907354 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fc4c96e33e069f7d08d33caf92cd0987668506f931fd134b246d5ce5e934f95c"} err="failed to get container status \"fc4c96e33e069f7d08d33caf92cd0987668506f931fd134b246d5ce5e934f95c\": rpc error: code = NotFound desc = could not find container \"fc4c96e33e069f7d08d33caf92cd0987668506f931fd134b246d5ce5e934f95c\": container with ID starting with fc4c96e33e069f7d08d33caf92cd0987668506f931fd134b246d5ce5e934f95c not found: ID does not exist" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.907419 4824 scope.go:117] "RemoveContainer" containerID="ddbb0e5ae02e9d5d82a80ead7efc91f1b8e557a9fb6f39940e5530fae8de2d38" Jan 21 11:20:26 crc kubenswrapper[4824]: E0121 11:20:26.907975 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ddbb0e5ae02e9d5d82a80ead7efc91f1b8e557a9fb6f39940e5530fae8de2d38\": container with ID starting with ddbb0e5ae02e9d5d82a80ead7efc91f1b8e557a9fb6f39940e5530fae8de2d38 not found: ID does not exist" containerID="ddbb0e5ae02e9d5d82a80ead7efc91f1b8e557a9fb6f39940e5530fae8de2d38" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.908001 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ddbb0e5ae02e9d5d82a80ead7efc91f1b8e557a9fb6f39940e5530fae8de2d38"} err="failed to get container status \"ddbb0e5ae02e9d5d82a80ead7efc91f1b8e557a9fb6f39940e5530fae8de2d38\": rpc error: code = NotFound desc = could not find container \"ddbb0e5ae02e9d5d82a80ead7efc91f1b8e557a9fb6f39940e5530fae8de2d38\": container with ID starting with ddbb0e5ae02e9d5d82a80ead7efc91f1b8e557a9fb6f39940e5530fae8de2d38 not found: ID does not exist" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.908016 4824 scope.go:117] "RemoveContainer" containerID="cee9cc9eec3fc815f1621a22a37805691288dbb60dd3ff64194a84be62abcddc" Jan 21 11:20:26 crc kubenswrapper[4824]: E0121 11:20:26.908212 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cee9cc9eec3fc815f1621a22a37805691288dbb60dd3ff64194a84be62abcddc\": container with ID starting with cee9cc9eec3fc815f1621a22a37805691288dbb60dd3ff64194a84be62abcddc not found: ID does not exist" containerID="cee9cc9eec3fc815f1621a22a37805691288dbb60dd3ff64194a84be62abcddc" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.908258 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cee9cc9eec3fc815f1621a22a37805691288dbb60dd3ff64194a84be62abcddc"} err="failed to get container status \"cee9cc9eec3fc815f1621a22a37805691288dbb60dd3ff64194a84be62abcddc\": rpc error: code = NotFound desc = could not find container \"cee9cc9eec3fc815f1621a22a37805691288dbb60dd3ff64194a84be62abcddc\": container with ID starting with cee9cc9eec3fc815f1621a22a37805691288dbb60dd3ff64194a84be62abcddc not found: ID does not exist" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.908272 4824 scope.go:117] "RemoveContainer" containerID="fc319bbc06568cc811ded2b418b04b986a31f321fed492e8be50104af79bb30b" Jan 21 11:20:26 crc kubenswrapper[4824]: E0121 11:20:26.908507 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fc319bbc06568cc811ded2b418b04b986a31f321fed492e8be50104af79bb30b\": container with ID starting with fc319bbc06568cc811ded2b418b04b986a31f321fed492e8be50104af79bb30b not found: ID does not exist" containerID="fc319bbc06568cc811ded2b418b04b986a31f321fed492e8be50104af79bb30b" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.908528 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fc319bbc06568cc811ded2b418b04b986a31f321fed492e8be50104af79bb30b"} err="failed to get container status \"fc319bbc06568cc811ded2b418b04b986a31f321fed492e8be50104af79bb30b\": rpc error: code = NotFound desc = could not find container \"fc319bbc06568cc811ded2b418b04b986a31f321fed492e8be50104af79bb30b\": container with ID starting with fc319bbc06568cc811ded2b418b04b986a31f321fed492e8be50104af79bb30b not found: ID does not exist" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.908544 4824 scope.go:117] "RemoveContainer" containerID="64a9065254f889b1c43d63759283335bf78fc4fa1531a4baae472c06c075dae3" Jan 21 11:20:26 crc kubenswrapper[4824]: E0121 11:20:26.908802 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"64a9065254f889b1c43d63759283335bf78fc4fa1531a4baae472c06c075dae3\": container with ID starting with 64a9065254f889b1c43d63759283335bf78fc4fa1531a4baae472c06c075dae3 not found: ID does not exist" containerID="64a9065254f889b1c43d63759283335bf78fc4fa1531a4baae472c06c075dae3" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.908829 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"64a9065254f889b1c43d63759283335bf78fc4fa1531a4baae472c06c075dae3"} err="failed to get container status \"64a9065254f889b1c43d63759283335bf78fc4fa1531a4baae472c06c075dae3\": rpc error: code = NotFound desc = could not find container \"64a9065254f889b1c43d63759283335bf78fc4fa1531a4baae472c06c075dae3\": container with ID starting with 64a9065254f889b1c43d63759283335bf78fc4fa1531a4baae472c06c075dae3 not found: ID does not exist" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.908843 4824 scope.go:117] "RemoveContainer" containerID="5e94ef8a8f47dc75ecfbab2e3fb8e7097c1b73eed32d7ca947f11ea198cea880" Jan 21 11:20:26 crc kubenswrapper[4824]: E0121 11:20:26.909127 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5e94ef8a8f47dc75ecfbab2e3fb8e7097c1b73eed32d7ca947f11ea198cea880\": container with ID starting with 5e94ef8a8f47dc75ecfbab2e3fb8e7097c1b73eed32d7ca947f11ea198cea880 not found: ID does not exist" containerID="5e94ef8a8f47dc75ecfbab2e3fb8e7097c1b73eed32d7ca947f11ea198cea880" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.909150 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5e94ef8a8f47dc75ecfbab2e3fb8e7097c1b73eed32d7ca947f11ea198cea880"} err="failed to get container status \"5e94ef8a8f47dc75ecfbab2e3fb8e7097c1b73eed32d7ca947f11ea198cea880\": rpc error: code = NotFound desc = could not find container \"5e94ef8a8f47dc75ecfbab2e3fb8e7097c1b73eed32d7ca947f11ea198cea880\": container with ID starting with 5e94ef8a8f47dc75ecfbab2e3fb8e7097c1b73eed32d7ca947f11ea198cea880 not found: ID does not exist" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.909167 4824 scope.go:117] "RemoveContainer" containerID="82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3" Jan 21 11:20:26 crc kubenswrapper[4824]: E0121 11:20:26.909389 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3\": container with ID starting with 82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3 not found: ID does not exist" containerID="82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.909408 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3"} err="failed to get container status \"82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3\": rpc error: code = NotFound desc = could not find container \"82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3\": container with ID starting with 82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3 not found: ID does not exist" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.909421 4824 scope.go:117] "RemoveContainer" containerID="99d1fdcb6aaacd8685fdad0e8e6e667a311f27767270fdc0f22419bc3b8af38a" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.909620 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"99d1fdcb6aaacd8685fdad0e8e6e667a311f27767270fdc0f22419bc3b8af38a"} err="failed to get container status \"99d1fdcb6aaacd8685fdad0e8e6e667a311f27767270fdc0f22419bc3b8af38a\": rpc error: code = NotFound desc = could not find container \"99d1fdcb6aaacd8685fdad0e8e6e667a311f27767270fdc0f22419bc3b8af38a\": container with ID starting with 99d1fdcb6aaacd8685fdad0e8e6e667a311f27767270fdc0f22419bc3b8af38a not found: ID does not exist" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.909639 4824 scope.go:117] "RemoveContainer" containerID="86a6881f109dab1f910fe505a21c7c800e587de4dded2635147997c51c05c483" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.909821 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"86a6881f109dab1f910fe505a21c7c800e587de4dded2635147997c51c05c483"} err="failed to get container status \"86a6881f109dab1f910fe505a21c7c800e587de4dded2635147997c51c05c483\": rpc error: code = NotFound desc = could not find container \"86a6881f109dab1f910fe505a21c7c800e587de4dded2635147997c51c05c483\": container with ID starting with 86a6881f109dab1f910fe505a21c7c800e587de4dded2635147997c51c05c483 not found: ID does not exist" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.909839 4824 scope.go:117] "RemoveContainer" containerID="948e0497eaffa7b0d84b4085c9ed8d987eb2c6527c9a82c10e576ed9c9c1e21b" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.910648 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"948e0497eaffa7b0d84b4085c9ed8d987eb2c6527c9a82c10e576ed9c9c1e21b"} err="failed to get container status \"948e0497eaffa7b0d84b4085c9ed8d987eb2c6527c9a82c10e576ed9c9c1e21b\": rpc error: code = NotFound desc = could not find container \"948e0497eaffa7b0d84b4085c9ed8d987eb2c6527c9a82c10e576ed9c9c1e21b\": container with ID starting with 948e0497eaffa7b0d84b4085c9ed8d987eb2c6527c9a82c10e576ed9c9c1e21b not found: ID does not exist" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.910668 4824 scope.go:117] "RemoveContainer" containerID="fc4c96e33e069f7d08d33caf92cd0987668506f931fd134b246d5ce5e934f95c" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.910999 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fc4c96e33e069f7d08d33caf92cd0987668506f931fd134b246d5ce5e934f95c"} err="failed to get container status \"fc4c96e33e069f7d08d33caf92cd0987668506f931fd134b246d5ce5e934f95c\": rpc error: code = NotFound desc = could not find container \"fc4c96e33e069f7d08d33caf92cd0987668506f931fd134b246d5ce5e934f95c\": container with ID starting with fc4c96e33e069f7d08d33caf92cd0987668506f931fd134b246d5ce5e934f95c not found: ID does not exist" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.911021 4824 scope.go:117] "RemoveContainer" containerID="ddbb0e5ae02e9d5d82a80ead7efc91f1b8e557a9fb6f39940e5530fae8de2d38" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.911339 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ddbb0e5ae02e9d5d82a80ead7efc91f1b8e557a9fb6f39940e5530fae8de2d38"} err="failed to get container status \"ddbb0e5ae02e9d5d82a80ead7efc91f1b8e557a9fb6f39940e5530fae8de2d38\": rpc error: code = NotFound desc = could not find container \"ddbb0e5ae02e9d5d82a80ead7efc91f1b8e557a9fb6f39940e5530fae8de2d38\": container with ID starting with ddbb0e5ae02e9d5d82a80ead7efc91f1b8e557a9fb6f39940e5530fae8de2d38 not found: ID does not exist" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.911419 4824 scope.go:117] "RemoveContainer" containerID="cee9cc9eec3fc815f1621a22a37805691288dbb60dd3ff64194a84be62abcddc" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.911619 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cee9cc9eec3fc815f1621a22a37805691288dbb60dd3ff64194a84be62abcddc"} err="failed to get container status \"cee9cc9eec3fc815f1621a22a37805691288dbb60dd3ff64194a84be62abcddc\": rpc error: code = NotFound desc = could not find container \"cee9cc9eec3fc815f1621a22a37805691288dbb60dd3ff64194a84be62abcddc\": container with ID starting with cee9cc9eec3fc815f1621a22a37805691288dbb60dd3ff64194a84be62abcddc not found: ID does not exist" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.911649 4824 scope.go:117] "RemoveContainer" containerID="fc319bbc06568cc811ded2b418b04b986a31f321fed492e8be50104af79bb30b" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.912123 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fc319bbc06568cc811ded2b418b04b986a31f321fed492e8be50104af79bb30b"} err="failed to get container status \"fc319bbc06568cc811ded2b418b04b986a31f321fed492e8be50104af79bb30b\": rpc error: code = NotFound desc = could not find container \"fc319bbc06568cc811ded2b418b04b986a31f321fed492e8be50104af79bb30b\": container with ID starting with fc319bbc06568cc811ded2b418b04b986a31f321fed492e8be50104af79bb30b not found: ID does not exist" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.912144 4824 scope.go:117] "RemoveContainer" containerID="64a9065254f889b1c43d63759283335bf78fc4fa1531a4baae472c06c075dae3" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.912397 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"64a9065254f889b1c43d63759283335bf78fc4fa1531a4baae472c06c075dae3"} err="failed to get container status \"64a9065254f889b1c43d63759283335bf78fc4fa1531a4baae472c06c075dae3\": rpc error: code = NotFound desc = could not find container \"64a9065254f889b1c43d63759283335bf78fc4fa1531a4baae472c06c075dae3\": container with ID starting with 64a9065254f889b1c43d63759283335bf78fc4fa1531a4baae472c06c075dae3 not found: ID does not exist" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.912436 4824 scope.go:117] "RemoveContainer" containerID="5e94ef8a8f47dc75ecfbab2e3fb8e7097c1b73eed32d7ca947f11ea198cea880" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.912839 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5e94ef8a8f47dc75ecfbab2e3fb8e7097c1b73eed32d7ca947f11ea198cea880"} err="failed to get container status \"5e94ef8a8f47dc75ecfbab2e3fb8e7097c1b73eed32d7ca947f11ea198cea880\": rpc error: code = NotFound desc = could not find container \"5e94ef8a8f47dc75ecfbab2e3fb8e7097c1b73eed32d7ca947f11ea198cea880\": container with ID starting with 5e94ef8a8f47dc75ecfbab2e3fb8e7097c1b73eed32d7ca947f11ea198cea880 not found: ID does not exist" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.912860 4824 scope.go:117] "RemoveContainer" containerID="82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.913346 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3"} err="failed to get container status \"82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3\": rpc error: code = NotFound desc = could not find container \"82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3\": container with ID starting with 82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3 not found: ID does not exist" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.913371 4824 scope.go:117] "RemoveContainer" containerID="99d1fdcb6aaacd8685fdad0e8e6e667a311f27767270fdc0f22419bc3b8af38a" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.914151 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"99d1fdcb6aaacd8685fdad0e8e6e667a311f27767270fdc0f22419bc3b8af38a"} err="failed to get container status \"99d1fdcb6aaacd8685fdad0e8e6e667a311f27767270fdc0f22419bc3b8af38a\": rpc error: code = NotFound desc = could not find container \"99d1fdcb6aaacd8685fdad0e8e6e667a311f27767270fdc0f22419bc3b8af38a\": container with ID starting with 99d1fdcb6aaacd8685fdad0e8e6e667a311f27767270fdc0f22419bc3b8af38a not found: ID does not exist" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.914175 4824 scope.go:117] "RemoveContainer" containerID="86a6881f109dab1f910fe505a21c7c800e587de4dded2635147997c51c05c483" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.914737 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"86a6881f109dab1f910fe505a21c7c800e587de4dded2635147997c51c05c483"} err="failed to get container status \"86a6881f109dab1f910fe505a21c7c800e587de4dded2635147997c51c05c483\": rpc error: code = NotFound desc = could not find container \"86a6881f109dab1f910fe505a21c7c800e587de4dded2635147997c51c05c483\": container with ID starting with 86a6881f109dab1f910fe505a21c7c800e587de4dded2635147997c51c05c483 not found: ID does not exist" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.914758 4824 scope.go:117] "RemoveContainer" containerID="948e0497eaffa7b0d84b4085c9ed8d987eb2c6527c9a82c10e576ed9c9c1e21b" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.915106 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"948e0497eaffa7b0d84b4085c9ed8d987eb2c6527c9a82c10e576ed9c9c1e21b"} err="failed to get container status \"948e0497eaffa7b0d84b4085c9ed8d987eb2c6527c9a82c10e576ed9c9c1e21b\": rpc error: code = NotFound desc = could not find container \"948e0497eaffa7b0d84b4085c9ed8d987eb2c6527c9a82c10e576ed9c9c1e21b\": container with ID starting with 948e0497eaffa7b0d84b4085c9ed8d987eb2c6527c9a82c10e576ed9c9c1e21b not found: ID does not exist" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.915178 4824 scope.go:117] "RemoveContainer" containerID="fc4c96e33e069f7d08d33caf92cd0987668506f931fd134b246d5ce5e934f95c" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.915707 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fc4c96e33e069f7d08d33caf92cd0987668506f931fd134b246d5ce5e934f95c"} err="failed to get container status \"fc4c96e33e069f7d08d33caf92cd0987668506f931fd134b246d5ce5e934f95c\": rpc error: code = NotFound desc = could not find container \"fc4c96e33e069f7d08d33caf92cd0987668506f931fd134b246d5ce5e934f95c\": container with ID starting with fc4c96e33e069f7d08d33caf92cd0987668506f931fd134b246d5ce5e934f95c not found: ID does not exist" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.915732 4824 scope.go:117] "RemoveContainer" containerID="ddbb0e5ae02e9d5d82a80ead7efc91f1b8e557a9fb6f39940e5530fae8de2d38" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.916030 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ddbb0e5ae02e9d5d82a80ead7efc91f1b8e557a9fb6f39940e5530fae8de2d38"} err="failed to get container status \"ddbb0e5ae02e9d5d82a80ead7efc91f1b8e557a9fb6f39940e5530fae8de2d38\": rpc error: code = NotFound desc = could not find container \"ddbb0e5ae02e9d5d82a80ead7efc91f1b8e557a9fb6f39940e5530fae8de2d38\": container with ID starting with ddbb0e5ae02e9d5d82a80ead7efc91f1b8e557a9fb6f39940e5530fae8de2d38 not found: ID does not exist" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.916053 4824 scope.go:117] "RemoveContainer" containerID="cee9cc9eec3fc815f1621a22a37805691288dbb60dd3ff64194a84be62abcddc" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.916226 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cee9cc9eec3fc815f1621a22a37805691288dbb60dd3ff64194a84be62abcddc"} err="failed to get container status \"cee9cc9eec3fc815f1621a22a37805691288dbb60dd3ff64194a84be62abcddc\": rpc error: code = NotFound desc = could not find container \"cee9cc9eec3fc815f1621a22a37805691288dbb60dd3ff64194a84be62abcddc\": container with ID starting with cee9cc9eec3fc815f1621a22a37805691288dbb60dd3ff64194a84be62abcddc not found: ID does not exist" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.916241 4824 scope.go:117] "RemoveContainer" containerID="fc319bbc06568cc811ded2b418b04b986a31f321fed492e8be50104af79bb30b" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.916397 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fc319bbc06568cc811ded2b418b04b986a31f321fed492e8be50104af79bb30b"} err="failed to get container status \"fc319bbc06568cc811ded2b418b04b986a31f321fed492e8be50104af79bb30b\": rpc error: code = NotFound desc = could not find container \"fc319bbc06568cc811ded2b418b04b986a31f321fed492e8be50104af79bb30b\": container with ID starting with fc319bbc06568cc811ded2b418b04b986a31f321fed492e8be50104af79bb30b not found: ID does not exist" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.916415 4824 scope.go:117] "RemoveContainer" containerID="64a9065254f889b1c43d63759283335bf78fc4fa1531a4baae472c06c075dae3" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.916582 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"64a9065254f889b1c43d63759283335bf78fc4fa1531a4baae472c06c075dae3"} err="failed to get container status \"64a9065254f889b1c43d63759283335bf78fc4fa1531a4baae472c06c075dae3\": rpc error: code = NotFound desc = could not find container \"64a9065254f889b1c43d63759283335bf78fc4fa1531a4baae472c06c075dae3\": container with ID starting with 64a9065254f889b1c43d63759283335bf78fc4fa1531a4baae472c06c075dae3 not found: ID does not exist" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.916599 4824 scope.go:117] "RemoveContainer" containerID="5e94ef8a8f47dc75ecfbab2e3fb8e7097c1b73eed32d7ca947f11ea198cea880" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.916819 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5e94ef8a8f47dc75ecfbab2e3fb8e7097c1b73eed32d7ca947f11ea198cea880"} err="failed to get container status \"5e94ef8a8f47dc75ecfbab2e3fb8e7097c1b73eed32d7ca947f11ea198cea880\": rpc error: code = NotFound desc = could not find container \"5e94ef8a8f47dc75ecfbab2e3fb8e7097c1b73eed32d7ca947f11ea198cea880\": container with ID starting with 5e94ef8a8f47dc75ecfbab2e3fb8e7097c1b73eed32d7ca947f11ea198cea880 not found: ID does not exist" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.916834 4824 scope.go:117] "RemoveContainer" containerID="82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.917203 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3"} err="failed to get container status \"82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3\": rpc error: code = NotFound desc = could not find container \"82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3\": container with ID starting with 82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3 not found: ID does not exist" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.917221 4824 scope.go:117] "RemoveContainer" containerID="99d1fdcb6aaacd8685fdad0e8e6e667a311f27767270fdc0f22419bc3b8af38a" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.917723 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"99d1fdcb6aaacd8685fdad0e8e6e667a311f27767270fdc0f22419bc3b8af38a"} err="failed to get container status \"99d1fdcb6aaacd8685fdad0e8e6e667a311f27767270fdc0f22419bc3b8af38a\": rpc error: code = NotFound desc = could not find container \"99d1fdcb6aaacd8685fdad0e8e6e667a311f27767270fdc0f22419bc3b8af38a\": container with ID starting with 99d1fdcb6aaacd8685fdad0e8e6e667a311f27767270fdc0f22419bc3b8af38a not found: ID does not exist" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.917738 4824 scope.go:117] "RemoveContainer" containerID="86a6881f109dab1f910fe505a21c7c800e587de4dded2635147997c51c05c483" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.917920 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"86a6881f109dab1f910fe505a21c7c800e587de4dded2635147997c51c05c483"} err="failed to get container status \"86a6881f109dab1f910fe505a21c7c800e587de4dded2635147997c51c05c483\": rpc error: code = NotFound desc = could not find container \"86a6881f109dab1f910fe505a21c7c800e587de4dded2635147997c51c05c483\": container with ID starting with 86a6881f109dab1f910fe505a21c7c800e587de4dded2635147997c51c05c483 not found: ID does not exist" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.917936 4824 scope.go:117] "RemoveContainer" containerID="948e0497eaffa7b0d84b4085c9ed8d987eb2c6527c9a82c10e576ed9c9c1e21b" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.918285 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"948e0497eaffa7b0d84b4085c9ed8d987eb2c6527c9a82c10e576ed9c9c1e21b"} err="failed to get container status \"948e0497eaffa7b0d84b4085c9ed8d987eb2c6527c9a82c10e576ed9c9c1e21b\": rpc error: code = NotFound desc = could not find container \"948e0497eaffa7b0d84b4085c9ed8d987eb2c6527c9a82c10e576ed9c9c1e21b\": container with ID starting with 948e0497eaffa7b0d84b4085c9ed8d987eb2c6527c9a82c10e576ed9c9c1e21b not found: ID does not exist" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.918317 4824 scope.go:117] "RemoveContainer" containerID="fc4c96e33e069f7d08d33caf92cd0987668506f931fd134b246d5ce5e934f95c" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.918538 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fc4c96e33e069f7d08d33caf92cd0987668506f931fd134b246d5ce5e934f95c"} err="failed to get container status \"fc4c96e33e069f7d08d33caf92cd0987668506f931fd134b246d5ce5e934f95c\": rpc error: code = NotFound desc = could not find container \"fc4c96e33e069f7d08d33caf92cd0987668506f931fd134b246d5ce5e934f95c\": container with ID starting with fc4c96e33e069f7d08d33caf92cd0987668506f931fd134b246d5ce5e934f95c not found: ID does not exist" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.918563 4824 scope.go:117] "RemoveContainer" containerID="ddbb0e5ae02e9d5d82a80ead7efc91f1b8e557a9fb6f39940e5530fae8de2d38" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.918839 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ddbb0e5ae02e9d5d82a80ead7efc91f1b8e557a9fb6f39940e5530fae8de2d38"} err="failed to get container status \"ddbb0e5ae02e9d5d82a80ead7efc91f1b8e557a9fb6f39940e5530fae8de2d38\": rpc error: code = NotFound desc = could not find container \"ddbb0e5ae02e9d5d82a80ead7efc91f1b8e557a9fb6f39940e5530fae8de2d38\": container with ID starting with ddbb0e5ae02e9d5d82a80ead7efc91f1b8e557a9fb6f39940e5530fae8de2d38 not found: ID does not exist" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.918866 4824 scope.go:117] "RemoveContainer" containerID="cee9cc9eec3fc815f1621a22a37805691288dbb60dd3ff64194a84be62abcddc" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.919270 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cee9cc9eec3fc815f1621a22a37805691288dbb60dd3ff64194a84be62abcddc"} err="failed to get container status \"cee9cc9eec3fc815f1621a22a37805691288dbb60dd3ff64194a84be62abcddc\": rpc error: code = NotFound desc = could not find container \"cee9cc9eec3fc815f1621a22a37805691288dbb60dd3ff64194a84be62abcddc\": container with ID starting with cee9cc9eec3fc815f1621a22a37805691288dbb60dd3ff64194a84be62abcddc not found: ID does not exist" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.919291 4824 scope.go:117] "RemoveContainer" containerID="fc319bbc06568cc811ded2b418b04b986a31f321fed492e8be50104af79bb30b" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.919456 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fc319bbc06568cc811ded2b418b04b986a31f321fed492e8be50104af79bb30b"} err="failed to get container status \"fc319bbc06568cc811ded2b418b04b986a31f321fed492e8be50104af79bb30b\": rpc error: code = NotFound desc = could not find container \"fc319bbc06568cc811ded2b418b04b986a31f321fed492e8be50104af79bb30b\": container with ID starting with fc319bbc06568cc811ded2b418b04b986a31f321fed492e8be50104af79bb30b not found: ID does not exist" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.919480 4824 scope.go:117] "RemoveContainer" containerID="64a9065254f889b1c43d63759283335bf78fc4fa1531a4baae472c06c075dae3" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.919764 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"64a9065254f889b1c43d63759283335bf78fc4fa1531a4baae472c06c075dae3"} err="failed to get container status \"64a9065254f889b1c43d63759283335bf78fc4fa1531a4baae472c06c075dae3\": rpc error: code = NotFound desc = could not find container \"64a9065254f889b1c43d63759283335bf78fc4fa1531a4baae472c06c075dae3\": container with ID starting with 64a9065254f889b1c43d63759283335bf78fc4fa1531a4baae472c06c075dae3 not found: ID does not exist" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.919787 4824 scope.go:117] "RemoveContainer" containerID="5e94ef8a8f47dc75ecfbab2e3fb8e7097c1b73eed32d7ca947f11ea198cea880" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.920120 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5e94ef8a8f47dc75ecfbab2e3fb8e7097c1b73eed32d7ca947f11ea198cea880"} err="failed to get container status \"5e94ef8a8f47dc75ecfbab2e3fb8e7097c1b73eed32d7ca947f11ea198cea880\": rpc error: code = NotFound desc = could not find container \"5e94ef8a8f47dc75ecfbab2e3fb8e7097c1b73eed32d7ca947f11ea198cea880\": container with ID starting with 5e94ef8a8f47dc75ecfbab2e3fb8e7097c1b73eed32d7ca947f11ea198cea880 not found: ID does not exist" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.920140 4824 scope.go:117] "RemoveContainer" containerID="82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.920309 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3"} err="failed to get container status \"82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3\": rpc error: code = NotFound desc = could not find container \"82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3\": container with ID starting with 82dea4417ac419c0aa08482780496180d7620af1e64b2c9c1db4c26f3e5970d3 not found: ID does not exist" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.920326 4824 scope.go:117] "RemoveContainer" containerID="99d1fdcb6aaacd8685fdad0e8e6e667a311f27767270fdc0f22419bc3b8af38a" Jan 21 11:20:26 crc kubenswrapper[4824]: I0121 11:20:26.920520 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"99d1fdcb6aaacd8685fdad0e8e6e667a311f27767270fdc0f22419bc3b8af38a"} err="failed to get container status \"99d1fdcb6aaacd8685fdad0e8e6e667a311f27767270fdc0f22419bc3b8af38a\": rpc error: code = NotFound desc = could not find container \"99d1fdcb6aaacd8685fdad0e8e6e667a311f27767270fdc0f22419bc3b8af38a\": container with ID starting with 99d1fdcb6aaacd8685fdad0e8e6e667a311f27767270fdc0f22419bc3b8af38a not found: ID does not exist" Jan 21 11:20:27 crc kubenswrapper[4824]: I0121 11:20:27.793797 4824 generic.go:334] "Generic (PLEG): container finished" podID="177dd5b5-7311-483f-8c21-0dae78cd3d09" containerID="dba9e5aa729335a50db7092a27a944bd16d57947741a4fe22c5ebbecd3bd67d3" exitCode=0 Jan 21 11:20:27 crc kubenswrapper[4824]: I0121 11:20:27.793847 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-r8p75" event={"ID":"177dd5b5-7311-483f-8c21-0dae78cd3d09","Type":"ContainerDied","Data":"dba9e5aa729335a50db7092a27a944bd16d57947741a4fe22c5ebbecd3bd67d3"} Jan 21 11:20:27 crc kubenswrapper[4824]: I0121 11:20:27.794026 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-r8p75" event={"ID":"177dd5b5-7311-483f-8c21-0dae78cd3d09","Type":"ContainerStarted","Data":"2d3f3ce0ea528595b97242e1fdaded28cfa81f33c89f8948464949ba03f16b59"} Jan 21 11:20:28 crc kubenswrapper[4824]: I0121 11:20:28.054083 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="26b1108a-4b98-4e39-a3cd-e0d055089fd8" path="/var/lib/kubelet/pods/26b1108a-4b98-4e39-a3cd-e0d055089fd8/volumes" Jan 21 11:20:28 crc kubenswrapper[4824]: I0121 11:20:28.801000 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-r8p75" event={"ID":"177dd5b5-7311-483f-8c21-0dae78cd3d09","Type":"ContainerStarted","Data":"9ce84d1b9a0df30346d416335b9702c66fa67375f6060d703f03b3698a0f8673"} Jan 21 11:20:28 crc kubenswrapper[4824]: I0121 11:20:28.801192 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-r8p75" event={"ID":"177dd5b5-7311-483f-8c21-0dae78cd3d09","Type":"ContainerStarted","Data":"7d8ed1d69da2c094a0c9e24cdf8fce33f5e3c95e0b157a8b595e4220bed4a1da"} Jan 21 11:20:28 crc kubenswrapper[4824]: I0121 11:20:28.801205 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-r8p75" event={"ID":"177dd5b5-7311-483f-8c21-0dae78cd3d09","Type":"ContainerStarted","Data":"b5bb983c3637b349c5675ac216689b556a6378d13e14e9565faaa1a42d4f10bf"} Jan 21 11:20:28 crc kubenswrapper[4824]: I0121 11:20:28.801214 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-r8p75" event={"ID":"177dd5b5-7311-483f-8c21-0dae78cd3d09","Type":"ContainerStarted","Data":"1b36a9038b6c5163b2384d2b4ba0cbfe0a0f50318802f964ccd26242e63c7f48"} Jan 21 11:20:28 crc kubenswrapper[4824]: I0121 11:20:28.801223 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-r8p75" event={"ID":"177dd5b5-7311-483f-8c21-0dae78cd3d09","Type":"ContainerStarted","Data":"54d7dc9977042eacfe440060eef65173d1c15fa89af0c71f3f672e6c5a24c539"} Jan 21 11:20:28 crc kubenswrapper[4824]: I0121 11:20:28.801231 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-r8p75" event={"ID":"177dd5b5-7311-483f-8c21-0dae78cd3d09","Type":"ContainerStarted","Data":"0ec365d42cddf75d8f253505f55690d7185fa643d4972f13b804be54193e743c"} Jan 21 11:20:30 crc kubenswrapper[4824]: I0121 11:20:30.814837 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-r8p75" event={"ID":"177dd5b5-7311-483f-8c21-0dae78cd3d09","Type":"ContainerStarted","Data":"92696eaafafab924a846f8ec543265b14a1d4eeddd3ea442d5e797aa54f4e71b"} Jan 21 11:20:32 crc kubenswrapper[4824]: I0121 11:20:32.825462 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-r8p75" event={"ID":"177dd5b5-7311-483f-8c21-0dae78cd3d09","Type":"ContainerStarted","Data":"87209c53211cdf230973b41ca812317014158ddab635b4d6b0acc166ec380ccd"} Jan 21 11:20:32 crc kubenswrapper[4824]: I0121 11:20:32.825700 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-r8p75" Jan 21 11:20:32 crc kubenswrapper[4824]: I0121 11:20:32.825886 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-r8p75" Jan 21 11:20:32 crc kubenswrapper[4824]: I0121 11:20:32.825900 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-r8p75" Jan 21 11:20:32 crc kubenswrapper[4824]: I0121 11:20:32.845445 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-r8p75" podStartSLOduration=6.845431265 podStartE2EDuration="6.845431265s" podCreationTimestamp="2026-01-21 11:20:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:20:32.843111304 +0000 UTC m=+575.136140596" watchObservedRunningTime="2026-01-21 11:20:32.845431265 +0000 UTC m=+575.138460557" Jan 21 11:20:32 crc kubenswrapper[4824]: I0121 11:20:32.845807 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-r8p75" Jan 21 11:20:32 crc kubenswrapper[4824]: I0121 11:20:32.850436 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-r8p75" Jan 21 11:20:40 crc kubenswrapper[4824]: I0121 11:20:40.049532 4824 scope.go:117] "RemoveContainer" containerID="80258959697dd4e8883d646587a5ac979419b23c4e94eace61a8b23e0885b470" Jan 21 11:20:40 crc kubenswrapper[4824]: E0121 11:20:40.049939 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-x94nb_openshift-multus(6b9588d1-5876-4c2c-8de5-e5c0ad46f04e)\"" pod="openshift-multus/multus-x94nb" podUID="6b9588d1-5876-4c2c-8de5-e5c0ad46f04e" Jan 21 11:20:46 crc kubenswrapper[4824]: I0121 11:20:46.064668 4824 patch_prober.go:28] interesting pod/machine-config-daemon-zqjg9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 21 11:20:46 crc kubenswrapper[4824]: I0121 11:20:46.065060 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 21 11:20:46 crc kubenswrapper[4824]: I0121 11:20:46.065092 4824 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" Jan 21 11:20:46 crc kubenswrapper[4824]: I0121 11:20:46.065519 4824 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"0494d41399c844b692156a76450e15e5791154f06b98545910f6ac8899bc8509"} pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 21 11:20:46 crc kubenswrapper[4824]: I0121 11:20:46.065569 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" containerName="machine-config-daemon" containerID="cri-o://0494d41399c844b692156a76450e15e5791154f06b98545910f6ac8899bc8509" gracePeriod=600 Jan 21 11:20:46 crc kubenswrapper[4824]: I0121 11:20:46.883214 4824 generic.go:334] "Generic (PLEG): container finished" podID="33f3d922-4ffe-409b-a49a-d88c85898260" containerID="0494d41399c844b692156a76450e15e5791154f06b98545910f6ac8899bc8509" exitCode=0 Jan 21 11:20:46 crc kubenswrapper[4824]: I0121 11:20:46.883271 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" event={"ID":"33f3d922-4ffe-409b-a49a-d88c85898260","Type":"ContainerDied","Data":"0494d41399c844b692156a76450e15e5791154f06b98545910f6ac8899bc8509"} Jan 21 11:20:46 crc kubenswrapper[4824]: I0121 11:20:46.883520 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" event={"ID":"33f3d922-4ffe-409b-a49a-d88c85898260","Type":"ContainerStarted","Data":"0c03e9f8dae3ff83ead8b2e38560a1562de223bd29105640743d2bd8e0cdca44"} Jan 21 11:20:46 crc kubenswrapper[4824]: I0121 11:20:46.883539 4824 scope.go:117] "RemoveContainer" containerID="04e919af77a770ce30313c5a8770baf390f15a9d134ed04fa80518ff38656c8a" Jan 21 11:20:53 crc kubenswrapper[4824]: I0121 11:20:53.049401 4824 scope.go:117] "RemoveContainer" containerID="80258959697dd4e8883d646587a5ac979419b23c4e94eace61a8b23e0885b470" Jan 21 11:20:53 crc kubenswrapper[4824]: I0121 11:20:53.916115 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-x94nb_6b9588d1-5876-4c2c-8de5-e5c0ad46f04e/kube-multus/2.log" Jan 21 11:20:53 crc kubenswrapper[4824]: I0121 11:20:53.916724 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-x94nb_6b9588d1-5876-4c2c-8de5-e5c0ad46f04e/kube-multus/1.log" Jan 21 11:20:53 crc kubenswrapper[4824]: I0121 11:20:53.916865 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-x94nb" event={"ID":"6b9588d1-5876-4c2c-8de5-e5c0ad46f04e","Type":"ContainerStarted","Data":"29d5c1eba74f1cd555791e1c196c0fb86bfc85e8e7e102815995a7488969c8b1"} Jan 21 11:20:54 crc kubenswrapper[4824]: I0121 11:20:54.524444 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713g7gj4"] Jan 21 11:20:54 crc kubenswrapper[4824]: I0121 11:20:54.525469 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713g7gj4" Jan 21 11:20:54 crc kubenswrapper[4824]: I0121 11:20:54.527033 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Jan 21 11:20:54 crc kubenswrapper[4824]: I0121 11:20:54.532465 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713g7gj4"] Jan 21 11:20:54 crc kubenswrapper[4824]: I0121 11:20:54.653385 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sm5bq\" (UniqueName: \"kubernetes.io/projected/d2dd714a-8d35-49a1-b6c2-7bec802eed41-kube-api-access-sm5bq\") pod \"53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713g7gj4\" (UID: \"d2dd714a-8d35-49a1-b6c2-7bec802eed41\") " pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713g7gj4" Jan 21 11:20:54 crc kubenswrapper[4824]: I0121 11:20:54.653489 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d2dd714a-8d35-49a1-b6c2-7bec802eed41-util\") pod \"53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713g7gj4\" (UID: \"d2dd714a-8d35-49a1-b6c2-7bec802eed41\") " pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713g7gj4" Jan 21 11:20:54 crc kubenswrapper[4824]: I0121 11:20:54.653528 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d2dd714a-8d35-49a1-b6c2-7bec802eed41-bundle\") pod \"53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713g7gj4\" (UID: \"d2dd714a-8d35-49a1-b6c2-7bec802eed41\") " pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713g7gj4" Jan 21 11:20:54 crc kubenswrapper[4824]: I0121 11:20:54.754342 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d2dd714a-8d35-49a1-b6c2-7bec802eed41-util\") pod \"53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713g7gj4\" (UID: \"d2dd714a-8d35-49a1-b6c2-7bec802eed41\") " pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713g7gj4" Jan 21 11:20:54 crc kubenswrapper[4824]: I0121 11:20:54.754391 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d2dd714a-8d35-49a1-b6c2-7bec802eed41-bundle\") pod \"53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713g7gj4\" (UID: \"d2dd714a-8d35-49a1-b6c2-7bec802eed41\") " pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713g7gj4" Jan 21 11:20:54 crc kubenswrapper[4824]: I0121 11:20:54.754445 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sm5bq\" (UniqueName: \"kubernetes.io/projected/d2dd714a-8d35-49a1-b6c2-7bec802eed41-kube-api-access-sm5bq\") pod \"53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713g7gj4\" (UID: \"d2dd714a-8d35-49a1-b6c2-7bec802eed41\") " pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713g7gj4" Jan 21 11:20:54 crc kubenswrapper[4824]: I0121 11:20:54.754877 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d2dd714a-8d35-49a1-b6c2-7bec802eed41-util\") pod \"53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713g7gj4\" (UID: \"d2dd714a-8d35-49a1-b6c2-7bec802eed41\") " pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713g7gj4" Jan 21 11:20:54 crc kubenswrapper[4824]: I0121 11:20:54.754924 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d2dd714a-8d35-49a1-b6c2-7bec802eed41-bundle\") pod \"53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713g7gj4\" (UID: \"d2dd714a-8d35-49a1-b6c2-7bec802eed41\") " pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713g7gj4" Jan 21 11:20:54 crc kubenswrapper[4824]: I0121 11:20:54.768574 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sm5bq\" (UniqueName: \"kubernetes.io/projected/d2dd714a-8d35-49a1-b6c2-7bec802eed41-kube-api-access-sm5bq\") pod \"53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713g7gj4\" (UID: \"d2dd714a-8d35-49a1-b6c2-7bec802eed41\") " pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713g7gj4" Jan 21 11:20:54 crc kubenswrapper[4824]: I0121 11:20:54.836407 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713g7gj4" Jan 21 11:20:55 crc kubenswrapper[4824]: I0121 11:20:55.171003 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713g7gj4"] Jan 21 11:20:55 crc kubenswrapper[4824]: W0121 11:20:55.174841 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd2dd714a_8d35_49a1_b6c2_7bec802eed41.slice/crio-221fa46d7b92efb02152aa910e2106f2970814766b26d4e7889b1e275f167029 WatchSource:0}: Error finding container 221fa46d7b92efb02152aa910e2106f2970814766b26d4e7889b1e275f167029: Status 404 returned error can't find the container with id 221fa46d7b92efb02152aa910e2106f2970814766b26d4e7889b1e275f167029 Jan 21 11:20:55 crc kubenswrapper[4824]: I0121 11:20:55.933367 4824 generic.go:334] "Generic (PLEG): container finished" podID="d2dd714a-8d35-49a1-b6c2-7bec802eed41" containerID="ce25b5463ed4235aca7c296c4ce401a068db4acbc17350d0ada1212aa35d509a" exitCode=0 Jan 21 11:20:55 crc kubenswrapper[4824]: I0121 11:20:55.933429 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713g7gj4" event={"ID":"d2dd714a-8d35-49a1-b6c2-7bec802eed41","Type":"ContainerDied","Data":"ce25b5463ed4235aca7c296c4ce401a068db4acbc17350d0ada1212aa35d509a"} Jan 21 11:20:55 crc kubenswrapper[4824]: I0121 11:20:55.934143 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713g7gj4" event={"ID":"d2dd714a-8d35-49a1-b6c2-7bec802eed41","Type":"ContainerStarted","Data":"221fa46d7b92efb02152aa910e2106f2970814766b26d4e7889b1e275f167029"} Jan 21 11:20:56 crc kubenswrapper[4824]: I0121 11:20:56.837141 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-r8p75" Jan 21 11:20:57 crc kubenswrapper[4824]: I0121 11:20:57.943164 4824 generic.go:334] "Generic (PLEG): container finished" podID="d2dd714a-8d35-49a1-b6c2-7bec802eed41" containerID="fe4aa38cf65a78901ded8e31785ce9f1dc40110d372e124bc39a42bc8328b91e" exitCode=0 Jan 21 11:20:57 crc kubenswrapper[4824]: I0121 11:20:57.943260 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713g7gj4" event={"ID":"d2dd714a-8d35-49a1-b6c2-7bec802eed41","Type":"ContainerDied","Data":"fe4aa38cf65a78901ded8e31785ce9f1dc40110d372e124bc39a42bc8328b91e"} Jan 21 11:20:58 crc kubenswrapper[4824]: I0121 11:20:58.210078 4824 scope.go:117] "RemoveContainer" containerID="49ce055ec8090d6957cb7e23fe75edb82c0c2d7f87e7203d280606368dd3179f" Jan 21 11:20:58 crc kubenswrapper[4824]: I0121 11:20:58.948543 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-x94nb_6b9588d1-5876-4c2c-8de5-e5c0ad46f04e/kube-multus/2.log" Jan 21 11:20:58 crc kubenswrapper[4824]: I0121 11:20:58.950234 4824 generic.go:334] "Generic (PLEG): container finished" podID="d2dd714a-8d35-49a1-b6c2-7bec802eed41" containerID="e59780e1ea6b956c60ca193737940a3132676cbfa71031ce443e2652855dc396" exitCode=0 Jan 21 11:20:58 crc kubenswrapper[4824]: I0121 11:20:58.950264 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713g7gj4" event={"ID":"d2dd714a-8d35-49a1-b6c2-7bec802eed41","Type":"ContainerDied","Data":"e59780e1ea6b956c60ca193737940a3132676cbfa71031ce443e2652855dc396"} Jan 21 11:21:00 crc kubenswrapper[4824]: I0121 11:21:00.137337 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713g7gj4" Jan 21 11:21:00 crc kubenswrapper[4824]: I0121 11:21:00.313641 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d2dd714a-8d35-49a1-b6c2-7bec802eed41-bundle\") pod \"d2dd714a-8d35-49a1-b6c2-7bec802eed41\" (UID: \"d2dd714a-8d35-49a1-b6c2-7bec802eed41\") " Jan 21 11:21:00 crc kubenswrapper[4824]: I0121 11:21:00.313691 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sm5bq\" (UniqueName: \"kubernetes.io/projected/d2dd714a-8d35-49a1-b6c2-7bec802eed41-kube-api-access-sm5bq\") pod \"d2dd714a-8d35-49a1-b6c2-7bec802eed41\" (UID: \"d2dd714a-8d35-49a1-b6c2-7bec802eed41\") " Jan 21 11:21:00 crc kubenswrapper[4824]: I0121 11:21:00.313795 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d2dd714a-8d35-49a1-b6c2-7bec802eed41-util\") pod \"d2dd714a-8d35-49a1-b6c2-7bec802eed41\" (UID: \"d2dd714a-8d35-49a1-b6c2-7bec802eed41\") " Jan 21 11:21:00 crc kubenswrapper[4824]: I0121 11:21:00.314145 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d2dd714a-8d35-49a1-b6c2-7bec802eed41-bundle" (OuterVolumeSpecName: "bundle") pod "d2dd714a-8d35-49a1-b6c2-7bec802eed41" (UID: "d2dd714a-8d35-49a1-b6c2-7bec802eed41"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:21:00 crc kubenswrapper[4824]: I0121 11:21:00.318152 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d2dd714a-8d35-49a1-b6c2-7bec802eed41-kube-api-access-sm5bq" (OuterVolumeSpecName: "kube-api-access-sm5bq") pod "d2dd714a-8d35-49a1-b6c2-7bec802eed41" (UID: "d2dd714a-8d35-49a1-b6c2-7bec802eed41"). InnerVolumeSpecName "kube-api-access-sm5bq". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:21:00 crc kubenswrapper[4824]: I0121 11:21:00.323923 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d2dd714a-8d35-49a1-b6c2-7bec802eed41-util" (OuterVolumeSpecName: "util") pod "d2dd714a-8d35-49a1-b6c2-7bec802eed41" (UID: "d2dd714a-8d35-49a1-b6c2-7bec802eed41"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:21:00 crc kubenswrapper[4824]: I0121 11:21:00.414683 4824 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d2dd714a-8d35-49a1-b6c2-7bec802eed41-util\") on node \"crc\" DevicePath \"\"" Jan 21 11:21:00 crc kubenswrapper[4824]: I0121 11:21:00.414708 4824 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d2dd714a-8d35-49a1-b6c2-7bec802eed41-bundle\") on node \"crc\" DevicePath \"\"" Jan 21 11:21:00 crc kubenswrapper[4824]: I0121 11:21:00.414718 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sm5bq\" (UniqueName: \"kubernetes.io/projected/d2dd714a-8d35-49a1-b6c2-7bec802eed41-kube-api-access-sm5bq\") on node \"crc\" DevicePath \"\"" Jan 21 11:21:00 crc kubenswrapper[4824]: I0121 11:21:00.961168 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713g7gj4" event={"ID":"d2dd714a-8d35-49a1-b6c2-7bec802eed41","Type":"ContainerDied","Data":"221fa46d7b92efb02152aa910e2106f2970814766b26d4e7889b1e275f167029"} Jan 21 11:21:00 crc kubenswrapper[4824]: I0121 11:21:00.961203 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="221fa46d7b92efb02152aa910e2106f2970814766b26d4e7889b1e275f167029" Jan 21 11:21:00 crc kubenswrapper[4824]: I0121 11:21:00.961214 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713g7gj4" Jan 21 11:21:02 crc kubenswrapper[4824]: I0121 11:21:02.363258 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-646758c888-zmc4k"] Jan 21 11:21:02 crc kubenswrapper[4824]: E0121 11:21:02.363434 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2dd714a-8d35-49a1-b6c2-7bec802eed41" containerName="util" Jan 21 11:21:02 crc kubenswrapper[4824]: I0121 11:21:02.363445 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2dd714a-8d35-49a1-b6c2-7bec802eed41" containerName="util" Jan 21 11:21:02 crc kubenswrapper[4824]: E0121 11:21:02.363462 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2dd714a-8d35-49a1-b6c2-7bec802eed41" containerName="pull" Jan 21 11:21:02 crc kubenswrapper[4824]: I0121 11:21:02.363467 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2dd714a-8d35-49a1-b6c2-7bec802eed41" containerName="pull" Jan 21 11:21:02 crc kubenswrapper[4824]: E0121 11:21:02.363473 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2dd714a-8d35-49a1-b6c2-7bec802eed41" containerName="extract" Jan 21 11:21:02 crc kubenswrapper[4824]: I0121 11:21:02.363500 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2dd714a-8d35-49a1-b6c2-7bec802eed41" containerName="extract" Jan 21 11:21:02 crc kubenswrapper[4824]: I0121 11:21:02.363617 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="d2dd714a-8d35-49a1-b6c2-7bec802eed41" containerName="extract" Jan 21 11:21:02 crc kubenswrapper[4824]: I0121 11:21:02.363937 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-646758c888-zmc4k" Jan 21 11:21:02 crc kubenswrapper[4824]: I0121 11:21:02.365994 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-txpxx" Jan 21 11:21:02 crc kubenswrapper[4824]: I0121 11:21:02.365993 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Jan 21 11:21:02 crc kubenswrapper[4824]: I0121 11:21:02.366706 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Jan 21 11:21:02 crc kubenswrapper[4824]: I0121 11:21:02.376100 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-646758c888-zmc4k"] Jan 21 11:21:02 crc kubenswrapper[4824]: I0121 11:21:02.536329 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5kmt6\" (UniqueName: \"kubernetes.io/projected/df60ca53-0573-4a6a-9235-86e0a1c17f91-kube-api-access-5kmt6\") pod \"nmstate-operator-646758c888-zmc4k\" (UID: \"df60ca53-0573-4a6a-9235-86e0a1c17f91\") " pod="openshift-nmstate/nmstate-operator-646758c888-zmc4k" Jan 21 11:21:02 crc kubenswrapper[4824]: I0121 11:21:02.637858 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5kmt6\" (UniqueName: \"kubernetes.io/projected/df60ca53-0573-4a6a-9235-86e0a1c17f91-kube-api-access-5kmt6\") pod \"nmstate-operator-646758c888-zmc4k\" (UID: \"df60ca53-0573-4a6a-9235-86e0a1c17f91\") " pod="openshift-nmstate/nmstate-operator-646758c888-zmc4k" Jan 21 11:21:02 crc kubenswrapper[4824]: I0121 11:21:02.651724 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5kmt6\" (UniqueName: \"kubernetes.io/projected/df60ca53-0573-4a6a-9235-86e0a1c17f91-kube-api-access-5kmt6\") pod \"nmstate-operator-646758c888-zmc4k\" (UID: \"df60ca53-0573-4a6a-9235-86e0a1c17f91\") " pod="openshift-nmstate/nmstate-operator-646758c888-zmc4k" Jan 21 11:21:02 crc kubenswrapper[4824]: I0121 11:21:02.676127 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-646758c888-zmc4k" Jan 21 11:21:03 crc kubenswrapper[4824]: I0121 11:21:03.044416 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-646758c888-zmc4k"] Jan 21 11:21:03 crc kubenswrapper[4824]: W0121 11:21:03.049478 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddf60ca53_0573_4a6a_9235_86e0a1c17f91.slice/crio-2ebdc5a7fcbbac549ba525a48829fc07aa41ebd21fa133f17578560b4ad02d03 WatchSource:0}: Error finding container 2ebdc5a7fcbbac549ba525a48829fc07aa41ebd21fa133f17578560b4ad02d03: Status 404 returned error can't find the container with id 2ebdc5a7fcbbac549ba525a48829fc07aa41ebd21fa133f17578560b4ad02d03 Jan 21 11:21:03 crc kubenswrapper[4824]: I0121 11:21:03.974338 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-646758c888-zmc4k" event={"ID":"df60ca53-0573-4a6a-9235-86e0a1c17f91","Type":"ContainerStarted","Data":"2ebdc5a7fcbbac549ba525a48829fc07aa41ebd21fa133f17578560b4ad02d03"} Jan 21 11:21:04 crc kubenswrapper[4824]: I0121 11:21:04.980027 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-646758c888-zmc4k" event={"ID":"df60ca53-0573-4a6a-9235-86e0a1c17f91","Type":"ContainerStarted","Data":"2bf180cba71ee6b1e9512a8e513b6e4e50a4d442dae7a116fb2209f499f0cbbf"} Jan 21 11:21:05 crc kubenswrapper[4824]: I0121 11:21:05.735460 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-646758c888-zmc4k" podStartSLOduration=2.122936639 podStartE2EDuration="3.735444162s" podCreationTimestamp="2026-01-21 11:21:02 +0000 UTC" firstStartedPulling="2026-01-21 11:21:03.050562383 +0000 UTC m=+605.343591675" lastFinishedPulling="2026-01-21 11:21:04.663069906 +0000 UTC m=+606.956099198" observedRunningTime="2026-01-21 11:21:04.990183871 +0000 UTC m=+607.283213162" watchObservedRunningTime="2026-01-21 11:21:05.735444162 +0000 UTC m=+608.028473454" Jan 21 11:21:05 crc kubenswrapper[4824]: I0121 11:21:05.735697 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-54757c584b-w2j6n"] Jan 21 11:21:05 crc kubenswrapper[4824]: I0121 11:21:05.736445 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-54757c584b-w2j6n" Jan 21 11:21:05 crc kubenswrapper[4824]: I0121 11:21:05.737734 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-nm9q7" Jan 21 11:21:05 crc kubenswrapper[4824]: I0121 11:21:05.746998 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-8474b5b9d8-h6kkh"] Jan 21 11:21:05 crc kubenswrapper[4824]: I0121 11:21:05.747655 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-h6kkh" Jan 21 11:21:05 crc kubenswrapper[4824]: I0121 11:21:05.748904 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Jan 21 11:21:05 crc kubenswrapper[4824]: I0121 11:21:05.758827 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-54757c584b-w2j6n"] Jan 21 11:21:05 crc kubenswrapper[4824]: I0121 11:21:05.760702 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-8474b5b9d8-h6kkh"] Jan 21 11:21:05 crc kubenswrapper[4824]: I0121 11:21:05.763429 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-qjpv9"] Jan 21 11:21:05 crc kubenswrapper[4824]: I0121 11:21:05.764089 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-qjpv9" Jan 21 11:21:05 crc kubenswrapper[4824]: I0121 11:21:05.768175 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h849b\" (UniqueName: \"kubernetes.io/projected/edf54909-b53e-44de-90bc-355e23bbc767-kube-api-access-h849b\") pod \"nmstate-webhook-8474b5b9d8-h6kkh\" (UID: \"edf54909-b53e-44de-90bc-355e23bbc767\") " pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-h6kkh" Jan 21 11:21:05 crc kubenswrapper[4824]: I0121 11:21:05.768250 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/edf54909-b53e-44de-90bc-355e23bbc767-tls-key-pair\") pod \"nmstate-webhook-8474b5b9d8-h6kkh\" (UID: \"edf54909-b53e-44de-90bc-355e23bbc767\") " pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-h6kkh" Jan 21 11:21:05 crc kubenswrapper[4824]: I0121 11:21:05.835672 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7754f76f8b-xjqs5"] Jan 21 11:21:05 crc kubenswrapper[4824]: I0121 11:21:05.841834 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-xjqs5" Jan 21 11:21:05 crc kubenswrapper[4824]: I0121 11:21:05.843414 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Jan 21 11:21:05 crc kubenswrapper[4824]: I0121 11:21:05.843630 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-7tptd" Jan 21 11:21:05 crc kubenswrapper[4824]: I0121 11:21:05.844708 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Jan 21 11:21:05 crc kubenswrapper[4824]: I0121 11:21:05.845604 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7754f76f8b-xjqs5"] Jan 21 11:21:05 crc kubenswrapper[4824]: I0121 11:21:05.870268 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/edf54909-b53e-44de-90bc-355e23bbc767-tls-key-pair\") pod \"nmstate-webhook-8474b5b9d8-h6kkh\" (UID: \"edf54909-b53e-44de-90bc-355e23bbc767\") " pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-h6kkh" Jan 21 11:21:05 crc kubenswrapper[4824]: I0121 11:21:05.870316 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/a01ecff0-9051-4ddb-8c8e-a4f379b6c823-dbus-socket\") pod \"nmstate-handler-qjpv9\" (UID: \"a01ecff0-9051-4ddb-8c8e-a4f379b6c823\") " pod="openshift-nmstate/nmstate-handler-qjpv9" Jan 21 11:21:05 crc kubenswrapper[4824]: I0121 11:21:05.870342 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-znknp\" (UniqueName: \"kubernetes.io/projected/a01ecff0-9051-4ddb-8c8e-a4f379b6c823-kube-api-access-znknp\") pod \"nmstate-handler-qjpv9\" (UID: \"a01ecff0-9051-4ddb-8c8e-a4f379b6c823\") " pod="openshift-nmstate/nmstate-handler-qjpv9" Jan 21 11:21:05 crc kubenswrapper[4824]: I0121 11:21:05.870384 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/a01ecff0-9051-4ddb-8c8e-a4f379b6c823-nmstate-lock\") pod \"nmstate-handler-qjpv9\" (UID: \"a01ecff0-9051-4ddb-8c8e-a4f379b6c823\") " pod="openshift-nmstate/nmstate-handler-qjpv9" Jan 21 11:21:05 crc kubenswrapper[4824]: E0121 11:21:05.870396 4824 secret.go:188] Couldn't get secret openshift-nmstate/openshift-nmstate-webhook: secret "openshift-nmstate-webhook" not found Jan 21 11:21:05 crc kubenswrapper[4824]: I0121 11:21:05.870427 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h849b\" (UniqueName: \"kubernetes.io/projected/edf54909-b53e-44de-90bc-355e23bbc767-kube-api-access-h849b\") pod \"nmstate-webhook-8474b5b9d8-h6kkh\" (UID: \"edf54909-b53e-44de-90bc-355e23bbc767\") " pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-h6kkh" Jan 21 11:21:05 crc kubenswrapper[4824]: E0121 11:21:05.870445 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/edf54909-b53e-44de-90bc-355e23bbc767-tls-key-pair podName:edf54909-b53e-44de-90bc-355e23bbc767 nodeName:}" failed. No retries permitted until 2026-01-21 11:21:06.370431815 +0000 UTC m=+608.663461107 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "tls-key-pair" (UniqueName: "kubernetes.io/secret/edf54909-b53e-44de-90bc-355e23bbc767-tls-key-pair") pod "nmstate-webhook-8474b5b9d8-h6kkh" (UID: "edf54909-b53e-44de-90bc-355e23bbc767") : secret "openshift-nmstate-webhook" not found Jan 21 11:21:05 crc kubenswrapper[4824]: I0121 11:21:05.870468 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j8dmq\" (UniqueName: \"kubernetes.io/projected/30b70686-bf05-4798-b5dc-3c7dc9cee56a-kube-api-access-j8dmq\") pod \"nmstate-metrics-54757c584b-w2j6n\" (UID: \"30b70686-bf05-4798-b5dc-3c7dc9cee56a\") " pod="openshift-nmstate/nmstate-metrics-54757c584b-w2j6n" Jan 21 11:21:05 crc kubenswrapper[4824]: I0121 11:21:05.870519 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/1575ab9d-016f-403a-990b-3b4c4d53f6f1-plugin-serving-cert\") pod \"nmstate-console-plugin-7754f76f8b-xjqs5\" (UID: \"1575ab9d-016f-403a-990b-3b4c4d53f6f1\") " pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-xjqs5" Jan 21 11:21:05 crc kubenswrapper[4824]: I0121 11:21:05.870553 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nv4cc\" (UniqueName: \"kubernetes.io/projected/1575ab9d-016f-403a-990b-3b4c4d53f6f1-kube-api-access-nv4cc\") pod \"nmstate-console-plugin-7754f76f8b-xjqs5\" (UID: \"1575ab9d-016f-403a-990b-3b4c4d53f6f1\") " pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-xjqs5" Jan 21 11:21:05 crc kubenswrapper[4824]: I0121 11:21:05.870573 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/1575ab9d-016f-403a-990b-3b4c4d53f6f1-nginx-conf\") pod \"nmstate-console-plugin-7754f76f8b-xjqs5\" (UID: \"1575ab9d-016f-403a-990b-3b4c4d53f6f1\") " pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-xjqs5" Jan 21 11:21:05 crc kubenswrapper[4824]: I0121 11:21:05.870632 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/a01ecff0-9051-4ddb-8c8e-a4f379b6c823-ovs-socket\") pod \"nmstate-handler-qjpv9\" (UID: \"a01ecff0-9051-4ddb-8c8e-a4f379b6c823\") " pod="openshift-nmstate/nmstate-handler-qjpv9" Jan 21 11:21:05 crc kubenswrapper[4824]: I0121 11:21:05.893993 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h849b\" (UniqueName: \"kubernetes.io/projected/edf54909-b53e-44de-90bc-355e23bbc767-kube-api-access-h849b\") pod \"nmstate-webhook-8474b5b9d8-h6kkh\" (UID: \"edf54909-b53e-44de-90bc-355e23bbc767\") " pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-h6kkh" Jan 21 11:21:05 crc kubenswrapper[4824]: I0121 11:21:05.971095 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/a01ecff0-9051-4ddb-8c8e-a4f379b6c823-ovs-socket\") pod \"nmstate-handler-qjpv9\" (UID: \"a01ecff0-9051-4ddb-8c8e-a4f379b6c823\") " pod="openshift-nmstate/nmstate-handler-qjpv9" Jan 21 11:21:05 crc kubenswrapper[4824]: I0121 11:21:05.971400 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/a01ecff0-9051-4ddb-8c8e-a4f379b6c823-dbus-socket\") pod \"nmstate-handler-qjpv9\" (UID: \"a01ecff0-9051-4ddb-8c8e-a4f379b6c823\") " pod="openshift-nmstate/nmstate-handler-qjpv9" Jan 21 11:21:05 crc kubenswrapper[4824]: I0121 11:21:05.971423 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-znknp\" (UniqueName: \"kubernetes.io/projected/a01ecff0-9051-4ddb-8c8e-a4f379b6c823-kube-api-access-znknp\") pod \"nmstate-handler-qjpv9\" (UID: \"a01ecff0-9051-4ddb-8c8e-a4f379b6c823\") " pod="openshift-nmstate/nmstate-handler-qjpv9" Jan 21 11:21:05 crc kubenswrapper[4824]: I0121 11:21:05.971448 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/a01ecff0-9051-4ddb-8c8e-a4f379b6c823-nmstate-lock\") pod \"nmstate-handler-qjpv9\" (UID: \"a01ecff0-9051-4ddb-8c8e-a4f379b6c823\") " pod="openshift-nmstate/nmstate-handler-qjpv9" Jan 21 11:21:05 crc kubenswrapper[4824]: I0121 11:21:05.971499 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j8dmq\" (UniqueName: \"kubernetes.io/projected/30b70686-bf05-4798-b5dc-3c7dc9cee56a-kube-api-access-j8dmq\") pod \"nmstate-metrics-54757c584b-w2j6n\" (UID: \"30b70686-bf05-4798-b5dc-3c7dc9cee56a\") " pod="openshift-nmstate/nmstate-metrics-54757c584b-w2j6n" Jan 21 11:21:05 crc kubenswrapper[4824]: I0121 11:21:05.971519 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/1575ab9d-016f-403a-990b-3b4c4d53f6f1-plugin-serving-cert\") pod \"nmstate-console-plugin-7754f76f8b-xjqs5\" (UID: \"1575ab9d-016f-403a-990b-3b4c4d53f6f1\") " pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-xjqs5" Jan 21 11:21:05 crc kubenswrapper[4824]: I0121 11:21:05.971541 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nv4cc\" (UniqueName: \"kubernetes.io/projected/1575ab9d-016f-403a-990b-3b4c4d53f6f1-kube-api-access-nv4cc\") pod \"nmstate-console-plugin-7754f76f8b-xjqs5\" (UID: \"1575ab9d-016f-403a-990b-3b4c4d53f6f1\") " pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-xjqs5" Jan 21 11:21:05 crc kubenswrapper[4824]: I0121 11:21:05.971557 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/1575ab9d-016f-403a-990b-3b4c4d53f6f1-nginx-conf\") pod \"nmstate-console-plugin-7754f76f8b-xjqs5\" (UID: \"1575ab9d-016f-403a-990b-3b4c4d53f6f1\") " pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-xjqs5" Jan 21 11:21:05 crc kubenswrapper[4824]: I0121 11:21:05.971653 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/a01ecff0-9051-4ddb-8c8e-a4f379b6c823-dbus-socket\") pod \"nmstate-handler-qjpv9\" (UID: \"a01ecff0-9051-4ddb-8c8e-a4f379b6c823\") " pod="openshift-nmstate/nmstate-handler-qjpv9" Jan 21 11:21:05 crc kubenswrapper[4824]: I0121 11:21:05.971211 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/a01ecff0-9051-4ddb-8c8e-a4f379b6c823-ovs-socket\") pod \"nmstate-handler-qjpv9\" (UID: \"a01ecff0-9051-4ddb-8c8e-a4f379b6c823\") " pod="openshift-nmstate/nmstate-handler-qjpv9" Jan 21 11:21:05 crc kubenswrapper[4824]: E0121 11:21:05.971806 4824 secret.go:188] Couldn't get secret openshift-nmstate/plugin-serving-cert: secret "plugin-serving-cert" not found Jan 21 11:21:05 crc kubenswrapper[4824]: E0121 11:21:05.971849 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1575ab9d-016f-403a-990b-3b4c4d53f6f1-plugin-serving-cert podName:1575ab9d-016f-403a-990b-3b4c4d53f6f1 nodeName:}" failed. No retries permitted until 2026-01-21 11:21:06.471836561 +0000 UTC m=+608.764865853 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "plugin-serving-cert" (UniqueName: "kubernetes.io/secret/1575ab9d-016f-403a-990b-3b4c4d53f6f1-plugin-serving-cert") pod "nmstate-console-plugin-7754f76f8b-xjqs5" (UID: "1575ab9d-016f-403a-990b-3b4c4d53f6f1") : secret "plugin-serving-cert" not found Jan 21 11:21:05 crc kubenswrapper[4824]: I0121 11:21:05.971916 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/a01ecff0-9051-4ddb-8c8e-a4f379b6c823-nmstate-lock\") pod \"nmstate-handler-qjpv9\" (UID: \"a01ecff0-9051-4ddb-8c8e-a4f379b6c823\") " pod="openshift-nmstate/nmstate-handler-qjpv9" Jan 21 11:21:05 crc kubenswrapper[4824]: I0121 11:21:05.972287 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/1575ab9d-016f-403a-990b-3b4c4d53f6f1-nginx-conf\") pod \"nmstate-console-plugin-7754f76f8b-xjqs5\" (UID: \"1575ab9d-016f-403a-990b-3b4c4d53f6f1\") " pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-xjqs5" Jan 21 11:21:05 crc kubenswrapper[4824]: I0121 11:21:05.994640 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nv4cc\" (UniqueName: \"kubernetes.io/projected/1575ab9d-016f-403a-990b-3b4c4d53f6f1-kube-api-access-nv4cc\") pod \"nmstate-console-plugin-7754f76f8b-xjqs5\" (UID: \"1575ab9d-016f-403a-990b-3b4c4d53f6f1\") " pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-xjqs5" Jan 21 11:21:05 crc kubenswrapper[4824]: I0121 11:21:05.995126 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-znknp\" (UniqueName: \"kubernetes.io/projected/a01ecff0-9051-4ddb-8c8e-a4f379b6c823-kube-api-access-znknp\") pod \"nmstate-handler-qjpv9\" (UID: \"a01ecff0-9051-4ddb-8c8e-a4f379b6c823\") " pod="openshift-nmstate/nmstate-handler-qjpv9" Jan 21 11:21:06 crc kubenswrapper[4824]: I0121 11:21:06.003540 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j8dmq\" (UniqueName: \"kubernetes.io/projected/30b70686-bf05-4798-b5dc-3c7dc9cee56a-kube-api-access-j8dmq\") pod \"nmstate-metrics-54757c584b-w2j6n\" (UID: \"30b70686-bf05-4798-b5dc-3c7dc9cee56a\") " pod="openshift-nmstate/nmstate-metrics-54757c584b-w2j6n" Jan 21 11:21:06 crc kubenswrapper[4824]: I0121 11:21:06.048205 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-54757c584b-w2j6n" Jan 21 11:21:06 crc kubenswrapper[4824]: I0121 11:21:06.076061 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-qjpv9" Jan 21 11:21:06 crc kubenswrapper[4824]: I0121 11:21:06.090179 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-68f8959f95-5v54n"] Jan 21 11:21:06 crc kubenswrapper[4824]: I0121 11:21:06.090811 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-68f8959f95-5v54n" Jan 21 11:21:06 crc kubenswrapper[4824]: W0121 11:21:06.094314 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda01ecff0_9051_4ddb_8c8e_a4f379b6c823.slice/crio-28b033c87d7f0443ef7f4ad888ac03af97dfddab52c519537fc47d807edac1d3 WatchSource:0}: Error finding container 28b033c87d7f0443ef7f4ad888ac03af97dfddab52c519537fc47d807edac1d3: Status 404 returned error can't find the container with id 28b033c87d7f0443ef7f4ad888ac03af97dfddab52c519537fc47d807edac1d3 Jan 21 11:21:06 crc kubenswrapper[4824]: I0121 11:21:06.105380 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-68f8959f95-5v54n"] Jan 21 11:21:06 crc kubenswrapper[4824]: I0121 11:21:06.173450 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/7bf3d16e-5b03-46d9-a09c-076fbea86f1d-console-serving-cert\") pod \"console-68f8959f95-5v54n\" (UID: \"7bf3d16e-5b03-46d9-a09c-076fbea86f1d\") " pod="openshift-console/console-68f8959f95-5v54n" Jan 21 11:21:06 crc kubenswrapper[4824]: I0121 11:21:06.173725 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/7bf3d16e-5b03-46d9-a09c-076fbea86f1d-oauth-serving-cert\") pod \"console-68f8959f95-5v54n\" (UID: \"7bf3d16e-5b03-46d9-a09c-076fbea86f1d\") " pod="openshift-console/console-68f8959f95-5v54n" Jan 21 11:21:06 crc kubenswrapper[4824]: I0121 11:21:06.173747 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7bf3d16e-5b03-46d9-a09c-076fbea86f1d-trusted-ca-bundle\") pod \"console-68f8959f95-5v54n\" (UID: \"7bf3d16e-5b03-46d9-a09c-076fbea86f1d\") " pod="openshift-console/console-68f8959f95-5v54n" Jan 21 11:21:06 crc kubenswrapper[4824]: I0121 11:21:06.173767 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/7bf3d16e-5b03-46d9-a09c-076fbea86f1d-console-oauth-config\") pod \"console-68f8959f95-5v54n\" (UID: \"7bf3d16e-5b03-46d9-a09c-076fbea86f1d\") " pod="openshift-console/console-68f8959f95-5v54n" Jan 21 11:21:06 crc kubenswrapper[4824]: I0121 11:21:06.173784 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/7bf3d16e-5b03-46d9-a09c-076fbea86f1d-service-ca\") pod \"console-68f8959f95-5v54n\" (UID: \"7bf3d16e-5b03-46d9-a09c-076fbea86f1d\") " pod="openshift-console/console-68f8959f95-5v54n" Jan 21 11:21:06 crc kubenswrapper[4824]: I0121 11:21:06.173952 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/7bf3d16e-5b03-46d9-a09c-076fbea86f1d-console-config\") pod \"console-68f8959f95-5v54n\" (UID: \"7bf3d16e-5b03-46d9-a09c-076fbea86f1d\") " pod="openshift-console/console-68f8959f95-5v54n" Jan 21 11:21:06 crc kubenswrapper[4824]: I0121 11:21:06.174034 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-48fpp\" (UniqueName: \"kubernetes.io/projected/7bf3d16e-5b03-46d9-a09c-076fbea86f1d-kube-api-access-48fpp\") pod \"console-68f8959f95-5v54n\" (UID: \"7bf3d16e-5b03-46d9-a09c-076fbea86f1d\") " pod="openshift-console/console-68f8959f95-5v54n" Jan 21 11:21:06 crc kubenswrapper[4824]: I0121 11:21:06.228378 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-54757c584b-w2j6n"] Jan 21 11:21:06 crc kubenswrapper[4824]: I0121 11:21:06.274559 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/7bf3d16e-5b03-46d9-a09c-076fbea86f1d-console-config\") pod \"console-68f8959f95-5v54n\" (UID: \"7bf3d16e-5b03-46d9-a09c-076fbea86f1d\") " pod="openshift-console/console-68f8959f95-5v54n" Jan 21 11:21:06 crc kubenswrapper[4824]: I0121 11:21:06.274602 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-48fpp\" (UniqueName: \"kubernetes.io/projected/7bf3d16e-5b03-46d9-a09c-076fbea86f1d-kube-api-access-48fpp\") pod \"console-68f8959f95-5v54n\" (UID: \"7bf3d16e-5b03-46d9-a09c-076fbea86f1d\") " pod="openshift-console/console-68f8959f95-5v54n" Jan 21 11:21:06 crc kubenswrapper[4824]: I0121 11:21:06.274656 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/7bf3d16e-5b03-46d9-a09c-076fbea86f1d-console-serving-cert\") pod \"console-68f8959f95-5v54n\" (UID: \"7bf3d16e-5b03-46d9-a09c-076fbea86f1d\") " pod="openshift-console/console-68f8959f95-5v54n" Jan 21 11:21:06 crc kubenswrapper[4824]: I0121 11:21:06.274683 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/7bf3d16e-5b03-46d9-a09c-076fbea86f1d-oauth-serving-cert\") pod \"console-68f8959f95-5v54n\" (UID: \"7bf3d16e-5b03-46d9-a09c-076fbea86f1d\") " pod="openshift-console/console-68f8959f95-5v54n" Jan 21 11:21:06 crc kubenswrapper[4824]: I0121 11:21:06.274699 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7bf3d16e-5b03-46d9-a09c-076fbea86f1d-trusted-ca-bundle\") pod \"console-68f8959f95-5v54n\" (UID: \"7bf3d16e-5b03-46d9-a09c-076fbea86f1d\") " pod="openshift-console/console-68f8959f95-5v54n" Jan 21 11:21:06 crc kubenswrapper[4824]: I0121 11:21:06.274713 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/7bf3d16e-5b03-46d9-a09c-076fbea86f1d-console-oauth-config\") pod \"console-68f8959f95-5v54n\" (UID: \"7bf3d16e-5b03-46d9-a09c-076fbea86f1d\") " pod="openshift-console/console-68f8959f95-5v54n" Jan 21 11:21:06 crc kubenswrapper[4824]: I0121 11:21:06.274731 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/7bf3d16e-5b03-46d9-a09c-076fbea86f1d-service-ca\") pod \"console-68f8959f95-5v54n\" (UID: \"7bf3d16e-5b03-46d9-a09c-076fbea86f1d\") " pod="openshift-console/console-68f8959f95-5v54n" Jan 21 11:21:06 crc kubenswrapper[4824]: I0121 11:21:06.275632 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/7bf3d16e-5b03-46d9-a09c-076fbea86f1d-service-ca\") pod \"console-68f8959f95-5v54n\" (UID: \"7bf3d16e-5b03-46d9-a09c-076fbea86f1d\") " pod="openshift-console/console-68f8959f95-5v54n" Jan 21 11:21:06 crc kubenswrapper[4824]: I0121 11:21:06.275658 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/7bf3d16e-5b03-46d9-a09c-076fbea86f1d-oauth-serving-cert\") pod \"console-68f8959f95-5v54n\" (UID: \"7bf3d16e-5b03-46d9-a09c-076fbea86f1d\") " pod="openshift-console/console-68f8959f95-5v54n" Jan 21 11:21:06 crc kubenswrapper[4824]: I0121 11:21:06.275794 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7bf3d16e-5b03-46d9-a09c-076fbea86f1d-trusted-ca-bundle\") pod \"console-68f8959f95-5v54n\" (UID: \"7bf3d16e-5b03-46d9-a09c-076fbea86f1d\") " pod="openshift-console/console-68f8959f95-5v54n" Jan 21 11:21:06 crc kubenswrapper[4824]: I0121 11:21:06.275918 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/7bf3d16e-5b03-46d9-a09c-076fbea86f1d-console-config\") pod \"console-68f8959f95-5v54n\" (UID: \"7bf3d16e-5b03-46d9-a09c-076fbea86f1d\") " pod="openshift-console/console-68f8959f95-5v54n" Jan 21 11:21:06 crc kubenswrapper[4824]: I0121 11:21:06.277795 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/7bf3d16e-5b03-46d9-a09c-076fbea86f1d-console-oauth-config\") pod \"console-68f8959f95-5v54n\" (UID: \"7bf3d16e-5b03-46d9-a09c-076fbea86f1d\") " pod="openshift-console/console-68f8959f95-5v54n" Jan 21 11:21:06 crc kubenswrapper[4824]: I0121 11:21:06.277984 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/7bf3d16e-5b03-46d9-a09c-076fbea86f1d-console-serving-cert\") pod \"console-68f8959f95-5v54n\" (UID: \"7bf3d16e-5b03-46d9-a09c-076fbea86f1d\") " pod="openshift-console/console-68f8959f95-5v54n" Jan 21 11:21:06 crc kubenswrapper[4824]: I0121 11:21:06.287780 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-48fpp\" (UniqueName: \"kubernetes.io/projected/7bf3d16e-5b03-46d9-a09c-076fbea86f1d-kube-api-access-48fpp\") pod \"console-68f8959f95-5v54n\" (UID: \"7bf3d16e-5b03-46d9-a09c-076fbea86f1d\") " pod="openshift-console/console-68f8959f95-5v54n" Jan 21 11:21:06 crc kubenswrapper[4824]: I0121 11:21:06.375319 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/edf54909-b53e-44de-90bc-355e23bbc767-tls-key-pair\") pod \"nmstate-webhook-8474b5b9d8-h6kkh\" (UID: \"edf54909-b53e-44de-90bc-355e23bbc767\") " pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-h6kkh" Jan 21 11:21:06 crc kubenswrapper[4824]: I0121 11:21:06.378589 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/edf54909-b53e-44de-90bc-355e23bbc767-tls-key-pair\") pod \"nmstate-webhook-8474b5b9d8-h6kkh\" (UID: \"edf54909-b53e-44de-90bc-355e23bbc767\") " pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-h6kkh" Jan 21 11:21:06 crc kubenswrapper[4824]: I0121 11:21:06.409541 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-68f8959f95-5v54n" Jan 21 11:21:06 crc kubenswrapper[4824]: I0121 11:21:06.476187 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/1575ab9d-016f-403a-990b-3b4c4d53f6f1-plugin-serving-cert\") pod \"nmstate-console-plugin-7754f76f8b-xjqs5\" (UID: \"1575ab9d-016f-403a-990b-3b4c4d53f6f1\") " pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-xjqs5" Jan 21 11:21:06 crc kubenswrapper[4824]: I0121 11:21:06.478770 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/1575ab9d-016f-403a-990b-3b4c4d53f6f1-plugin-serving-cert\") pod \"nmstate-console-plugin-7754f76f8b-xjqs5\" (UID: \"1575ab9d-016f-403a-990b-3b4c4d53f6f1\") " pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-xjqs5" Jan 21 11:21:06 crc kubenswrapper[4824]: I0121 11:21:06.661876 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-h6kkh" Jan 21 11:21:06 crc kubenswrapper[4824]: I0121 11:21:06.755187 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-xjqs5" Jan 21 11:21:06 crc kubenswrapper[4824]: I0121 11:21:06.761482 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-68f8959f95-5v54n"] Jan 21 11:21:06 crc kubenswrapper[4824]: W0121 11:21:06.766385 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7bf3d16e_5b03_46d9_a09c_076fbea86f1d.slice/crio-f2bd8103f9f0316e4fa3b3650c5acf4310b28769b7af0d9f813d086f15b076e6 WatchSource:0}: Error finding container f2bd8103f9f0316e4fa3b3650c5acf4310b28769b7af0d9f813d086f15b076e6: Status 404 returned error can't find the container with id f2bd8103f9f0316e4fa3b3650c5acf4310b28769b7af0d9f813d086f15b076e6 Jan 21 11:21:06 crc kubenswrapper[4824]: I0121 11:21:06.903282 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7754f76f8b-xjqs5"] Jan 21 11:21:06 crc kubenswrapper[4824]: W0121 11:21:06.912518 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1575ab9d_016f_403a_990b_3b4c4d53f6f1.slice/crio-4ff43fd17a851104dc40b9a7cffa012b1cff642cd009615e3cb6e93c57b2498f WatchSource:0}: Error finding container 4ff43fd17a851104dc40b9a7cffa012b1cff642cd009615e3cb6e93c57b2498f: Status 404 returned error can't find the container with id 4ff43fd17a851104dc40b9a7cffa012b1cff642cd009615e3cb6e93c57b2498f Jan 21 11:21:06 crc kubenswrapper[4824]: I0121 11:21:06.990758 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-68f8959f95-5v54n" event={"ID":"7bf3d16e-5b03-46d9-a09c-076fbea86f1d","Type":"ContainerStarted","Data":"f9d67e2a9ab1d7f288440eb669207a4003bbd4b49cdb9299daaa81316fdddd61"} Jan 21 11:21:06 crc kubenswrapper[4824]: I0121 11:21:06.991117 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-68f8959f95-5v54n" event={"ID":"7bf3d16e-5b03-46d9-a09c-076fbea86f1d","Type":"ContainerStarted","Data":"f2bd8103f9f0316e4fa3b3650c5acf4310b28769b7af0d9f813d086f15b076e6"} Jan 21 11:21:06 crc kubenswrapper[4824]: I0121 11:21:06.992456 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-xjqs5" event={"ID":"1575ab9d-016f-403a-990b-3b4c4d53f6f1","Type":"ContainerStarted","Data":"4ff43fd17a851104dc40b9a7cffa012b1cff642cd009615e3cb6e93c57b2498f"} Jan 21 11:21:06 crc kubenswrapper[4824]: I0121 11:21:06.994454 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-qjpv9" event={"ID":"a01ecff0-9051-4ddb-8c8e-a4f379b6c823","Type":"ContainerStarted","Data":"28b033c87d7f0443ef7f4ad888ac03af97dfddab52c519537fc47d807edac1d3"} Jan 21 11:21:06 crc kubenswrapper[4824]: I0121 11:21:06.995999 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-54757c584b-w2j6n" event={"ID":"30b70686-bf05-4798-b5dc-3c7dc9cee56a","Type":"ContainerStarted","Data":"77bcedfac8711709d0029f27edf7cb259166fb4d4d40ca365fd9d40e7a422c44"} Jan 21 11:21:07 crc kubenswrapper[4824]: I0121 11:21:07.012066 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-68f8959f95-5v54n" podStartSLOduration=1.012028192 podStartE2EDuration="1.012028192s" podCreationTimestamp="2026-01-21 11:21:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:21:07.003775946 +0000 UTC m=+609.296805238" watchObservedRunningTime="2026-01-21 11:21:07.012028192 +0000 UTC m=+609.305057484" Jan 21 11:21:07 crc kubenswrapper[4824]: I0121 11:21:07.052263 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-8474b5b9d8-h6kkh"] Jan 21 11:21:08 crc kubenswrapper[4824]: I0121 11:21:08.003227 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-h6kkh" event={"ID":"edf54909-b53e-44de-90bc-355e23bbc767","Type":"ContainerStarted","Data":"af018cf706cab903b3bc5faab1eb0afff311f9420b7439621c8d45b257a9ff07"} Jan 21 11:21:09 crc kubenswrapper[4824]: I0121 11:21:09.008241 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-54757c584b-w2j6n" event={"ID":"30b70686-bf05-4798-b5dc-3c7dc9cee56a","Type":"ContainerStarted","Data":"b8ea1c7781977f653e93553dad86c7e755fa043f3b3876c6acb9122e2e388446"} Jan 21 11:21:09 crc kubenswrapper[4824]: I0121 11:21:09.010583 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-h6kkh" event={"ID":"edf54909-b53e-44de-90bc-355e23bbc767","Type":"ContainerStarted","Data":"593c0deeb8c98823905656739795f6161a9cdf083e9c64bceec49e676cedb143"} Jan 21 11:21:09 crc kubenswrapper[4824]: I0121 11:21:09.010704 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-h6kkh" Jan 21 11:21:09 crc kubenswrapper[4824]: I0121 11:21:09.011991 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-xjqs5" event={"ID":"1575ab9d-016f-403a-990b-3b4c4d53f6f1","Type":"ContainerStarted","Data":"1b059417217898b40ec29c93b3b17a039f8ababce0e90ca359107aacd389cd30"} Jan 21 11:21:09 crc kubenswrapper[4824]: I0121 11:21:09.013088 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-qjpv9" event={"ID":"a01ecff0-9051-4ddb-8c8e-a4f379b6c823","Type":"ContainerStarted","Data":"24e5f66c433e84206edd442af9dec71753b7880999d768b147b02732caf21946"} Jan 21 11:21:09 crc kubenswrapper[4824]: I0121 11:21:09.013226 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-qjpv9" Jan 21 11:21:09 crc kubenswrapper[4824]: I0121 11:21:09.026370 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-h6kkh" podStartSLOduration=3.034105345 podStartE2EDuration="4.026357168s" podCreationTimestamp="2026-01-21 11:21:05 +0000 UTC" firstStartedPulling="2026-01-21 11:21:07.052261514 +0000 UTC m=+609.345290806" lastFinishedPulling="2026-01-21 11:21:08.044513337 +0000 UTC m=+610.337542629" observedRunningTime="2026-01-21 11:21:09.024772539 +0000 UTC m=+611.317801831" watchObservedRunningTime="2026-01-21 11:21:09.026357168 +0000 UTC m=+611.319386460" Jan 21 11:21:09 crc kubenswrapper[4824]: I0121 11:21:09.036716 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-xjqs5" podStartSLOduration=2.192533945 podStartE2EDuration="4.0367012s" podCreationTimestamp="2026-01-21 11:21:05 +0000 UTC" firstStartedPulling="2026-01-21 11:21:06.915380479 +0000 UTC m=+609.208409771" lastFinishedPulling="2026-01-21 11:21:08.759547733 +0000 UTC m=+611.052577026" observedRunningTime="2026-01-21 11:21:09.03436363 +0000 UTC m=+611.327392922" watchObservedRunningTime="2026-01-21 11:21:09.0367012 +0000 UTC m=+611.329730492" Jan 21 11:21:11 crc kubenswrapper[4824]: I0121 11:21:11.023072 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-54757c584b-w2j6n" event={"ID":"30b70686-bf05-4798-b5dc-3c7dc9cee56a","Type":"ContainerStarted","Data":"3d7e837c034443be3ebd5618c9a2e57a07267b1da725bf4eadf39c08a24b5208"} Jan 21 11:21:11 crc kubenswrapper[4824]: I0121 11:21:11.036413 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-qjpv9" podStartSLOduration=4.083457826 podStartE2EDuration="6.036401262s" podCreationTimestamp="2026-01-21 11:21:05 +0000 UTC" firstStartedPulling="2026-01-21 11:21:06.096059158 +0000 UTC m=+608.389088451" lastFinishedPulling="2026-01-21 11:21:08.049002594 +0000 UTC m=+610.342031887" observedRunningTime="2026-01-21 11:21:09.045900172 +0000 UTC m=+611.338929464" watchObservedRunningTime="2026-01-21 11:21:11.036401262 +0000 UTC m=+613.329430553" Jan 21 11:21:11 crc kubenswrapper[4824]: I0121 11:21:11.037644 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-54757c584b-w2j6n" podStartSLOduration=2.255024729 podStartE2EDuration="6.037636171s" podCreationTimestamp="2026-01-21 11:21:05 +0000 UTC" firstStartedPulling="2026-01-21 11:21:06.238621921 +0000 UTC m=+608.531651212" lastFinishedPulling="2026-01-21 11:21:10.021233362 +0000 UTC m=+612.314262654" observedRunningTime="2026-01-21 11:21:11.034036752 +0000 UTC m=+613.327066044" watchObservedRunningTime="2026-01-21 11:21:11.037636171 +0000 UTC m=+613.330665464" Jan 21 11:21:16 crc kubenswrapper[4824]: I0121 11:21:16.095474 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-qjpv9" Jan 21 11:21:16 crc kubenswrapper[4824]: I0121 11:21:16.410631 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-68f8959f95-5v54n" Jan 21 11:21:16 crc kubenswrapper[4824]: I0121 11:21:16.410678 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-68f8959f95-5v54n" Jan 21 11:21:16 crc kubenswrapper[4824]: I0121 11:21:16.414480 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-68f8959f95-5v54n" Jan 21 11:21:17 crc kubenswrapper[4824]: I0121 11:21:17.051904 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-68f8959f95-5v54n" Jan 21 11:21:17 crc kubenswrapper[4824]: I0121 11:21:17.082349 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-nwlwd"] Jan 21 11:21:26 crc kubenswrapper[4824]: I0121 11:21:26.666478 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-h6kkh" Jan 21 11:21:35 crc kubenswrapper[4824]: I0121 11:21:35.024296 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcwqkxq"] Jan 21 11:21:35 crc kubenswrapper[4824]: I0121 11:21:35.025543 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcwqkxq" Jan 21 11:21:35 crc kubenswrapper[4824]: I0121 11:21:35.026727 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Jan 21 11:21:35 crc kubenswrapper[4824]: I0121 11:21:35.036063 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcwqkxq"] Jan 21 11:21:35 crc kubenswrapper[4824]: I0121 11:21:35.180784 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tmm2f\" (UniqueName: \"kubernetes.io/projected/d9005d95-8246-435d-aaea-5c60d58175c6-kube-api-access-tmm2f\") pod \"270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcwqkxq\" (UID: \"d9005d95-8246-435d-aaea-5c60d58175c6\") " pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcwqkxq" Jan 21 11:21:35 crc kubenswrapper[4824]: I0121 11:21:35.180818 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d9005d95-8246-435d-aaea-5c60d58175c6-bundle\") pod \"270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcwqkxq\" (UID: \"d9005d95-8246-435d-aaea-5c60d58175c6\") " pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcwqkxq" Jan 21 11:21:35 crc kubenswrapper[4824]: I0121 11:21:35.180878 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d9005d95-8246-435d-aaea-5c60d58175c6-util\") pod \"270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcwqkxq\" (UID: \"d9005d95-8246-435d-aaea-5c60d58175c6\") " pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcwqkxq" Jan 21 11:21:35 crc kubenswrapper[4824]: I0121 11:21:35.282101 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tmm2f\" (UniqueName: \"kubernetes.io/projected/d9005d95-8246-435d-aaea-5c60d58175c6-kube-api-access-tmm2f\") pod \"270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcwqkxq\" (UID: \"d9005d95-8246-435d-aaea-5c60d58175c6\") " pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcwqkxq" Jan 21 11:21:35 crc kubenswrapper[4824]: I0121 11:21:35.282132 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d9005d95-8246-435d-aaea-5c60d58175c6-bundle\") pod \"270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcwqkxq\" (UID: \"d9005d95-8246-435d-aaea-5c60d58175c6\") " pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcwqkxq" Jan 21 11:21:35 crc kubenswrapper[4824]: I0121 11:21:35.282184 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d9005d95-8246-435d-aaea-5c60d58175c6-util\") pod \"270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcwqkxq\" (UID: \"d9005d95-8246-435d-aaea-5c60d58175c6\") " pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcwqkxq" Jan 21 11:21:35 crc kubenswrapper[4824]: I0121 11:21:35.282603 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d9005d95-8246-435d-aaea-5c60d58175c6-bundle\") pod \"270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcwqkxq\" (UID: \"d9005d95-8246-435d-aaea-5c60d58175c6\") " pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcwqkxq" Jan 21 11:21:35 crc kubenswrapper[4824]: I0121 11:21:35.282643 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d9005d95-8246-435d-aaea-5c60d58175c6-util\") pod \"270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcwqkxq\" (UID: \"d9005d95-8246-435d-aaea-5c60d58175c6\") " pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcwqkxq" Jan 21 11:21:35 crc kubenswrapper[4824]: I0121 11:21:35.296244 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tmm2f\" (UniqueName: \"kubernetes.io/projected/d9005d95-8246-435d-aaea-5c60d58175c6-kube-api-access-tmm2f\") pod \"270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcwqkxq\" (UID: \"d9005d95-8246-435d-aaea-5c60d58175c6\") " pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcwqkxq" Jan 21 11:21:35 crc kubenswrapper[4824]: I0121 11:21:35.337730 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcwqkxq" Jan 21 11:21:35 crc kubenswrapper[4824]: I0121 11:21:35.673022 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcwqkxq"] Jan 21 11:21:36 crc kubenswrapper[4824]: I0121 11:21:36.122277 4824 generic.go:334] "Generic (PLEG): container finished" podID="d9005d95-8246-435d-aaea-5c60d58175c6" containerID="27732c80d1e351e888b3afdbfb3574b4168cb28379f32229565b988ce91bccf7" exitCode=0 Jan 21 11:21:36 crc kubenswrapper[4824]: I0121 11:21:36.122351 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcwqkxq" event={"ID":"d9005d95-8246-435d-aaea-5c60d58175c6","Type":"ContainerDied","Data":"27732c80d1e351e888b3afdbfb3574b4168cb28379f32229565b988ce91bccf7"} Jan 21 11:21:36 crc kubenswrapper[4824]: I0121 11:21:36.122477 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcwqkxq" event={"ID":"d9005d95-8246-435d-aaea-5c60d58175c6","Type":"ContainerStarted","Data":"3169369e4043b11755f263f8f5ffe3764db94ba00d13ead4ec3ef289b12bce38"} Jan 21 11:21:38 crc kubenswrapper[4824]: I0121 11:21:38.131061 4824 generic.go:334] "Generic (PLEG): container finished" podID="d9005d95-8246-435d-aaea-5c60d58175c6" containerID="67331e20e703337d471521971fe82b6954704b075e1322eff28c99c0a434fb47" exitCode=0 Jan 21 11:21:38 crc kubenswrapper[4824]: I0121 11:21:38.131091 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcwqkxq" event={"ID":"d9005d95-8246-435d-aaea-5c60d58175c6","Type":"ContainerDied","Data":"67331e20e703337d471521971fe82b6954704b075e1322eff28c99c0a434fb47"} Jan 21 11:21:39 crc kubenswrapper[4824]: I0121 11:21:39.135920 4824 generic.go:334] "Generic (PLEG): container finished" podID="d9005d95-8246-435d-aaea-5c60d58175c6" containerID="bfd4bec6d81589bdf71cf6fa238cbe1dc1a1166fc05f16b127d0eb3075647620" exitCode=0 Jan 21 11:21:39 crc kubenswrapper[4824]: I0121 11:21:39.135984 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcwqkxq" event={"ID":"d9005d95-8246-435d-aaea-5c60d58175c6","Type":"ContainerDied","Data":"bfd4bec6d81589bdf71cf6fa238cbe1dc1a1166fc05f16b127d0eb3075647620"} Jan 21 11:21:40 crc kubenswrapper[4824]: I0121 11:21:40.305327 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcwqkxq" Jan 21 11:21:40 crc kubenswrapper[4824]: I0121 11:21:40.431625 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d9005d95-8246-435d-aaea-5c60d58175c6-bundle\") pod \"d9005d95-8246-435d-aaea-5c60d58175c6\" (UID: \"d9005d95-8246-435d-aaea-5c60d58175c6\") " Jan 21 11:21:40 crc kubenswrapper[4824]: I0121 11:21:40.431667 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tmm2f\" (UniqueName: \"kubernetes.io/projected/d9005d95-8246-435d-aaea-5c60d58175c6-kube-api-access-tmm2f\") pod \"d9005d95-8246-435d-aaea-5c60d58175c6\" (UID: \"d9005d95-8246-435d-aaea-5c60d58175c6\") " Jan 21 11:21:40 crc kubenswrapper[4824]: I0121 11:21:40.431705 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d9005d95-8246-435d-aaea-5c60d58175c6-util\") pod \"d9005d95-8246-435d-aaea-5c60d58175c6\" (UID: \"d9005d95-8246-435d-aaea-5c60d58175c6\") " Jan 21 11:21:40 crc kubenswrapper[4824]: I0121 11:21:40.432502 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d9005d95-8246-435d-aaea-5c60d58175c6-bundle" (OuterVolumeSpecName: "bundle") pod "d9005d95-8246-435d-aaea-5c60d58175c6" (UID: "d9005d95-8246-435d-aaea-5c60d58175c6"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:21:40 crc kubenswrapper[4824]: I0121 11:21:40.435948 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d9005d95-8246-435d-aaea-5c60d58175c6-kube-api-access-tmm2f" (OuterVolumeSpecName: "kube-api-access-tmm2f") pod "d9005d95-8246-435d-aaea-5c60d58175c6" (UID: "d9005d95-8246-435d-aaea-5c60d58175c6"). InnerVolumeSpecName "kube-api-access-tmm2f". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:21:40 crc kubenswrapper[4824]: I0121 11:21:40.441590 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d9005d95-8246-435d-aaea-5c60d58175c6-util" (OuterVolumeSpecName: "util") pod "d9005d95-8246-435d-aaea-5c60d58175c6" (UID: "d9005d95-8246-435d-aaea-5c60d58175c6"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:21:40 crc kubenswrapper[4824]: I0121 11:21:40.532694 4824 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d9005d95-8246-435d-aaea-5c60d58175c6-bundle\") on node \"crc\" DevicePath \"\"" Jan 21 11:21:40 crc kubenswrapper[4824]: I0121 11:21:40.532719 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tmm2f\" (UniqueName: \"kubernetes.io/projected/d9005d95-8246-435d-aaea-5c60d58175c6-kube-api-access-tmm2f\") on node \"crc\" DevicePath \"\"" Jan 21 11:21:40 crc kubenswrapper[4824]: I0121 11:21:40.532729 4824 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d9005d95-8246-435d-aaea-5c60d58175c6-util\") on node \"crc\" DevicePath \"\"" Jan 21 11:21:41 crc kubenswrapper[4824]: I0121 11:21:41.145123 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcwqkxq" event={"ID":"d9005d95-8246-435d-aaea-5c60d58175c6","Type":"ContainerDied","Data":"3169369e4043b11755f263f8f5ffe3764db94ba00d13ead4ec3ef289b12bce38"} Jan 21 11:21:41 crc kubenswrapper[4824]: I0121 11:21:41.145155 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3169369e4043b11755f263f8f5ffe3764db94ba00d13ead4ec3ef289b12bce38" Jan 21 11:21:41 crc kubenswrapper[4824]: I0121 11:21:41.145171 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcwqkxq" Jan 21 11:21:42 crc kubenswrapper[4824]: I0121 11:21:42.106596 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-nwlwd" podUID="c585ef2f-3e29-4b81-918f-11b075006ca0" containerName="console" containerID="cri-o://3a9966ed610a4f72371cd05219170eb5e70c5b477a36cba97f1fa20640f0ccd7" gracePeriod=15 Jan 21 11:21:42 crc kubenswrapper[4824]: I0121 11:21:42.384074 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-nwlwd_c585ef2f-3e29-4b81-918f-11b075006ca0/console/0.log" Jan 21 11:21:42 crc kubenswrapper[4824]: I0121 11:21:42.384137 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-nwlwd" Jan 21 11:21:42 crc kubenswrapper[4824]: I0121 11:21:42.552190 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/c585ef2f-3e29-4b81-918f-11b075006ca0-service-ca\") pod \"c585ef2f-3e29-4b81-918f-11b075006ca0\" (UID: \"c585ef2f-3e29-4b81-918f-11b075006ca0\") " Jan 21 11:21:42 crc kubenswrapper[4824]: I0121 11:21:42.552249 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q8qjw\" (UniqueName: \"kubernetes.io/projected/c585ef2f-3e29-4b81-918f-11b075006ca0-kube-api-access-q8qjw\") pod \"c585ef2f-3e29-4b81-918f-11b075006ca0\" (UID: \"c585ef2f-3e29-4b81-918f-11b075006ca0\") " Jan 21 11:21:42 crc kubenswrapper[4824]: I0121 11:21:42.552276 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/c585ef2f-3e29-4b81-918f-11b075006ca0-console-serving-cert\") pod \"c585ef2f-3e29-4b81-918f-11b075006ca0\" (UID: \"c585ef2f-3e29-4b81-918f-11b075006ca0\") " Jan 21 11:21:42 crc kubenswrapper[4824]: I0121 11:21:42.552303 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c585ef2f-3e29-4b81-918f-11b075006ca0-trusted-ca-bundle\") pod \"c585ef2f-3e29-4b81-918f-11b075006ca0\" (UID: \"c585ef2f-3e29-4b81-918f-11b075006ca0\") " Jan 21 11:21:42 crc kubenswrapper[4824]: I0121 11:21:42.552338 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/c585ef2f-3e29-4b81-918f-11b075006ca0-console-config\") pod \"c585ef2f-3e29-4b81-918f-11b075006ca0\" (UID: \"c585ef2f-3e29-4b81-918f-11b075006ca0\") " Jan 21 11:21:42 crc kubenswrapper[4824]: I0121 11:21:42.552385 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/c585ef2f-3e29-4b81-918f-11b075006ca0-oauth-serving-cert\") pod \"c585ef2f-3e29-4b81-918f-11b075006ca0\" (UID: \"c585ef2f-3e29-4b81-918f-11b075006ca0\") " Jan 21 11:21:42 crc kubenswrapper[4824]: I0121 11:21:42.552402 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/c585ef2f-3e29-4b81-918f-11b075006ca0-console-oauth-config\") pod \"c585ef2f-3e29-4b81-918f-11b075006ca0\" (UID: \"c585ef2f-3e29-4b81-918f-11b075006ca0\") " Jan 21 11:21:42 crc kubenswrapper[4824]: I0121 11:21:42.552871 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c585ef2f-3e29-4b81-918f-11b075006ca0-service-ca" (OuterVolumeSpecName: "service-ca") pod "c585ef2f-3e29-4b81-918f-11b075006ca0" (UID: "c585ef2f-3e29-4b81-918f-11b075006ca0"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:21:42 crc kubenswrapper[4824]: I0121 11:21:42.553148 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c585ef2f-3e29-4b81-918f-11b075006ca0-console-config" (OuterVolumeSpecName: "console-config") pod "c585ef2f-3e29-4b81-918f-11b075006ca0" (UID: "c585ef2f-3e29-4b81-918f-11b075006ca0"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:21:42 crc kubenswrapper[4824]: I0121 11:21:42.553310 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c585ef2f-3e29-4b81-918f-11b075006ca0-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "c585ef2f-3e29-4b81-918f-11b075006ca0" (UID: "c585ef2f-3e29-4b81-918f-11b075006ca0"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:21:42 crc kubenswrapper[4824]: I0121 11:21:42.553690 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c585ef2f-3e29-4b81-918f-11b075006ca0-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "c585ef2f-3e29-4b81-918f-11b075006ca0" (UID: "c585ef2f-3e29-4b81-918f-11b075006ca0"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:21:42 crc kubenswrapper[4824]: I0121 11:21:42.554080 4824 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c585ef2f-3e29-4b81-918f-11b075006ca0-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 21 11:21:42 crc kubenswrapper[4824]: I0121 11:21:42.554147 4824 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/c585ef2f-3e29-4b81-918f-11b075006ca0-console-config\") on node \"crc\" DevicePath \"\"" Jan 21 11:21:42 crc kubenswrapper[4824]: I0121 11:21:42.554161 4824 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/c585ef2f-3e29-4b81-918f-11b075006ca0-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 21 11:21:42 crc kubenswrapper[4824]: I0121 11:21:42.554171 4824 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/c585ef2f-3e29-4b81-918f-11b075006ca0-service-ca\") on node \"crc\" DevicePath \"\"" Jan 21 11:21:42 crc kubenswrapper[4824]: I0121 11:21:42.556849 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c585ef2f-3e29-4b81-918f-11b075006ca0-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "c585ef2f-3e29-4b81-918f-11b075006ca0" (UID: "c585ef2f-3e29-4b81-918f-11b075006ca0"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:21:42 crc kubenswrapper[4824]: I0121 11:21:42.556928 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c585ef2f-3e29-4b81-918f-11b075006ca0-kube-api-access-q8qjw" (OuterVolumeSpecName: "kube-api-access-q8qjw") pod "c585ef2f-3e29-4b81-918f-11b075006ca0" (UID: "c585ef2f-3e29-4b81-918f-11b075006ca0"). InnerVolumeSpecName "kube-api-access-q8qjw". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:21:42 crc kubenswrapper[4824]: I0121 11:21:42.557022 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c585ef2f-3e29-4b81-918f-11b075006ca0-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "c585ef2f-3e29-4b81-918f-11b075006ca0" (UID: "c585ef2f-3e29-4b81-918f-11b075006ca0"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:21:42 crc kubenswrapper[4824]: I0121 11:21:42.654804 4824 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/c585ef2f-3e29-4b81-918f-11b075006ca0-console-oauth-config\") on node \"crc\" DevicePath \"\"" Jan 21 11:21:42 crc kubenswrapper[4824]: I0121 11:21:42.655319 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q8qjw\" (UniqueName: \"kubernetes.io/projected/c585ef2f-3e29-4b81-918f-11b075006ca0-kube-api-access-q8qjw\") on node \"crc\" DevicePath \"\"" Jan 21 11:21:42 crc kubenswrapper[4824]: I0121 11:21:42.655388 4824 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/c585ef2f-3e29-4b81-918f-11b075006ca0-console-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 21 11:21:43 crc kubenswrapper[4824]: I0121 11:21:43.154739 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-nwlwd_c585ef2f-3e29-4b81-918f-11b075006ca0/console/0.log" Jan 21 11:21:43 crc kubenswrapper[4824]: I0121 11:21:43.154790 4824 generic.go:334] "Generic (PLEG): container finished" podID="c585ef2f-3e29-4b81-918f-11b075006ca0" containerID="3a9966ed610a4f72371cd05219170eb5e70c5b477a36cba97f1fa20640f0ccd7" exitCode=2 Jan 21 11:21:43 crc kubenswrapper[4824]: I0121 11:21:43.154819 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-nwlwd" event={"ID":"c585ef2f-3e29-4b81-918f-11b075006ca0","Type":"ContainerDied","Data":"3a9966ed610a4f72371cd05219170eb5e70c5b477a36cba97f1fa20640f0ccd7"} Jan 21 11:21:43 crc kubenswrapper[4824]: I0121 11:21:43.154844 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-nwlwd" event={"ID":"c585ef2f-3e29-4b81-918f-11b075006ca0","Type":"ContainerDied","Data":"0b16adc0798e89f22e1034c180951bc349e7c3b376c2bea592b22d2474b2bc5e"} Jan 21 11:21:43 crc kubenswrapper[4824]: I0121 11:21:43.154859 4824 scope.go:117] "RemoveContainer" containerID="3a9966ed610a4f72371cd05219170eb5e70c5b477a36cba97f1fa20640f0ccd7" Jan 21 11:21:43 crc kubenswrapper[4824]: I0121 11:21:43.154865 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-nwlwd" Jan 21 11:21:43 crc kubenswrapper[4824]: I0121 11:21:43.170339 4824 scope.go:117] "RemoveContainer" containerID="3a9966ed610a4f72371cd05219170eb5e70c5b477a36cba97f1fa20640f0ccd7" Jan 21 11:21:43 crc kubenswrapper[4824]: E0121 11:21:43.171519 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3a9966ed610a4f72371cd05219170eb5e70c5b477a36cba97f1fa20640f0ccd7\": container with ID starting with 3a9966ed610a4f72371cd05219170eb5e70c5b477a36cba97f1fa20640f0ccd7 not found: ID does not exist" containerID="3a9966ed610a4f72371cd05219170eb5e70c5b477a36cba97f1fa20640f0ccd7" Jan 21 11:21:43 crc kubenswrapper[4824]: I0121 11:21:43.171572 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3a9966ed610a4f72371cd05219170eb5e70c5b477a36cba97f1fa20640f0ccd7"} err="failed to get container status \"3a9966ed610a4f72371cd05219170eb5e70c5b477a36cba97f1fa20640f0ccd7\": rpc error: code = NotFound desc = could not find container \"3a9966ed610a4f72371cd05219170eb5e70c5b477a36cba97f1fa20640f0ccd7\": container with ID starting with 3a9966ed610a4f72371cd05219170eb5e70c5b477a36cba97f1fa20640f0ccd7 not found: ID does not exist" Jan 21 11:21:43 crc kubenswrapper[4824]: I0121 11:21:43.181752 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-nwlwd"] Jan 21 11:21:43 crc kubenswrapper[4824]: I0121 11:21:43.184411 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-nwlwd"] Jan 21 11:21:44 crc kubenswrapper[4824]: I0121 11:21:44.053299 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c585ef2f-3e29-4b81-918f-11b075006ca0" path="/var/lib/kubelet/pods/c585ef2f-3e29-4b81-918f-11b075006ca0/volumes" Jan 21 11:21:49 crc kubenswrapper[4824]: I0121 11:21:49.470615 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-5c58bcfcff-69lct"] Jan 21 11:21:49 crc kubenswrapper[4824]: E0121 11:21:49.471142 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d9005d95-8246-435d-aaea-5c60d58175c6" containerName="util" Jan 21 11:21:49 crc kubenswrapper[4824]: I0121 11:21:49.471155 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="d9005d95-8246-435d-aaea-5c60d58175c6" containerName="util" Jan 21 11:21:49 crc kubenswrapper[4824]: E0121 11:21:49.471163 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c585ef2f-3e29-4b81-918f-11b075006ca0" containerName="console" Jan 21 11:21:49 crc kubenswrapper[4824]: I0121 11:21:49.471168 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="c585ef2f-3e29-4b81-918f-11b075006ca0" containerName="console" Jan 21 11:21:49 crc kubenswrapper[4824]: E0121 11:21:49.471189 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d9005d95-8246-435d-aaea-5c60d58175c6" containerName="extract" Jan 21 11:21:49 crc kubenswrapper[4824]: I0121 11:21:49.471194 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="d9005d95-8246-435d-aaea-5c60d58175c6" containerName="extract" Jan 21 11:21:49 crc kubenswrapper[4824]: E0121 11:21:49.471200 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d9005d95-8246-435d-aaea-5c60d58175c6" containerName="pull" Jan 21 11:21:49 crc kubenswrapper[4824]: I0121 11:21:49.471206 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="d9005d95-8246-435d-aaea-5c60d58175c6" containerName="pull" Jan 21 11:21:49 crc kubenswrapper[4824]: I0121 11:21:49.471315 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="d9005d95-8246-435d-aaea-5c60d58175c6" containerName="extract" Jan 21 11:21:49 crc kubenswrapper[4824]: I0121 11:21:49.471324 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="c585ef2f-3e29-4b81-918f-11b075006ca0" containerName="console" Jan 21 11:21:49 crc kubenswrapper[4824]: I0121 11:21:49.471666 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-5c58bcfcff-69lct" Jan 21 11:21:49 crc kubenswrapper[4824]: I0121 11:21:49.474827 4824 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Jan 21 11:21:49 crc kubenswrapper[4824]: I0121 11:21:49.475013 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Jan 21 11:21:49 crc kubenswrapper[4824]: I0121 11:21:49.475180 4824 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-485hc" Jan 21 11:21:49 crc kubenswrapper[4824]: I0121 11:21:49.475026 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Jan 21 11:21:49 crc kubenswrapper[4824]: I0121 11:21:49.475057 4824 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Jan 21 11:21:49 crc kubenswrapper[4824]: I0121 11:21:49.480554 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-5c58bcfcff-69lct"] Jan 21 11:21:49 crc kubenswrapper[4824]: I0121 11:21:49.527061 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/714d2131-afc8-4828-840f-e2f0ce727d5d-webhook-cert\") pod \"metallb-operator-controller-manager-5c58bcfcff-69lct\" (UID: \"714d2131-afc8-4828-840f-e2f0ce727d5d\") " pod="metallb-system/metallb-operator-controller-manager-5c58bcfcff-69lct" Jan 21 11:21:49 crc kubenswrapper[4824]: I0121 11:21:49.527117 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v5lrt\" (UniqueName: \"kubernetes.io/projected/714d2131-afc8-4828-840f-e2f0ce727d5d-kube-api-access-v5lrt\") pod \"metallb-operator-controller-manager-5c58bcfcff-69lct\" (UID: \"714d2131-afc8-4828-840f-e2f0ce727d5d\") " pod="metallb-system/metallb-operator-controller-manager-5c58bcfcff-69lct" Jan 21 11:21:49 crc kubenswrapper[4824]: I0121 11:21:49.527147 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/714d2131-afc8-4828-840f-e2f0ce727d5d-apiservice-cert\") pod \"metallb-operator-controller-manager-5c58bcfcff-69lct\" (UID: \"714d2131-afc8-4828-840f-e2f0ce727d5d\") " pod="metallb-system/metallb-operator-controller-manager-5c58bcfcff-69lct" Jan 21 11:21:49 crc kubenswrapper[4824]: I0121 11:21:49.627843 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/714d2131-afc8-4828-840f-e2f0ce727d5d-webhook-cert\") pod \"metallb-operator-controller-manager-5c58bcfcff-69lct\" (UID: \"714d2131-afc8-4828-840f-e2f0ce727d5d\") " pod="metallb-system/metallb-operator-controller-manager-5c58bcfcff-69lct" Jan 21 11:21:49 crc kubenswrapper[4824]: I0121 11:21:49.627884 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v5lrt\" (UniqueName: \"kubernetes.io/projected/714d2131-afc8-4828-840f-e2f0ce727d5d-kube-api-access-v5lrt\") pod \"metallb-operator-controller-manager-5c58bcfcff-69lct\" (UID: \"714d2131-afc8-4828-840f-e2f0ce727d5d\") " pod="metallb-system/metallb-operator-controller-manager-5c58bcfcff-69lct" Jan 21 11:21:49 crc kubenswrapper[4824]: I0121 11:21:49.627908 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/714d2131-afc8-4828-840f-e2f0ce727d5d-apiservice-cert\") pod \"metallb-operator-controller-manager-5c58bcfcff-69lct\" (UID: \"714d2131-afc8-4828-840f-e2f0ce727d5d\") " pod="metallb-system/metallb-operator-controller-manager-5c58bcfcff-69lct" Jan 21 11:21:49 crc kubenswrapper[4824]: I0121 11:21:49.633183 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/714d2131-afc8-4828-840f-e2f0ce727d5d-apiservice-cert\") pod \"metallb-operator-controller-manager-5c58bcfcff-69lct\" (UID: \"714d2131-afc8-4828-840f-e2f0ce727d5d\") " pod="metallb-system/metallb-operator-controller-manager-5c58bcfcff-69lct" Jan 21 11:21:49 crc kubenswrapper[4824]: I0121 11:21:49.637521 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/714d2131-afc8-4828-840f-e2f0ce727d5d-webhook-cert\") pod \"metallb-operator-controller-manager-5c58bcfcff-69lct\" (UID: \"714d2131-afc8-4828-840f-e2f0ce727d5d\") " pod="metallb-system/metallb-operator-controller-manager-5c58bcfcff-69lct" Jan 21 11:21:49 crc kubenswrapper[4824]: I0121 11:21:49.641873 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v5lrt\" (UniqueName: \"kubernetes.io/projected/714d2131-afc8-4828-840f-e2f0ce727d5d-kube-api-access-v5lrt\") pod \"metallb-operator-controller-manager-5c58bcfcff-69lct\" (UID: \"714d2131-afc8-4828-840f-e2f0ce727d5d\") " pod="metallb-system/metallb-operator-controller-manager-5c58bcfcff-69lct" Jan 21 11:21:49 crc kubenswrapper[4824]: I0121 11:21:49.692250 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-68bc9bfcb6-hzmp5"] Jan 21 11:21:49 crc kubenswrapper[4824]: I0121 11:21:49.692864 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-68bc9bfcb6-hzmp5" Jan 21 11:21:49 crc kubenswrapper[4824]: I0121 11:21:49.694293 4824 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Jan 21 11:21:49 crc kubenswrapper[4824]: I0121 11:21:49.694315 4824 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Jan 21 11:21:49 crc kubenswrapper[4824]: I0121 11:21:49.694508 4824 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-j4wrz" Jan 21 11:21:49 crc kubenswrapper[4824]: I0121 11:21:49.704214 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-68bc9bfcb6-hzmp5"] Jan 21 11:21:49 crc kubenswrapper[4824]: I0121 11:21:49.729260 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/81fed5ab-ec3d-46f0-8998-199393bffba0-apiservice-cert\") pod \"metallb-operator-webhook-server-68bc9bfcb6-hzmp5\" (UID: \"81fed5ab-ec3d-46f0-8998-199393bffba0\") " pod="metallb-system/metallb-operator-webhook-server-68bc9bfcb6-hzmp5" Jan 21 11:21:49 crc kubenswrapper[4824]: I0121 11:21:49.729314 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/81fed5ab-ec3d-46f0-8998-199393bffba0-webhook-cert\") pod \"metallb-operator-webhook-server-68bc9bfcb6-hzmp5\" (UID: \"81fed5ab-ec3d-46f0-8998-199393bffba0\") " pod="metallb-system/metallb-operator-webhook-server-68bc9bfcb6-hzmp5" Jan 21 11:21:49 crc kubenswrapper[4824]: I0121 11:21:49.729455 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wzb82\" (UniqueName: \"kubernetes.io/projected/81fed5ab-ec3d-46f0-8998-199393bffba0-kube-api-access-wzb82\") pod \"metallb-operator-webhook-server-68bc9bfcb6-hzmp5\" (UID: \"81fed5ab-ec3d-46f0-8998-199393bffba0\") " pod="metallb-system/metallb-operator-webhook-server-68bc9bfcb6-hzmp5" Jan 21 11:21:49 crc kubenswrapper[4824]: I0121 11:21:49.785946 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-5c58bcfcff-69lct" Jan 21 11:21:49 crc kubenswrapper[4824]: I0121 11:21:49.830398 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/81fed5ab-ec3d-46f0-8998-199393bffba0-apiservice-cert\") pod \"metallb-operator-webhook-server-68bc9bfcb6-hzmp5\" (UID: \"81fed5ab-ec3d-46f0-8998-199393bffba0\") " pod="metallb-system/metallb-operator-webhook-server-68bc9bfcb6-hzmp5" Jan 21 11:21:49 crc kubenswrapper[4824]: I0121 11:21:49.830648 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/81fed5ab-ec3d-46f0-8998-199393bffba0-webhook-cert\") pod \"metallb-operator-webhook-server-68bc9bfcb6-hzmp5\" (UID: \"81fed5ab-ec3d-46f0-8998-199393bffba0\") " pod="metallb-system/metallb-operator-webhook-server-68bc9bfcb6-hzmp5" Jan 21 11:21:49 crc kubenswrapper[4824]: I0121 11:21:49.830692 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wzb82\" (UniqueName: \"kubernetes.io/projected/81fed5ab-ec3d-46f0-8998-199393bffba0-kube-api-access-wzb82\") pod \"metallb-operator-webhook-server-68bc9bfcb6-hzmp5\" (UID: \"81fed5ab-ec3d-46f0-8998-199393bffba0\") " pod="metallb-system/metallb-operator-webhook-server-68bc9bfcb6-hzmp5" Jan 21 11:21:49 crc kubenswrapper[4824]: I0121 11:21:49.837560 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/81fed5ab-ec3d-46f0-8998-199393bffba0-webhook-cert\") pod \"metallb-operator-webhook-server-68bc9bfcb6-hzmp5\" (UID: \"81fed5ab-ec3d-46f0-8998-199393bffba0\") " pod="metallb-system/metallb-operator-webhook-server-68bc9bfcb6-hzmp5" Jan 21 11:21:49 crc kubenswrapper[4824]: I0121 11:21:49.847783 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/81fed5ab-ec3d-46f0-8998-199393bffba0-apiservice-cert\") pod \"metallb-operator-webhook-server-68bc9bfcb6-hzmp5\" (UID: \"81fed5ab-ec3d-46f0-8998-199393bffba0\") " pod="metallb-system/metallb-operator-webhook-server-68bc9bfcb6-hzmp5" Jan 21 11:21:49 crc kubenswrapper[4824]: I0121 11:21:49.848090 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wzb82\" (UniqueName: \"kubernetes.io/projected/81fed5ab-ec3d-46f0-8998-199393bffba0-kube-api-access-wzb82\") pod \"metallb-operator-webhook-server-68bc9bfcb6-hzmp5\" (UID: \"81fed5ab-ec3d-46f0-8998-199393bffba0\") " pod="metallb-system/metallb-operator-webhook-server-68bc9bfcb6-hzmp5" Jan 21 11:21:50 crc kubenswrapper[4824]: I0121 11:21:50.010168 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-68bc9bfcb6-hzmp5" Jan 21 11:21:50 crc kubenswrapper[4824]: I0121 11:21:50.158721 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-68bc9bfcb6-hzmp5"] Jan 21 11:21:50 crc kubenswrapper[4824]: W0121 11:21:50.163688 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod81fed5ab_ec3d_46f0_8998_199393bffba0.slice/crio-e13fd4ef1e0aa49f805e3011bcab6e7503dfccd9f20ebd3374b90352ef667dd8 WatchSource:0}: Error finding container e13fd4ef1e0aa49f805e3011bcab6e7503dfccd9f20ebd3374b90352ef667dd8: Status 404 returned error can't find the container with id e13fd4ef1e0aa49f805e3011bcab6e7503dfccd9f20ebd3374b90352ef667dd8 Jan 21 11:21:50 crc kubenswrapper[4824]: I0121 11:21:50.176708 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-5c58bcfcff-69lct"] Jan 21 11:21:50 crc kubenswrapper[4824]: W0121 11:21:50.184179 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod714d2131_afc8_4828_840f_e2f0ce727d5d.slice/crio-02da55d074eb6b9f3efed95f9be4a2b12e2ee23fabe98f8b1db6fddea084e66f WatchSource:0}: Error finding container 02da55d074eb6b9f3efed95f9be4a2b12e2ee23fabe98f8b1db6fddea084e66f: Status 404 returned error can't find the container with id 02da55d074eb6b9f3efed95f9be4a2b12e2ee23fabe98f8b1db6fddea084e66f Jan 21 11:21:50 crc kubenswrapper[4824]: I0121 11:21:50.191892 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-68bc9bfcb6-hzmp5" event={"ID":"81fed5ab-ec3d-46f0-8998-199393bffba0","Type":"ContainerStarted","Data":"e13fd4ef1e0aa49f805e3011bcab6e7503dfccd9f20ebd3374b90352ef667dd8"} Jan 21 11:21:51 crc kubenswrapper[4824]: I0121 11:21:51.197458 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-5c58bcfcff-69lct" event={"ID":"714d2131-afc8-4828-840f-e2f0ce727d5d","Type":"ContainerStarted","Data":"02da55d074eb6b9f3efed95f9be4a2b12e2ee23fabe98f8b1db6fddea084e66f"} Jan 21 11:21:54 crc kubenswrapper[4824]: I0121 11:21:54.215051 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-5c58bcfcff-69lct" event={"ID":"714d2131-afc8-4828-840f-e2f0ce727d5d","Type":"ContainerStarted","Data":"946b563520e00ffbab367928e26c2cfb8af9b079810c7e19ceaa7ee88b6e11e5"} Jan 21 11:21:54 crc kubenswrapper[4824]: I0121 11:21:54.215391 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-5c58bcfcff-69lct" Jan 21 11:21:54 crc kubenswrapper[4824]: I0121 11:21:54.217172 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-68bc9bfcb6-hzmp5" event={"ID":"81fed5ab-ec3d-46f0-8998-199393bffba0","Type":"ContainerStarted","Data":"9f9097f18e9d26887efe54591063136fb92d1431b65090d400d4bb491708021c"} Jan 21 11:21:54 crc kubenswrapper[4824]: I0121 11:21:54.217322 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-68bc9bfcb6-hzmp5" Jan 21 11:21:54 crc kubenswrapper[4824]: I0121 11:21:54.229017 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-5c58bcfcff-69lct" podStartSLOduration=2.20324148 podStartE2EDuration="5.229005823s" podCreationTimestamp="2026-01-21 11:21:49 +0000 UTC" firstStartedPulling="2026-01-21 11:21:50.193665085 +0000 UTC m=+652.486694368" lastFinishedPulling="2026-01-21 11:21:53.219429418 +0000 UTC m=+655.512458711" observedRunningTime="2026-01-21 11:21:54.227360286 +0000 UTC m=+656.520389578" watchObservedRunningTime="2026-01-21 11:21:54.229005823 +0000 UTC m=+656.522035115" Jan 21 11:21:54 crc kubenswrapper[4824]: I0121 11:21:54.245212 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-68bc9bfcb6-hzmp5" podStartSLOduration=2.184176941 podStartE2EDuration="5.245196505s" podCreationTimestamp="2026-01-21 11:21:49 +0000 UTC" firstStartedPulling="2026-01-21 11:21:50.164992755 +0000 UTC m=+652.458022047" lastFinishedPulling="2026-01-21 11:21:53.226012319 +0000 UTC m=+655.519041611" observedRunningTime="2026-01-21 11:21:54.242161865 +0000 UTC m=+656.535191157" watchObservedRunningTime="2026-01-21 11:21:54.245196505 +0000 UTC m=+656.538225797" Jan 21 11:22:10 crc kubenswrapper[4824]: I0121 11:22:10.013690 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-68bc9bfcb6-hzmp5" Jan 21 11:22:29 crc kubenswrapper[4824]: I0121 11:22:29.788874 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-5c58bcfcff-69lct" Jan 21 11:22:30 crc kubenswrapper[4824]: I0121 11:22:30.341128 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-9q9wj"] Jan 21 11:22:30 crc kubenswrapper[4824]: I0121 11:22:30.342907 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-9q9wj" Jan 21 11:22:30 crc kubenswrapper[4824]: I0121 11:22:30.344481 4824 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-dgcz7" Jan 21 11:22:30 crc kubenswrapper[4824]: I0121 11:22:30.345044 4824 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Jan 21 11:22:30 crc kubenswrapper[4824]: I0121 11:22:30.345476 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Jan 21 11:22:30 crc kubenswrapper[4824]: I0121 11:22:30.349122 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-7df86c4f6c-fdblb"] Jan 21 11:22:30 crc kubenswrapper[4824]: I0121 11:22:30.350065 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-fdblb" Jan 21 11:22:30 crc kubenswrapper[4824]: I0121 11:22:30.351671 4824 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Jan 21 11:22:30 crc kubenswrapper[4824]: I0121 11:22:30.354840 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7df86c4f6c-fdblb"] Jan 21 11:22:30 crc kubenswrapper[4824]: I0121 11:22:30.374979 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d71f993c-b2c7-47af-884f-4cbe6dcf3c1a-cert\") pod \"frr-k8s-webhook-server-7df86c4f6c-fdblb\" (UID: \"d71f993c-b2c7-47af-884f-4cbe6dcf3c1a\") " pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-fdblb" Jan 21 11:22:30 crc kubenswrapper[4824]: I0121 11:22:30.375100 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5jxvb\" (UniqueName: \"kubernetes.io/projected/d71f993c-b2c7-47af-884f-4cbe6dcf3c1a-kube-api-access-5jxvb\") pod \"frr-k8s-webhook-server-7df86c4f6c-fdblb\" (UID: \"d71f993c-b2c7-47af-884f-4cbe6dcf3c1a\") " pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-fdblb" Jan 21 11:22:30 crc kubenswrapper[4824]: I0121 11:22:30.417791 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-8tsb9"] Jan 21 11:22:30 crc kubenswrapper[4824]: I0121 11:22:30.418990 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-8tsb9" Jan 21 11:22:30 crc kubenswrapper[4824]: I0121 11:22:30.422807 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-6968d8fdc4-bbv7p"] Jan 21 11:22:30 crc kubenswrapper[4824]: I0121 11:22:30.423281 4824 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Jan 21 11:22:30 crc kubenswrapper[4824]: I0121 11:22:30.423443 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Jan 21 11:22:30 crc kubenswrapper[4824]: I0121 11:22:30.423357 4824 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-2cwg7" Jan 21 11:22:30 crc kubenswrapper[4824]: I0121 11:22:30.423758 4824 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Jan 21 11:22:30 crc kubenswrapper[4824]: I0121 11:22:30.423876 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-6968d8fdc4-bbv7p" Jan 21 11:22:30 crc kubenswrapper[4824]: I0121 11:22:30.426728 4824 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Jan 21 11:22:30 crc kubenswrapper[4824]: I0121 11:22:30.430889 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-6968d8fdc4-bbv7p"] Jan 21 11:22:30 crc kubenswrapper[4824]: I0121 11:22:30.476697 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d30695b8-6f6a-48f0-88ed-9181484634b8-metrics-certs\") pod \"controller-6968d8fdc4-bbv7p\" (UID: \"d30695b8-6f6a-48f0-88ed-9181484634b8\") " pod="metallb-system/controller-6968d8fdc4-bbv7p" Jan 21 11:22:30 crc kubenswrapper[4824]: I0121 11:22:30.476739 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n6bd9\" (UniqueName: \"kubernetes.io/projected/baf5db51-d3ac-4b97-9c5f-afc598d59ab7-kube-api-access-n6bd9\") pod \"speaker-8tsb9\" (UID: \"baf5db51-d3ac-4b97-9c5f-afc598d59ab7\") " pod="metallb-system/speaker-8tsb9" Jan 21 11:22:30 crc kubenswrapper[4824]: I0121 11:22:30.476760 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/baf5db51-d3ac-4b97-9c5f-afc598d59ab7-metallb-excludel2\") pod \"speaker-8tsb9\" (UID: \"baf5db51-d3ac-4b97-9c5f-afc598d59ab7\") " pod="metallb-system/speaker-8tsb9" Jan 21 11:22:30 crc kubenswrapper[4824]: I0121 11:22:30.476777 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/baf5db51-d3ac-4b97-9c5f-afc598d59ab7-memberlist\") pod \"speaker-8tsb9\" (UID: \"baf5db51-d3ac-4b97-9c5f-afc598d59ab7\") " pod="metallb-system/speaker-8tsb9" Jan 21 11:22:30 crc kubenswrapper[4824]: I0121 11:22:30.476796 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/e38576f0-c423-497a-a30d-60c2e8edd93a-reloader\") pod \"frr-k8s-9q9wj\" (UID: \"e38576f0-c423-497a-a30d-60c2e8edd93a\") " pod="metallb-system/frr-k8s-9q9wj" Jan 21 11:22:30 crc kubenswrapper[4824]: I0121 11:22:30.476810 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d30695b8-6f6a-48f0-88ed-9181484634b8-cert\") pod \"controller-6968d8fdc4-bbv7p\" (UID: \"d30695b8-6f6a-48f0-88ed-9181484634b8\") " pod="metallb-system/controller-6968d8fdc4-bbv7p" Jan 21 11:22:30 crc kubenswrapper[4824]: I0121 11:22:30.476831 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5jxvb\" (UniqueName: \"kubernetes.io/projected/d71f993c-b2c7-47af-884f-4cbe6dcf3c1a-kube-api-access-5jxvb\") pod \"frr-k8s-webhook-server-7df86c4f6c-fdblb\" (UID: \"d71f993c-b2c7-47af-884f-4cbe6dcf3c1a\") " pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-fdblb" Jan 21 11:22:30 crc kubenswrapper[4824]: I0121 11:22:30.476855 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s8zz4\" (UniqueName: \"kubernetes.io/projected/d30695b8-6f6a-48f0-88ed-9181484634b8-kube-api-access-s8zz4\") pod \"controller-6968d8fdc4-bbv7p\" (UID: \"d30695b8-6f6a-48f0-88ed-9181484634b8\") " pod="metallb-system/controller-6968d8fdc4-bbv7p" Jan 21 11:22:30 crc kubenswrapper[4824]: I0121 11:22:30.476876 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e38576f0-c423-497a-a30d-60c2e8edd93a-metrics-certs\") pod \"frr-k8s-9q9wj\" (UID: \"e38576f0-c423-497a-a30d-60c2e8edd93a\") " pod="metallb-system/frr-k8s-9q9wj" Jan 21 11:22:30 crc kubenswrapper[4824]: I0121 11:22:30.476982 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/e38576f0-c423-497a-a30d-60c2e8edd93a-frr-startup\") pod \"frr-k8s-9q9wj\" (UID: \"e38576f0-c423-497a-a30d-60c2e8edd93a\") " pod="metallb-system/frr-k8s-9q9wj" Jan 21 11:22:30 crc kubenswrapper[4824]: I0121 11:22:30.477037 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/e38576f0-c423-497a-a30d-60c2e8edd93a-metrics\") pod \"frr-k8s-9q9wj\" (UID: \"e38576f0-c423-497a-a30d-60c2e8edd93a\") " pod="metallb-system/frr-k8s-9q9wj" Jan 21 11:22:30 crc kubenswrapper[4824]: I0121 11:22:30.477056 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/baf5db51-d3ac-4b97-9c5f-afc598d59ab7-metrics-certs\") pod \"speaker-8tsb9\" (UID: \"baf5db51-d3ac-4b97-9c5f-afc598d59ab7\") " pod="metallb-system/speaker-8tsb9" Jan 21 11:22:30 crc kubenswrapper[4824]: I0121 11:22:30.477073 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xf6fn\" (UniqueName: \"kubernetes.io/projected/e38576f0-c423-497a-a30d-60c2e8edd93a-kube-api-access-xf6fn\") pod \"frr-k8s-9q9wj\" (UID: \"e38576f0-c423-497a-a30d-60c2e8edd93a\") " pod="metallb-system/frr-k8s-9q9wj" Jan 21 11:22:30 crc kubenswrapper[4824]: I0121 11:22:30.477088 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/e38576f0-c423-497a-a30d-60c2e8edd93a-frr-conf\") pod \"frr-k8s-9q9wj\" (UID: \"e38576f0-c423-497a-a30d-60c2e8edd93a\") " pod="metallb-system/frr-k8s-9q9wj" Jan 21 11:22:30 crc kubenswrapper[4824]: I0121 11:22:30.477105 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d71f993c-b2c7-47af-884f-4cbe6dcf3c1a-cert\") pod \"frr-k8s-webhook-server-7df86c4f6c-fdblb\" (UID: \"d71f993c-b2c7-47af-884f-4cbe6dcf3c1a\") " pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-fdblb" Jan 21 11:22:30 crc kubenswrapper[4824]: I0121 11:22:30.477128 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/e38576f0-c423-497a-a30d-60c2e8edd93a-frr-sockets\") pod \"frr-k8s-9q9wj\" (UID: \"e38576f0-c423-497a-a30d-60c2e8edd93a\") " pod="metallb-system/frr-k8s-9q9wj" Jan 21 11:22:30 crc kubenswrapper[4824]: I0121 11:22:30.492709 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d71f993c-b2c7-47af-884f-4cbe6dcf3c1a-cert\") pod \"frr-k8s-webhook-server-7df86c4f6c-fdblb\" (UID: \"d71f993c-b2c7-47af-884f-4cbe6dcf3c1a\") " pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-fdblb" Jan 21 11:22:30 crc kubenswrapper[4824]: I0121 11:22:30.492814 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5jxvb\" (UniqueName: \"kubernetes.io/projected/d71f993c-b2c7-47af-884f-4cbe6dcf3c1a-kube-api-access-5jxvb\") pod \"frr-k8s-webhook-server-7df86c4f6c-fdblb\" (UID: \"d71f993c-b2c7-47af-884f-4cbe6dcf3c1a\") " pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-fdblb" Jan 21 11:22:30 crc kubenswrapper[4824]: I0121 11:22:30.578290 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/e38576f0-c423-497a-a30d-60c2e8edd93a-metrics\") pod \"frr-k8s-9q9wj\" (UID: \"e38576f0-c423-497a-a30d-60c2e8edd93a\") " pod="metallb-system/frr-k8s-9q9wj" Jan 21 11:22:30 crc kubenswrapper[4824]: I0121 11:22:30.578334 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/baf5db51-d3ac-4b97-9c5f-afc598d59ab7-metrics-certs\") pod \"speaker-8tsb9\" (UID: \"baf5db51-d3ac-4b97-9c5f-afc598d59ab7\") " pod="metallb-system/speaker-8tsb9" Jan 21 11:22:30 crc kubenswrapper[4824]: I0121 11:22:30.578358 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xf6fn\" (UniqueName: \"kubernetes.io/projected/e38576f0-c423-497a-a30d-60c2e8edd93a-kube-api-access-xf6fn\") pod \"frr-k8s-9q9wj\" (UID: \"e38576f0-c423-497a-a30d-60c2e8edd93a\") " pod="metallb-system/frr-k8s-9q9wj" Jan 21 11:22:30 crc kubenswrapper[4824]: I0121 11:22:30.578376 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/e38576f0-c423-497a-a30d-60c2e8edd93a-frr-conf\") pod \"frr-k8s-9q9wj\" (UID: \"e38576f0-c423-497a-a30d-60c2e8edd93a\") " pod="metallb-system/frr-k8s-9q9wj" Jan 21 11:22:30 crc kubenswrapper[4824]: I0121 11:22:30.578404 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/e38576f0-c423-497a-a30d-60c2e8edd93a-frr-sockets\") pod \"frr-k8s-9q9wj\" (UID: \"e38576f0-c423-497a-a30d-60c2e8edd93a\") " pod="metallb-system/frr-k8s-9q9wj" Jan 21 11:22:30 crc kubenswrapper[4824]: I0121 11:22:30.578441 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d30695b8-6f6a-48f0-88ed-9181484634b8-metrics-certs\") pod \"controller-6968d8fdc4-bbv7p\" (UID: \"d30695b8-6f6a-48f0-88ed-9181484634b8\") " pod="metallb-system/controller-6968d8fdc4-bbv7p" Jan 21 11:22:30 crc kubenswrapper[4824]: I0121 11:22:30.578459 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n6bd9\" (UniqueName: \"kubernetes.io/projected/baf5db51-d3ac-4b97-9c5f-afc598d59ab7-kube-api-access-n6bd9\") pod \"speaker-8tsb9\" (UID: \"baf5db51-d3ac-4b97-9c5f-afc598d59ab7\") " pod="metallb-system/speaker-8tsb9" Jan 21 11:22:30 crc kubenswrapper[4824]: I0121 11:22:30.578475 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/baf5db51-d3ac-4b97-9c5f-afc598d59ab7-metallb-excludel2\") pod \"speaker-8tsb9\" (UID: \"baf5db51-d3ac-4b97-9c5f-afc598d59ab7\") " pod="metallb-system/speaker-8tsb9" Jan 21 11:22:30 crc kubenswrapper[4824]: I0121 11:22:30.578488 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/baf5db51-d3ac-4b97-9c5f-afc598d59ab7-memberlist\") pod \"speaker-8tsb9\" (UID: \"baf5db51-d3ac-4b97-9c5f-afc598d59ab7\") " pod="metallb-system/speaker-8tsb9" Jan 21 11:22:30 crc kubenswrapper[4824]: I0121 11:22:30.578506 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/e38576f0-c423-497a-a30d-60c2e8edd93a-reloader\") pod \"frr-k8s-9q9wj\" (UID: \"e38576f0-c423-497a-a30d-60c2e8edd93a\") " pod="metallb-system/frr-k8s-9q9wj" Jan 21 11:22:30 crc kubenswrapper[4824]: I0121 11:22:30.578520 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d30695b8-6f6a-48f0-88ed-9181484634b8-cert\") pod \"controller-6968d8fdc4-bbv7p\" (UID: \"d30695b8-6f6a-48f0-88ed-9181484634b8\") " pod="metallb-system/controller-6968d8fdc4-bbv7p" Jan 21 11:22:30 crc kubenswrapper[4824]: I0121 11:22:30.578542 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s8zz4\" (UniqueName: \"kubernetes.io/projected/d30695b8-6f6a-48f0-88ed-9181484634b8-kube-api-access-s8zz4\") pod \"controller-6968d8fdc4-bbv7p\" (UID: \"d30695b8-6f6a-48f0-88ed-9181484634b8\") " pod="metallb-system/controller-6968d8fdc4-bbv7p" Jan 21 11:22:30 crc kubenswrapper[4824]: I0121 11:22:30.578560 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e38576f0-c423-497a-a30d-60c2e8edd93a-metrics-certs\") pod \"frr-k8s-9q9wj\" (UID: \"e38576f0-c423-497a-a30d-60c2e8edd93a\") " pod="metallb-system/frr-k8s-9q9wj" Jan 21 11:22:30 crc kubenswrapper[4824]: I0121 11:22:30.578600 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/e38576f0-c423-497a-a30d-60c2e8edd93a-frr-startup\") pod \"frr-k8s-9q9wj\" (UID: \"e38576f0-c423-497a-a30d-60c2e8edd93a\") " pod="metallb-system/frr-k8s-9q9wj" Jan 21 11:22:30 crc kubenswrapper[4824]: I0121 11:22:30.579420 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/e38576f0-c423-497a-a30d-60c2e8edd93a-frr-startup\") pod \"frr-k8s-9q9wj\" (UID: \"e38576f0-c423-497a-a30d-60c2e8edd93a\") " pod="metallb-system/frr-k8s-9q9wj" Jan 21 11:22:30 crc kubenswrapper[4824]: I0121 11:22:30.579641 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/e38576f0-c423-497a-a30d-60c2e8edd93a-metrics\") pod \"frr-k8s-9q9wj\" (UID: \"e38576f0-c423-497a-a30d-60c2e8edd93a\") " pod="metallb-system/frr-k8s-9q9wj" Jan 21 11:22:30 crc kubenswrapper[4824]: E0121 11:22:30.579726 4824 secret.go:188] Couldn't get secret metallb-system/speaker-certs-secret: secret "speaker-certs-secret" not found Jan 21 11:22:30 crc kubenswrapper[4824]: E0121 11:22:30.579769 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/baf5db51-d3ac-4b97-9c5f-afc598d59ab7-metrics-certs podName:baf5db51-d3ac-4b97-9c5f-afc598d59ab7 nodeName:}" failed. No retries permitted until 2026-01-21 11:22:31.079754946 +0000 UTC m=+693.372784238 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/baf5db51-d3ac-4b97-9c5f-afc598d59ab7-metrics-certs") pod "speaker-8tsb9" (UID: "baf5db51-d3ac-4b97-9c5f-afc598d59ab7") : secret "speaker-certs-secret" not found Jan 21 11:22:30 crc kubenswrapper[4824]: I0121 11:22:30.580283 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/e38576f0-c423-497a-a30d-60c2e8edd93a-frr-conf\") pod \"frr-k8s-9q9wj\" (UID: \"e38576f0-c423-497a-a30d-60c2e8edd93a\") " pod="metallb-system/frr-k8s-9q9wj" Jan 21 11:22:30 crc kubenswrapper[4824]: I0121 11:22:30.580451 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/e38576f0-c423-497a-a30d-60c2e8edd93a-frr-sockets\") pod \"frr-k8s-9q9wj\" (UID: \"e38576f0-c423-497a-a30d-60c2e8edd93a\") " pod="metallb-system/frr-k8s-9q9wj" Jan 21 11:22:30 crc kubenswrapper[4824]: I0121 11:22:30.580984 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/e38576f0-c423-497a-a30d-60c2e8edd93a-reloader\") pod \"frr-k8s-9q9wj\" (UID: \"e38576f0-c423-497a-a30d-60c2e8edd93a\") " pod="metallb-system/frr-k8s-9q9wj" Jan 21 11:22:30 crc kubenswrapper[4824]: E0121 11:22:30.580983 4824 secret.go:188] Couldn't get secret metallb-system/frr-k8s-certs-secret: secret "frr-k8s-certs-secret" not found Jan 21 11:22:30 crc kubenswrapper[4824]: E0121 11:22:30.581097 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e38576f0-c423-497a-a30d-60c2e8edd93a-metrics-certs podName:e38576f0-c423-497a-a30d-60c2e8edd93a nodeName:}" failed. No retries permitted until 2026-01-21 11:22:31.081086511 +0000 UTC m=+693.374115802 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/e38576f0-c423-497a-a30d-60c2e8edd93a-metrics-certs") pod "frr-k8s-9q9wj" (UID: "e38576f0-c423-497a-a30d-60c2e8edd93a") : secret "frr-k8s-certs-secret" not found Jan 21 11:22:30 crc kubenswrapper[4824]: E0121 11:22:30.581049 4824 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Jan 21 11:22:30 crc kubenswrapper[4824]: E0121 11:22:30.581128 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/baf5db51-d3ac-4b97-9c5f-afc598d59ab7-memberlist podName:baf5db51-d3ac-4b97-9c5f-afc598d59ab7 nodeName:}" failed. No retries permitted until 2026-01-21 11:22:31.081122678 +0000 UTC m=+693.374151970 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/baf5db51-d3ac-4b97-9c5f-afc598d59ab7-memberlist") pod "speaker-8tsb9" (UID: "baf5db51-d3ac-4b97-9c5f-afc598d59ab7") : secret "metallb-memberlist" not found Jan 21 11:22:30 crc kubenswrapper[4824]: I0121 11:22:30.581647 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/baf5db51-d3ac-4b97-9c5f-afc598d59ab7-metallb-excludel2\") pod \"speaker-8tsb9\" (UID: \"baf5db51-d3ac-4b97-9c5f-afc598d59ab7\") " pod="metallb-system/speaker-8tsb9" Jan 21 11:22:30 crc kubenswrapper[4824]: I0121 11:22:30.582539 4824 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Jan 21 11:22:30 crc kubenswrapper[4824]: I0121 11:22:30.583853 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d30695b8-6f6a-48f0-88ed-9181484634b8-metrics-certs\") pod \"controller-6968d8fdc4-bbv7p\" (UID: \"d30695b8-6f6a-48f0-88ed-9181484634b8\") " pod="metallb-system/controller-6968d8fdc4-bbv7p" Jan 21 11:22:30 crc kubenswrapper[4824]: I0121 11:22:30.593273 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xf6fn\" (UniqueName: \"kubernetes.io/projected/e38576f0-c423-497a-a30d-60c2e8edd93a-kube-api-access-xf6fn\") pod \"frr-k8s-9q9wj\" (UID: \"e38576f0-c423-497a-a30d-60c2e8edd93a\") " pod="metallb-system/frr-k8s-9q9wj" Jan 21 11:22:30 crc kubenswrapper[4824]: I0121 11:22:30.593777 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d30695b8-6f6a-48f0-88ed-9181484634b8-cert\") pod \"controller-6968d8fdc4-bbv7p\" (UID: \"d30695b8-6f6a-48f0-88ed-9181484634b8\") " pod="metallb-system/controller-6968d8fdc4-bbv7p" Jan 21 11:22:30 crc kubenswrapper[4824]: I0121 11:22:30.595358 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s8zz4\" (UniqueName: \"kubernetes.io/projected/d30695b8-6f6a-48f0-88ed-9181484634b8-kube-api-access-s8zz4\") pod \"controller-6968d8fdc4-bbv7p\" (UID: \"d30695b8-6f6a-48f0-88ed-9181484634b8\") " pod="metallb-system/controller-6968d8fdc4-bbv7p" Jan 21 11:22:30 crc kubenswrapper[4824]: I0121 11:22:30.596274 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n6bd9\" (UniqueName: \"kubernetes.io/projected/baf5db51-d3ac-4b97-9c5f-afc598d59ab7-kube-api-access-n6bd9\") pod \"speaker-8tsb9\" (UID: \"baf5db51-d3ac-4b97-9c5f-afc598d59ab7\") " pod="metallb-system/speaker-8tsb9" Jan 21 11:22:30 crc kubenswrapper[4824]: I0121 11:22:30.678821 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-fdblb" Jan 21 11:22:30 crc kubenswrapper[4824]: I0121 11:22:30.741016 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-6968d8fdc4-bbv7p" Jan 21 11:22:31 crc kubenswrapper[4824]: I0121 11:22:31.083694 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/baf5db51-d3ac-4b97-9c5f-afc598d59ab7-memberlist\") pod \"speaker-8tsb9\" (UID: \"baf5db51-d3ac-4b97-9c5f-afc598d59ab7\") " pod="metallb-system/speaker-8tsb9" Jan 21 11:22:31 crc kubenswrapper[4824]: E0121 11:22:31.083776 4824 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Jan 21 11:22:31 crc kubenswrapper[4824]: I0121 11:22:31.083989 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e38576f0-c423-497a-a30d-60c2e8edd93a-metrics-certs\") pod \"frr-k8s-9q9wj\" (UID: \"e38576f0-c423-497a-a30d-60c2e8edd93a\") " pod="metallb-system/frr-k8s-9q9wj" Jan 21 11:22:31 crc kubenswrapper[4824]: E0121 11:22:31.084034 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/baf5db51-d3ac-4b97-9c5f-afc598d59ab7-memberlist podName:baf5db51-d3ac-4b97-9c5f-afc598d59ab7 nodeName:}" failed. No retries permitted until 2026-01-21 11:22:32.084015757 +0000 UTC m=+694.377045049 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/baf5db51-d3ac-4b97-9c5f-afc598d59ab7-memberlist") pod "speaker-8tsb9" (UID: "baf5db51-d3ac-4b97-9c5f-afc598d59ab7") : secret "metallb-memberlist" not found Jan 21 11:22:31 crc kubenswrapper[4824]: I0121 11:22:31.084054 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/baf5db51-d3ac-4b97-9c5f-afc598d59ab7-metrics-certs\") pod \"speaker-8tsb9\" (UID: \"baf5db51-d3ac-4b97-9c5f-afc598d59ab7\") " pod="metallb-system/speaker-8tsb9" Jan 21 11:22:31 crc kubenswrapper[4824]: I0121 11:22:31.088394 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7df86c4f6c-fdblb"] Jan 21 11:22:31 crc kubenswrapper[4824]: I0121 11:22:31.093186 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e38576f0-c423-497a-a30d-60c2e8edd93a-metrics-certs\") pod \"frr-k8s-9q9wj\" (UID: \"e38576f0-c423-497a-a30d-60c2e8edd93a\") " pod="metallb-system/frr-k8s-9q9wj" Jan 21 11:22:31 crc kubenswrapper[4824]: I0121 11:22:31.093218 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/baf5db51-d3ac-4b97-9c5f-afc598d59ab7-metrics-certs\") pod \"speaker-8tsb9\" (UID: \"baf5db51-d3ac-4b97-9c5f-afc598d59ab7\") " pod="metallb-system/speaker-8tsb9" Jan 21 11:22:31 crc kubenswrapper[4824]: I0121 11:22:31.152264 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-6968d8fdc4-bbv7p"] Jan 21 11:22:31 crc kubenswrapper[4824]: W0121 11:22:31.158567 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd30695b8_6f6a_48f0_88ed_9181484634b8.slice/crio-e91b7eab093938e8669c4c30f13b245b7708535eb6b924ed385a4505933e198e WatchSource:0}: Error finding container e91b7eab093938e8669c4c30f13b245b7708535eb6b924ed385a4505933e198e: Status 404 returned error can't find the container with id e91b7eab093938e8669c4c30f13b245b7708535eb6b924ed385a4505933e198e Jan 21 11:22:31 crc kubenswrapper[4824]: I0121 11:22:31.258224 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-9q9wj" Jan 21 11:22:31 crc kubenswrapper[4824]: I0121 11:22:31.381999 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-6968d8fdc4-bbv7p" event={"ID":"d30695b8-6f6a-48f0-88ed-9181484634b8","Type":"ContainerStarted","Data":"6941c121390bf3d300db662a46888f442a975e5dcb6064e1bbafeffed095a417"} Jan 21 11:22:31 crc kubenswrapper[4824]: I0121 11:22:31.382040 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-6968d8fdc4-bbv7p" event={"ID":"d30695b8-6f6a-48f0-88ed-9181484634b8","Type":"ContainerStarted","Data":"cf5b3ae27e69bc46a35b70866df64d613f991437dc86f83bc3a382de36b29dd6"} Jan 21 11:22:31 crc kubenswrapper[4824]: I0121 11:22:31.382053 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-6968d8fdc4-bbv7p" event={"ID":"d30695b8-6f6a-48f0-88ed-9181484634b8","Type":"ContainerStarted","Data":"e91b7eab093938e8669c4c30f13b245b7708535eb6b924ed385a4505933e198e"} Jan 21 11:22:31 crc kubenswrapper[4824]: I0121 11:22:31.382133 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-6968d8fdc4-bbv7p" Jan 21 11:22:31 crc kubenswrapper[4824]: I0121 11:22:31.383072 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-9q9wj" event={"ID":"e38576f0-c423-497a-a30d-60c2e8edd93a","Type":"ContainerStarted","Data":"12be331cc3262346a8fa63374a3426fa6b02cfc67db2a8b9421fb800f4d5536f"} Jan 21 11:22:31 crc kubenswrapper[4824]: I0121 11:22:31.383936 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-fdblb" event={"ID":"d71f993c-b2c7-47af-884f-4cbe6dcf3c1a","Type":"ContainerStarted","Data":"e421d48f3b6d0a95ac7a49d2adc2bf51886d1df12f496efea31ce2742524a525"} Jan 21 11:22:31 crc kubenswrapper[4824]: I0121 11:22:31.395342 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-6968d8fdc4-bbv7p" podStartSLOduration=1.395330277 podStartE2EDuration="1.395330277s" podCreationTimestamp="2026-01-21 11:22:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:22:31.392841518 +0000 UTC m=+693.685870810" watchObservedRunningTime="2026-01-21 11:22:31.395330277 +0000 UTC m=+693.688359569" Jan 21 11:22:32 crc kubenswrapper[4824]: I0121 11:22:32.093758 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/baf5db51-d3ac-4b97-9c5f-afc598d59ab7-memberlist\") pod \"speaker-8tsb9\" (UID: \"baf5db51-d3ac-4b97-9c5f-afc598d59ab7\") " pod="metallb-system/speaker-8tsb9" Jan 21 11:22:32 crc kubenswrapper[4824]: I0121 11:22:32.110842 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/baf5db51-d3ac-4b97-9c5f-afc598d59ab7-memberlist\") pod \"speaker-8tsb9\" (UID: \"baf5db51-d3ac-4b97-9c5f-afc598d59ab7\") " pod="metallb-system/speaker-8tsb9" Jan 21 11:22:32 crc kubenswrapper[4824]: I0121 11:22:32.235201 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-8tsb9" Jan 21 11:22:32 crc kubenswrapper[4824]: I0121 11:22:32.401764 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-8tsb9" event={"ID":"baf5db51-d3ac-4b97-9c5f-afc598d59ab7","Type":"ContainerStarted","Data":"00408b77679c420ad9a736839489c2fafe53e9a4cdd59abe8d5fd7728fd5286a"} Jan 21 11:22:33 crc kubenswrapper[4824]: I0121 11:22:33.411348 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-8tsb9" event={"ID":"baf5db51-d3ac-4b97-9c5f-afc598d59ab7","Type":"ContainerStarted","Data":"d8500e599fa9c0d231db01e19baeca992b240d084b6c51b01ea472922e43afa2"} Jan 21 11:22:33 crc kubenswrapper[4824]: I0121 11:22:33.411673 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-8tsb9" event={"ID":"baf5db51-d3ac-4b97-9c5f-afc598d59ab7","Type":"ContainerStarted","Data":"cbffadfc9555b5b289cbd0ae205d85f95a914b260f3dfee6102d45e8d3ee91c3"} Jan 21 11:22:33 crc kubenswrapper[4824]: I0121 11:22:33.411690 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-8tsb9" Jan 21 11:22:33 crc kubenswrapper[4824]: I0121 11:22:33.426988 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-8tsb9" podStartSLOduration=3.426972368 podStartE2EDuration="3.426972368s" podCreationTimestamp="2026-01-21 11:22:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:22:33.424863806 +0000 UTC m=+695.717893099" watchObservedRunningTime="2026-01-21 11:22:33.426972368 +0000 UTC m=+695.720001660" Jan 21 11:22:37 crc kubenswrapper[4824]: I0121 11:22:37.439842 4824 generic.go:334] "Generic (PLEG): container finished" podID="e38576f0-c423-497a-a30d-60c2e8edd93a" containerID="0a95b49dd0f30bdcb9480450ec3e995683dcc1c5bef8adea27f3a85a5d236c1d" exitCode=0 Jan 21 11:22:37 crc kubenswrapper[4824]: I0121 11:22:37.439987 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-9q9wj" event={"ID":"e38576f0-c423-497a-a30d-60c2e8edd93a","Type":"ContainerDied","Data":"0a95b49dd0f30bdcb9480450ec3e995683dcc1c5bef8adea27f3a85a5d236c1d"} Jan 21 11:22:37 crc kubenswrapper[4824]: I0121 11:22:37.443813 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-fdblb" event={"ID":"d71f993c-b2c7-47af-884f-4cbe6dcf3c1a","Type":"ContainerStarted","Data":"344ca599880fd08d1fb66b1e7118337f7fdf955b7f9b5d0d142f0d2aee4d5487"} Jan 21 11:22:37 crc kubenswrapper[4824]: I0121 11:22:37.443988 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-fdblb" Jan 21 11:22:37 crc kubenswrapper[4824]: I0121 11:22:37.470942 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-fdblb" podStartSLOduration=1.917495246 podStartE2EDuration="7.470924236s" podCreationTimestamp="2026-01-21 11:22:30 +0000 UTC" firstStartedPulling="2026-01-21 11:22:31.095394333 +0000 UTC m=+693.388423614" lastFinishedPulling="2026-01-21 11:22:36.648823312 +0000 UTC m=+698.941852604" observedRunningTime="2026-01-21 11:22:37.470544929 +0000 UTC m=+699.763574221" watchObservedRunningTime="2026-01-21 11:22:37.470924236 +0000 UTC m=+699.763953528" Jan 21 11:22:38 crc kubenswrapper[4824]: I0121 11:22:38.449558 4824 generic.go:334] "Generic (PLEG): container finished" podID="e38576f0-c423-497a-a30d-60c2e8edd93a" containerID="02f6cd335f7b59d9369a0b173b948cb2181ed1b07a21195f07787f05898d6c8f" exitCode=0 Jan 21 11:22:38 crc kubenswrapper[4824]: I0121 11:22:38.449602 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-9q9wj" event={"ID":"e38576f0-c423-497a-a30d-60c2e8edd93a","Type":"ContainerDied","Data":"02f6cd335f7b59d9369a0b173b948cb2181ed1b07a21195f07787f05898d6c8f"} Jan 21 11:22:39 crc kubenswrapper[4824]: I0121 11:22:39.456453 4824 generic.go:334] "Generic (PLEG): container finished" podID="e38576f0-c423-497a-a30d-60c2e8edd93a" containerID="8e5df092e2d0209c1dd968162c6bb2a85f1a4e9f92796f3723722c15b87ebeb3" exitCode=0 Jan 21 11:22:39 crc kubenswrapper[4824]: I0121 11:22:39.456498 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-9q9wj" event={"ID":"e38576f0-c423-497a-a30d-60c2e8edd93a","Type":"ContainerDied","Data":"8e5df092e2d0209c1dd968162c6bb2a85f1a4e9f92796f3723722c15b87ebeb3"} Jan 21 11:22:40 crc kubenswrapper[4824]: I0121 11:22:40.465183 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-9q9wj" event={"ID":"e38576f0-c423-497a-a30d-60c2e8edd93a","Type":"ContainerStarted","Data":"dceffcf088cbc712e691f9a86c11e2343a15757b03b28b0eecd852c705dd889c"} Jan 21 11:22:40 crc kubenswrapper[4824]: I0121 11:22:40.465473 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-9q9wj" event={"ID":"e38576f0-c423-497a-a30d-60c2e8edd93a","Type":"ContainerStarted","Data":"61a0f960fe6522f77e86258dc114d4a59ad0c0f8644703aea226afefae7a28df"} Jan 21 11:22:40 crc kubenswrapper[4824]: I0121 11:22:40.465484 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-9q9wj" event={"ID":"e38576f0-c423-497a-a30d-60c2e8edd93a","Type":"ContainerStarted","Data":"9ba551b63fd42c126b2373cfc52ce8a5958522b4cd4bd230ea86d5c8ac0dcfc0"} Jan 21 11:22:40 crc kubenswrapper[4824]: I0121 11:22:40.465498 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-9q9wj" Jan 21 11:22:40 crc kubenswrapper[4824]: I0121 11:22:40.465507 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-9q9wj" event={"ID":"e38576f0-c423-497a-a30d-60c2e8edd93a","Type":"ContainerStarted","Data":"276db60e826eda049265f4eeb3b091ef6c8beb0d2e747421f737f99e465f7838"} Jan 21 11:22:40 crc kubenswrapper[4824]: I0121 11:22:40.465515 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-9q9wj" event={"ID":"e38576f0-c423-497a-a30d-60c2e8edd93a","Type":"ContainerStarted","Data":"077e5c87fa526fc502e122e29f0cb39c220b47fe3bb8ad2146bd1f65ac9b5519"} Jan 21 11:22:40 crc kubenswrapper[4824]: I0121 11:22:40.465524 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-9q9wj" event={"ID":"e38576f0-c423-497a-a30d-60c2e8edd93a","Type":"ContainerStarted","Data":"eac894334aecaf0c6194362c91ce71a737232cd8eacc273c58adeca41bc80a23"} Jan 21 11:22:40 crc kubenswrapper[4824]: I0121 11:22:40.481827 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-9q9wj" podStartSLOduration=5.160693349 podStartE2EDuration="10.481815485s" podCreationTimestamp="2026-01-21 11:22:30 +0000 UTC" firstStartedPulling="2026-01-21 11:22:31.333815642 +0000 UTC m=+693.626844924" lastFinishedPulling="2026-01-21 11:22:36.654937767 +0000 UTC m=+698.947967060" observedRunningTime="2026-01-21 11:22:40.481201355 +0000 UTC m=+702.774230647" watchObservedRunningTime="2026-01-21 11:22:40.481815485 +0000 UTC m=+702.774844777" Jan 21 11:22:41 crc kubenswrapper[4824]: I0121 11:22:41.259054 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-9q9wj" Jan 21 11:22:41 crc kubenswrapper[4824]: I0121 11:22:41.288052 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-9q9wj" Jan 21 11:22:42 crc kubenswrapper[4824]: I0121 11:22:42.238549 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-8tsb9" Jan 21 11:22:44 crc kubenswrapper[4824]: I0121 11:22:44.326361 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-2xq8c"] Jan 21 11:22:44 crc kubenswrapper[4824]: I0121 11:22:44.327008 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-2xq8c" Jan 21 11:22:44 crc kubenswrapper[4824]: I0121 11:22:44.329120 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-dlsfl" Jan 21 11:22:44 crc kubenswrapper[4824]: I0121 11:22:44.330282 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Jan 21 11:22:44 crc kubenswrapper[4824]: I0121 11:22:44.333767 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-2xq8c"] Jan 21 11:22:44 crc kubenswrapper[4824]: I0121 11:22:44.334574 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Jan 21 11:22:44 crc kubenswrapper[4824]: I0121 11:22:44.460138 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p6bsj\" (UniqueName: \"kubernetes.io/projected/3d4b4d56-a2a1-4fba-8d1d-449a222f8b9d-kube-api-access-p6bsj\") pod \"openstack-operator-index-2xq8c\" (UID: \"3d4b4d56-a2a1-4fba-8d1d-449a222f8b9d\") " pod="openstack-operators/openstack-operator-index-2xq8c" Jan 21 11:22:44 crc kubenswrapper[4824]: I0121 11:22:44.561494 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p6bsj\" (UniqueName: \"kubernetes.io/projected/3d4b4d56-a2a1-4fba-8d1d-449a222f8b9d-kube-api-access-p6bsj\") pod \"openstack-operator-index-2xq8c\" (UID: \"3d4b4d56-a2a1-4fba-8d1d-449a222f8b9d\") " pod="openstack-operators/openstack-operator-index-2xq8c" Jan 21 11:22:44 crc kubenswrapper[4824]: I0121 11:22:44.577836 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p6bsj\" (UniqueName: \"kubernetes.io/projected/3d4b4d56-a2a1-4fba-8d1d-449a222f8b9d-kube-api-access-p6bsj\") pod \"openstack-operator-index-2xq8c\" (UID: \"3d4b4d56-a2a1-4fba-8d1d-449a222f8b9d\") " pod="openstack-operators/openstack-operator-index-2xq8c" Jan 21 11:22:44 crc kubenswrapper[4824]: I0121 11:22:44.642528 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-2xq8c" Jan 21 11:22:44 crc kubenswrapper[4824]: I0121 11:22:44.978536 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-2xq8c"] Jan 21 11:22:45 crc kubenswrapper[4824]: I0121 11:22:45.489273 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-2xq8c" event={"ID":"3d4b4d56-a2a1-4fba-8d1d-449a222f8b9d","Type":"ContainerStarted","Data":"65770a4cd5ba23f67f4e7e3da13403242a193d33807de294fba0702d80bb3eec"} Jan 21 11:22:46 crc kubenswrapper[4824]: I0121 11:22:46.065275 4824 patch_prober.go:28] interesting pod/machine-config-daemon-zqjg9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 21 11:22:46 crc kubenswrapper[4824]: I0121 11:22:46.065316 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 21 11:22:47 crc kubenswrapper[4824]: I0121 11:22:47.499316 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-2xq8c" event={"ID":"3d4b4d56-a2a1-4fba-8d1d-449a222f8b9d","Type":"ContainerStarted","Data":"127db9144957d0346ae9b36f73dd36f254481811058d81a21f490169cf556266"} Jan 21 11:22:47 crc kubenswrapper[4824]: I0121 11:22:47.511918 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-2xq8c" podStartSLOduration=1.48636971 podStartE2EDuration="3.511903518s" podCreationTimestamp="2026-01-21 11:22:44 +0000 UTC" firstStartedPulling="2026-01-21 11:22:44.986417396 +0000 UTC m=+707.279446687" lastFinishedPulling="2026-01-21 11:22:47.011951202 +0000 UTC m=+709.304980495" observedRunningTime="2026-01-21 11:22:47.509012018 +0000 UTC m=+709.802041310" watchObservedRunningTime="2026-01-21 11:22:47.511903518 +0000 UTC m=+709.804932810" Jan 21 11:22:47 crc kubenswrapper[4824]: I0121 11:22:47.709671 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-2xq8c"] Jan 21 11:22:48 crc kubenswrapper[4824]: I0121 11:22:48.313047 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-zggxm"] Jan 21 11:22:48 crc kubenswrapper[4824]: I0121 11:22:48.313640 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-zggxm" Jan 21 11:22:48 crc kubenswrapper[4824]: I0121 11:22:48.317857 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-zggxm"] Jan 21 11:22:48 crc kubenswrapper[4824]: I0121 11:22:48.407994 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dfdz9\" (UniqueName: \"kubernetes.io/projected/78eb1906-d9e6-49f7-a512-ed3cc0a07da7-kube-api-access-dfdz9\") pod \"openstack-operator-index-zggxm\" (UID: \"78eb1906-d9e6-49f7-a512-ed3cc0a07da7\") " pod="openstack-operators/openstack-operator-index-zggxm" Jan 21 11:22:48 crc kubenswrapper[4824]: I0121 11:22:48.509349 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dfdz9\" (UniqueName: \"kubernetes.io/projected/78eb1906-d9e6-49f7-a512-ed3cc0a07da7-kube-api-access-dfdz9\") pod \"openstack-operator-index-zggxm\" (UID: \"78eb1906-d9e6-49f7-a512-ed3cc0a07da7\") " pod="openstack-operators/openstack-operator-index-zggxm" Jan 21 11:22:48 crc kubenswrapper[4824]: I0121 11:22:48.523699 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dfdz9\" (UniqueName: \"kubernetes.io/projected/78eb1906-d9e6-49f7-a512-ed3cc0a07da7-kube-api-access-dfdz9\") pod \"openstack-operator-index-zggxm\" (UID: \"78eb1906-d9e6-49f7-a512-ed3cc0a07da7\") " pod="openstack-operators/openstack-operator-index-zggxm" Jan 21 11:22:48 crc kubenswrapper[4824]: I0121 11:22:48.627122 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-zggxm" Jan 21 11:22:48 crc kubenswrapper[4824]: I0121 11:22:48.955828 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-zggxm"] Jan 21 11:22:48 crc kubenswrapper[4824]: W0121 11:22:48.958414 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod78eb1906_d9e6_49f7_a512_ed3cc0a07da7.slice/crio-817c80c60202a27b25381cfe512730c140beb8ad74df7f8cc7375b67d651a6ea WatchSource:0}: Error finding container 817c80c60202a27b25381cfe512730c140beb8ad74df7f8cc7375b67d651a6ea: Status 404 returned error can't find the container with id 817c80c60202a27b25381cfe512730c140beb8ad74df7f8cc7375b67d651a6ea Jan 21 11:22:49 crc kubenswrapper[4824]: I0121 11:22:49.509165 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-zggxm" event={"ID":"78eb1906-d9e6-49f7-a512-ed3cc0a07da7","Type":"ContainerStarted","Data":"88794dbd990089c8d2da8afa3daa81d9dd2222adf0cc7353a709e9d78811072d"} Jan 21 11:22:49 crc kubenswrapper[4824]: I0121 11:22:49.509216 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-zggxm" event={"ID":"78eb1906-d9e6-49f7-a512-ed3cc0a07da7","Type":"ContainerStarted","Data":"817c80c60202a27b25381cfe512730c140beb8ad74df7f8cc7375b67d651a6ea"} Jan 21 11:22:49 crc kubenswrapper[4824]: I0121 11:22:49.509255 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-index-2xq8c" podUID="3d4b4d56-a2a1-4fba-8d1d-449a222f8b9d" containerName="registry-server" containerID="cri-o://127db9144957d0346ae9b36f73dd36f254481811058d81a21f490169cf556266" gracePeriod=2 Jan 21 11:22:49 crc kubenswrapper[4824]: I0121 11:22:49.525754 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-zggxm" podStartSLOduration=1.142007744 podStartE2EDuration="1.525737543s" podCreationTimestamp="2026-01-21 11:22:48 +0000 UTC" firstStartedPulling="2026-01-21 11:22:48.961313879 +0000 UTC m=+711.254343171" lastFinishedPulling="2026-01-21 11:22:49.345043678 +0000 UTC m=+711.638072970" observedRunningTime="2026-01-21 11:22:49.522488348 +0000 UTC m=+711.815517640" watchObservedRunningTime="2026-01-21 11:22:49.525737543 +0000 UTC m=+711.818766835" Jan 21 11:22:49 crc kubenswrapper[4824]: I0121 11:22:49.800908 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-2xq8c" Jan 21 11:22:49 crc kubenswrapper[4824]: I0121 11:22:49.925879 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p6bsj\" (UniqueName: \"kubernetes.io/projected/3d4b4d56-a2a1-4fba-8d1d-449a222f8b9d-kube-api-access-p6bsj\") pod \"3d4b4d56-a2a1-4fba-8d1d-449a222f8b9d\" (UID: \"3d4b4d56-a2a1-4fba-8d1d-449a222f8b9d\") " Jan 21 11:22:49 crc kubenswrapper[4824]: I0121 11:22:49.930033 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3d4b4d56-a2a1-4fba-8d1d-449a222f8b9d-kube-api-access-p6bsj" (OuterVolumeSpecName: "kube-api-access-p6bsj") pod "3d4b4d56-a2a1-4fba-8d1d-449a222f8b9d" (UID: "3d4b4d56-a2a1-4fba-8d1d-449a222f8b9d"). InnerVolumeSpecName "kube-api-access-p6bsj". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:22:50 crc kubenswrapper[4824]: I0121 11:22:50.026876 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p6bsj\" (UniqueName: \"kubernetes.io/projected/3d4b4d56-a2a1-4fba-8d1d-449a222f8b9d-kube-api-access-p6bsj\") on node \"crc\" DevicePath \"\"" Jan 21 11:22:50 crc kubenswrapper[4824]: I0121 11:22:50.514250 4824 generic.go:334] "Generic (PLEG): container finished" podID="3d4b4d56-a2a1-4fba-8d1d-449a222f8b9d" containerID="127db9144957d0346ae9b36f73dd36f254481811058d81a21f490169cf556266" exitCode=0 Jan 21 11:22:50 crc kubenswrapper[4824]: I0121 11:22:50.514286 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-2xq8c" event={"ID":"3d4b4d56-a2a1-4fba-8d1d-449a222f8b9d","Type":"ContainerDied","Data":"127db9144957d0346ae9b36f73dd36f254481811058d81a21f490169cf556266"} Jan 21 11:22:50 crc kubenswrapper[4824]: I0121 11:22:50.514318 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-2xq8c" Jan 21 11:22:50 crc kubenswrapper[4824]: I0121 11:22:50.514337 4824 scope.go:117] "RemoveContainer" containerID="127db9144957d0346ae9b36f73dd36f254481811058d81a21f490169cf556266" Jan 21 11:22:50 crc kubenswrapper[4824]: I0121 11:22:50.514325 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-2xq8c" event={"ID":"3d4b4d56-a2a1-4fba-8d1d-449a222f8b9d","Type":"ContainerDied","Data":"65770a4cd5ba23f67f4e7e3da13403242a193d33807de294fba0702d80bb3eec"} Jan 21 11:22:50 crc kubenswrapper[4824]: I0121 11:22:50.527942 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-2xq8c"] Jan 21 11:22:50 crc kubenswrapper[4824]: I0121 11:22:50.529163 4824 scope.go:117] "RemoveContainer" containerID="127db9144957d0346ae9b36f73dd36f254481811058d81a21f490169cf556266" Jan 21 11:22:50 crc kubenswrapper[4824]: E0121 11:22:50.529466 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"127db9144957d0346ae9b36f73dd36f254481811058d81a21f490169cf556266\": container with ID starting with 127db9144957d0346ae9b36f73dd36f254481811058d81a21f490169cf556266 not found: ID does not exist" containerID="127db9144957d0346ae9b36f73dd36f254481811058d81a21f490169cf556266" Jan 21 11:22:50 crc kubenswrapper[4824]: I0121 11:22:50.529512 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"127db9144957d0346ae9b36f73dd36f254481811058d81a21f490169cf556266"} err="failed to get container status \"127db9144957d0346ae9b36f73dd36f254481811058d81a21f490169cf556266\": rpc error: code = NotFound desc = could not find container \"127db9144957d0346ae9b36f73dd36f254481811058d81a21f490169cf556266\": container with ID starting with 127db9144957d0346ae9b36f73dd36f254481811058d81a21f490169cf556266 not found: ID does not exist" Jan 21 11:22:50 crc kubenswrapper[4824]: I0121 11:22:50.531178 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-index-2xq8c"] Jan 21 11:22:50 crc kubenswrapper[4824]: I0121 11:22:50.683807 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-fdblb" Jan 21 11:22:50 crc kubenswrapper[4824]: I0121 11:22:50.744380 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-6968d8fdc4-bbv7p" Jan 21 11:22:51 crc kubenswrapper[4824]: I0121 11:22:51.260733 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-9q9wj" Jan 21 11:22:52 crc kubenswrapper[4824]: I0121 11:22:52.053953 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3d4b4d56-a2a1-4fba-8d1d-449a222f8b9d" path="/var/lib/kubelet/pods/3d4b4d56-a2a1-4fba-8d1d-449a222f8b9d/volumes" Jan 21 11:22:58 crc kubenswrapper[4824]: I0121 11:22:58.628202 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-zggxm" Jan 21 11:22:58 crc kubenswrapper[4824]: I0121 11:22:58.628387 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-zggxm" Jan 21 11:22:58 crc kubenswrapper[4824]: I0121 11:22:58.647797 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-zggxm" Jan 21 11:22:59 crc kubenswrapper[4824]: I0121 11:22:59.572301 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-zggxm" Jan 21 11:23:07 crc kubenswrapper[4824]: I0121 11:23:07.436062 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/db4b9aacd6f51b5aff2d66474158631eb1e58fb4d1358e39c8aaaca1bdn7cxr"] Jan 21 11:23:07 crc kubenswrapper[4824]: E0121 11:23:07.436602 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d4b4d56-a2a1-4fba-8d1d-449a222f8b9d" containerName="registry-server" Jan 21 11:23:07 crc kubenswrapper[4824]: I0121 11:23:07.436614 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d4b4d56-a2a1-4fba-8d1d-449a222f8b9d" containerName="registry-server" Jan 21 11:23:07 crc kubenswrapper[4824]: I0121 11:23:07.436745 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="3d4b4d56-a2a1-4fba-8d1d-449a222f8b9d" containerName="registry-server" Jan 21 11:23:07 crc kubenswrapper[4824]: I0121 11:23:07.437436 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/db4b9aacd6f51b5aff2d66474158631eb1e58fb4d1358e39c8aaaca1bdn7cxr" Jan 21 11:23:07 crc kubenswrapper[4824]: I0121 11:23:07.439219 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-gw7z7" Jan 21 11:23:07 crc kubenswrapper[4824]: I0121 11:23:07.447703 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/db4b9aacd6f51b5aff2d66474158631eb1e58fb4d1358e39c8aaaca1bdn7cxr"] Jan 21 11:23:07 crc kubenswrapper[4824]: I0121 11:23:07.532449 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/72ab26cb-149e-4825-9587-3acfa66a368f-bundle\") pod \"db4b9aacd6f51b5aff2d66474158631eb1e58fb4d1358e39c8aaaca1bdn7cxr\" (UID: \"72ab26cb-149e-4825-9587-3acfa66a368f\") " pod="openstack-operators/db4b9aacd6f51b5aff2d66474158631eb1e58fb4d1358e39c8aaaca1bdn7cxr" Jan 21 11:23:07 crc kubenswrapper[4824]: I0121 11:23:07.532554 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-clhll\" (UniqueName: \"kubernetes.io/projected/72ab26cb-149e-4825-9587-3acfa66a368f-kube-api-access-clhll\") pod \"db4b9aacd6f51b5aff2d66474158631eb1e58fb4d1358e39c8aaaca1bdn7cxr\" (UID: \"72ab26cb-149e-4825-9587-3acfa66a368f\") " pod="openstack-operators/db4b9aacd6f51b5aff2d66474158631eb1e58fb4d1358e39c8aaaca1bdn7cxr" Jan 21 11:23:07 crc kubenswrapper[4824]: I0121 11:23:07.532588 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/72ab26cb-149e-4825-9587-3acfa66a368f-util\") pod \"db4b9aacd6f51b5aff2d66474158631eb1e58fb4d1358e39c8aaaca1bdn7cxr\" (UID: \"72ab26cb-149e-4825-9587-3acfa66a368f\") " pod="openstack-operators/db4b9aacd6f51b5aff2d66474158631eb1e58fb4d1358e39c8aaaca1bdn7cxr" Jan 21 11:23:07 crc kubenswrapper[4824]: I0121 11:23:07.633464 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/72ab26cb-149e-4825-9587-3acfa66a368f-bundle\") pod \"db4b9aacd6f51b5aff2d66474158631eb1e58fb4d1358e39c8aaaca1bdn7cxr\" (UID: \"72ab26cb-149e-4825-9587-3acfa66a368f\") " pod="openstack-operators/db4b9aacd6f51b5aff2d66474158631eb1e58fb4d1358e39c8aaaca1bdn7cxr" Jan 21 11:23:07 crc kubenswrapper[4824]: I0121 11:23:07.633555 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-clhll\" (UniqueName: \"kubernetes.io/projected/72ab26cb-149e-4825-9587-3acfa66a368f-kube-api-access-clhll\") pod \"db4b9aacd6f51b5aff2d66474158631eb1e58fb4d1358e39c8aaaca1bdn7cxr\" (UID: \"72ab26cb-149e-4825-9587-3acfa66a368f\") " pod="openstack-operators/db4b9aacd6f51b5aff2d66474158631eb1e58fb4d1358e39c8aaaca1bdn7cxr" Jan 21 11:23:07 crc kubenswrapper[4824]: I0121 11:23:07.633579 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/72ab26cb-149e-4825-9587-3acfa66a368f-util\") pod \"db4b9aacd6f51b5aff2d66474158631eb1e58fb4d1358e39c8aaaca1bdn7cxr\" (UID: \"72ab26cb-149e-4825-9587-3acfa66a368f\") " pod="openstack-operators/db4b9aacd6f51b5aff2d66474158631eb1e58fb4d1358e39c8aaaca1bdn7cxr" Jan 21 11:23:07 crc kubenswrapper[4824]: I0121 11:23:07.633938 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/72ab26cb-149e-4825-9587-3acfa66a368f-util\") pod \"db4b9aacd6f51b5aff2d66474158631eb1e58fb4d1358e39c8aaaca1bdn7cxr\" (UID: \"72ab26cb-149e-4825-9587-3acfa66a368f\") " pod="openstack-operators/db4b9aacd6f51b5aff2d66474158631eb1e58fb4d1358e39c8aaaca1bdn7cxr" Jan 21 11:23:07 crc kubenswrapper[4824]: I0121 11:23:07.634050 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/72ab26cb-149e-4825-9587-3acfa66a368f-bundle\") pod \"db4b9aacd6f51b5aff2d66474158631eb1e58fb4d1358e39c8aaaca1bdn7cxr\" (UID: \"72ab26cb-149e-4825-9587-3acfa66a368f\") " pod="openstack-operators/db4b9aacd6f51b5aff2d66474158631eb1e58fb4d1358e39c8aaaca1bdn7cxr" Jan 21 11:23:07 crc kubenswrapper[4824]: I0121 11:23:07.649042 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-clhll\" (UniqueName: \"kubernetes.io/projected/72ab26cb-149e-4825-9587-3acfa66a368f-kube-api-access-clhll\") pod \"db4b9aacd6f51b5aff2d66474158631eb1e58fb4d1358e39c8aaaca1bdn7cxr\" (UID: \"72ab26cb-149e-4825-9587-3acfa66a368f\") " pod="openstack-operators/db4b9aacd6f51b5aff2d66474158631eb1e58fb4d1358e39c8aaaca1bdn7cxr" Jan 21 11:23:07 crc kubenswrapper[4824]: I0121 11:23:07.750571 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/db4b9aacd6f51b5aff2d66474158631eb1e58fb4d1358e39c8aaaca1bdn7cxr" Jan 21 11:23:08 crc kubenswrapper[4824]: I0121 11:23:08.089073 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/db4b9aacd6f51b5aff2d66474158631eb1e58fb4d1358e39c8aaaca1bdn7cxr"] Jan 21 11:23:08 crc kubenswrapper[4824]: W0121 11:23:08.093812 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod72ab26cb_149e_4825_9587_3acfa66a368f.slice/crio-3a388ea26d3419c18e1784c7a0bc1491ce1ea2643b305a49427345a44f845d83 WatchSource:0}: Error finding container 3a388ea26d3419c18e1784c7a0bc1491ce1ea2643b305a49427345a44f845d83: Status 404 returned error can't find the container with id 3a388ea26d3419c18e1784c7a0bc1491ce1ea2643b305a49427345a44f845d83 Jan 21 11:23:08 crc kubenswrapper[4824]: I0121 11:23:08.594389 4824 generic.go:334] "Generic (PLEG): container finished" podID="72ab26cb-149e-4825-9587-3acfa66a368f" containerID="fc23b7440efb654b0a6f0f66596889abdd1f9ef8f8fc616eb9d053707e783e4b" exitCode=0 Jan 21 11:23:08 crc kubenswrapper[4824]: I0121 11:23:08.594436 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/db4b9aacd6f51b5aff2d66474158631eb1e58fb4d1358e39c8aaaca1bdn7cxr" event={"ID":"72ab26cb-149e-4825-9587-3acfa66a368f","Type":"ContainerDied","Data":"fc23b7440efb654b0a6f0f66596889abdd1f9ef8f8fc616eb9d053707e783e4b"} Jan 21 11:23:08 crc kubenswrapper[4824]: I0121 11:23:08.594460 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/db4b9aacd6f51b5aff2d66474158631eb1e58fb4d1358e39c8aaaca1bdn7cxr" event={"ID":"72ab26cb-149e-4825-9587-3acfa66a368f","Type":"ContainerStarted","Data":"3a388ea26d3419c18e1784c7a0bc1491ce1ea2643b305a49427345a44f845d83"} Jan 21 11:23:09 crc kubenswrapper[4824]: I0121 11:23:09.599404 4824 generic.go:334] "Generic (PLEG): container finished" podID="72ab26cb-149e-4825-9587-3acfa66a368f" containerID="c380719aa42b9f83e4a40c333480e1664ba50c6d0ecce4d8e13a8c5281aac946" exitCode=0 Jan 21 11:23:09 crc kubenswrapper[4824]: I0121 11:23:09.599464 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/db4b9aacd6f51b5aff2d66474158631eb1e58fb4d1358e39c8aaaca1bdn7cxr" event={"ID":"72ab26cb-149e-4825-9587-3acfa66a368f","Type":"ContainerDied","Data":"c380719aa42b9f83e4a40c333480e1664ba50c6d0ecce4d8e13a8c5281aac946"} Jan 21 11:23:10 crc kubenswrapper[4824]: I0121 11:23:10.609859 4824 generic.go:334] "Generic (PLEG): container finished" podID="72ab26cb-149e-4825-9587-3acfa66a368f" containerID="7b18087c00cc8c90c5738a6f8ee462f79b3296105378b8976f0a6d0af38ecccf" exitCode=0 Jan 21 11:23:10 crc kubenswrapper[4824]: I0121 11:23:10.609901 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/db4b9aacd6f51b5aff2d66474158631eb1e58fb4d1358e39c8aaaca1bdn7cxr" event={"ID":"72ab26cb-149e-4825-9587-3acfa66a368f","Type":"ContainerDied","Data":"7b18087c00cc8c90c5738a6f8ee462f79b3296105378b8976f0a6d0af38ecccf"} Jan 21 11:23:11 crc kubenswrapper[4824]: I0121 11:23:11.797531 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/db4b9aacd6f51b5aff2d66474158631eb1e58fb4d1358e39c8aaaca1bdn7cxr" Jan 21 11:23:11 crc kubenswrapper[4824]: I0121 11:23:11.978247 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/72ab26cb-149e-4825-9587-3acfa66a368f-bundle\") pod \"72ab26cb-149e-4825-9587-3acfa66a368f\" (UID: \"72ab26cb-149e-4825-9587-3acfa66a368f\") " Jan 21 11:23:11 crc kubenswrapper[4824]: I0121 11:23:11.978351 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-clhll\" (UniqueName: \"kubernetes.io/projected/72ab26cb-149e-4825-9587-3acfa66a368f-kube-api-access-clhll\") pod \"72ab26cb-149e-4825-9587-3acfa66a368f\" (UID: \"72ab26cb-149e-4825-9587-3acfa66a368f\") " Jan 21 11:23:11 crc kubenswrapper[4824]: I0121 11:23:11.978417 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/72ab26cb-149e-4825-9587-3acfa66a368f-util\") pod \"72ab26cb-149e-4825-9587-3acfa66a368f\" (UID: \"72ab26cb-149e-4825-9587-3acfa66a368f\") " Jan 21 11:23:11 crc kubenswrapper[4824]: I0121 11:23:11.978844 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/72ab26cb-149e-4825-9587-3acfa66a368f-bundle" (OuterVolumeSpecName: "bundle") pod "72ab26cb-149e-4825-9587-3acfa66a368f" (UID: "72ab26cb-149e-4825-9587-3acfa66a368f"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:23:11 crc kubenswrapper[4824]: I0121 11:23:11.982938 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/72ab26cb-149e-4825-9587-3acfa66a368f-kube-api-access-clhll" (OuterVolumeSpecName: "kube-api-access-clhll") pod "72ab26cb-149e-4825-9587-3acfa66a368f" (UID: "72ab26cb-149e-4825-9587-3acfa66a368f"). InnerVolumeSpecName "kube-api-access-clhll". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:23:11 crc kubenswrapper[4824]: I0121 11:23:11.988254 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/72ab26cb-149e-4825-9587-3acfa66a368f-util" (OuterVolumeSpecName: "util") pod "72ab26cb-149e-4825-9587-3acfa66a368f" (UID: "72ab26cb-149e-4825-9587-3acfa66a368f"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:23:12 crc kubenswrapper[4824]: I0121 11:23:12.080931 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-clhll\" (UniqueName: \"kubernetes.io/projected/72ab26cb-149e-4825-9587-3acfa66a368f-kube-api-access-clhll\") on node \"crc\" DevicePath \"\"" Jan 21 11:23:12 crc kubenswrapper[4824]: I0121 11:23:12.081001 4824 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/72ab26cb-149e-4825-9587-3acfa66a368f-util\") on node \"crc\" DevicePath \"\"" Jan 21 11:23:12 crc kubenswrapper[4824]: I0121 11:23:12.081015 4824 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/72ab26cb-149e-4825-9587-3acfa66a368f-bundle\") on node \"crc\" DevicePath \"\"" Jan 21 11:23:12 crc kubenswrapper[4824]: I0121 11:23:12.620230 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/db4b9aacd6f51b5aff2d66474158631eb1e58fb4d1358e39c8aaaca1bdn7cxr" event={"ID":"72ab26cb-149e-4825-9587-3acfa66a368f","Type":"ContainerDied","Data":"3a388ea26d3419c18e1784c7a0bc1491ce1ea2643b305a49427345a44f845d83"} Jan 21 11:23:12 crc kubenswrapper[4824]: I0121 11:23:12.620275 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3a388ea26d3419c18e1784c7a0bc1491ce1ea2643b305a49427345a44f845d83" Jan 21 11:23:12 crc kubenswrapper[4824]: I0121 11:23:12.620337 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/db4b9aacd6f51b5aff2d66474158631eb1e58fb4d1358e39c8aaaca1bdn7cxr" Jan 21 11:23:14 crc kubenswrapper[4824]: I0121 11:23:14.830760 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-init-dbdbd6855-gkm5b"] Jan 21 11:23:14 crc kubenswrapper[4824]: E0121 11:23:14.831179 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="72ab26cb-149e-4825-9587-3acfa66a368f" containerName="pull" Jan 21 11:23:14 crc kubenswrapper[4824]: I0121 11:23:14.831191 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="72ab26cb-149e-4825-9587-3acfa66a368f" containerName="pull" Jan 21 11:23:14 crc kubenswrapper[4824]: E0121 11:23:14.831200 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="72ab26cb-149e-4825-9587-3acfa66a368f" containerName="util" Jan 21 11:23:14 crc kubenswrapper[4824]: I0121 11:23:14.831205 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="72ab26cb-149e-4825-9587-3acfa66a368f" containerName="util" Jan 21 11:23:14 crc kubenswrapper[4824]: E0121 11:23:14.831219 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="72ab26cb-149e-4825-9587-3acfa66a368f" containerName="extract" Jan 21 11:23:14 crc kubenswrapper[4824]: I0121 11:23:14.831226 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="72ab26cb-149e-4825-9587-3acfa66a368f" containerName="extract" Jan 21 11:23:14 crc kubenswrapper[4824]: I0121 11:23:14.831323 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="72ab26cb-149e-4825-9587-3acfa66a368f" containerName="extract" Jan 21 11:23:14 crc kubenswrapper[4824]: I0121 11:23:14.831682 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-init-dbdbd6855-gkm5b" Jan 21 11:23:14 crc kubenswrapper[4824]: I0121 11:23:14.834137 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-init-dockercfg-9hmt7" Jan 21 11:23:14 crc kubenswrapper[4824]: I0121 11:23:14.856112 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-init-dbdbd6855-gkm5b"] Jan 21 11:23:14 crc kubenswrapper[4824]: I0121 11:23:14.916437 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d2hv9\" (UniqueName: \"kubernetes.io/projected/35334861-8e16-4b16-8cf5-167a049836af-kube-api-access-d2hv9\") pod \"openstack-operator-controller-init-dbdbd6855-gkm5b\" (UID: \"35334861-8e16-4b16-8cf5-167a049836af\") " pod="openstack-operators/openstack-operator-controller-init-dbdbd6855-gkm5b" Jan 21 11:23:15 crc kubenswrapper[4824]: I0121 11:23:15.017433 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d2hv9\" (UniqueName: \"kubernetes.io/projected/35334861-8e16-4b16-8cf5-167a049836af-kube-api-access-d2hv9\") pod \"openstack-operator-controller-init-dbdbd6855-gkm5b\" (UID: \"35334861-8e16-4b16-8cf5-167a049836af\") " pod="openstack-operators/openstack-operator-controller-init-dbdbd6855-gkm5b" Jan 21 11:23:15 crc kubenswrapper[4824]: I0121 11:23:15.035179 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d2hv9\" (UniqueName: \"kubernetes.io/projected/35334861-8e16-4b16-8cf5-167a049836af-kube-api-access-d2hv9\") pod \"openstack-operator-controller-init-dbdbd6855-gkm5b\" (UID: \"35334861-8e16-4b16-8cf5-167a049836af\") " pod="openstack-operators/openstack-operator-controller-init-dbdbd6855-gkm5b" Jan 21 11:23:15 crc kubenswrapper[4824]: I0121 11:23:15.145202 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-init-dbdbd6855-gkm5b" Jan 21 11:23:15 crc kubenswrapper[4824]: I0121 11:23:15.511373 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-init-dbdbd6855-gkm5b"] Jan 21 11:23:15 crc kubenswrapper[4824]: I0121 11:23:15.635716 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-init-dbdbd6855-gkm5b" event={"ID":"35334861-8e16-4b16-8cf5-167a049836af","Type":"ContainerStarted","Data":"f53867a58b3b691c841b96163206155dfec28ba34222d55889b813a2952b7204"} Jan 21 11:23:16 crc kubenswrapper[4824]: I0121 11:23:16.065069 4824 patch_prober.go:28] interesting pod/machine-config-daemon-zqjg9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 21 11:23:16 crc kubenswrapper[4824]: I0121 11:23:16.065139 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 21 11:23:19 crc kubenswrapper[4824]: I0121 11:23:19.655701 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-init-dbdbd6855-gkm5b" event={"ID":"35334861-8e16-4b16-8cf5-167a049836af","Type":"ContainerStarted","Data":"b45e8370768dd0af119e29d056ccdfde86899f4c9d7e9d40974aab7097005c2e"} Jan 21 11:23:19 crc kubenswrapper[4824]: I0121 11:23:19.656119 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-init-dbdbd6855-gkm5b" Jan 21 11:23:19 crc kubenswrapper[4824]: I0121 11:23:19.675920 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-init-dbdbd6855-gkm5b" podStartSLOduration=2.071477087 podStartE2EDuration="5.675906248s" podCreationTimestamp="2026-01-21 11:23:14 +0000 UTC" firstStartedPulling="2026-01-21 11:23:15.529139576 +0000 UTC m=+737.822168868" lastFinishedPulling="2026-01-21 11:23:19.133568737 +0000 UTC m=+741.426598029" observedRunningTime="2026-01-21 11:23:19.67331292 +0000 UTC m=+741.966342211" watchObservedRunningTime="2026-01-21 11:23:19.675906248 +0000 UTC m=+741.968935540" Jan 21 11:23:24 crc kubenswrapper[4824]: I0121 11:23:24.014384 4824 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Jan 21 11:23:25 crc kubenswrapper[4824]: I0121 11:23:25.148276 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-init-dbdbd6855-gkm5b" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.547309 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7ddb5c749-bvptq"] Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.548329 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-7ddb5c749-bvptq" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.549846 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-6jjxv" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.554375 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-9b68f5989-24r54"] Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.555065 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-9b68f5989-24r54" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.558951 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7ddb5c749-bvptq"] Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.560576 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-2r2m7" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.567370 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-9b68f5989-24r54"] Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.576377 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-9f958b845-wtfj4"] Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.577820 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-9f958b845-wtfj4" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.579023 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-w7vhc" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.582673 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-c6994669c-hvq6s"] Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.583354 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-c6994669c-hvq6s" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.587854 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-9f958b845-wtfj4"] Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.588776 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-vt9bz" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.595981 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-c6994669c-hvq6s"] Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.608005 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-77d5c5b54f-mp5ml"] Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.608790 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-77d5c5b54f-mp5ml" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.614159 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-rpx84" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.619558 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-5fbc59b5b6-hgmf4"] Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.620265 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5fbc59b5b6-hgmf4" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.624482 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-x922p" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.628416 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5fbc59b5b6-hgmf4"] Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.634021 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-77d5c5b54f-mp5ml"] Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.636598 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h66xm\" (UniqueName: \"kubernetes.io/projected/a25733d4-a5f8-4c97-a897-8d96e637c253-kube-api-access-h66xm\") pod \"cinder-operator-controller-manager-9b68f5989-24r54\" (UID: \"a25733d4-a5f8-4c97-a897-8d96e637c253\") " pod="openstack-operators/cinder-operator-controller-manager-9b68f5989-24r54" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.636685 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vdzth\" (UniqueName: \"kubernetes.io/projected/8ed6bb2c-2770-4785-bf38-9b6fd0bd8472-kube-api-access-vdzth\") pod \"barbican-operator-controller-manager-7ddb5c749-bvptq\" (UID: \"8ed6bb2c-2770-4785-bf38-9b6fd0bd8472\") " pod="openstack-operators/barbican-operator-controller-manager-7ddb5c749-bvptq" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.636740 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2kwld\" (UniqueName: \"kubernetes.io/projected/14dd1f12-3943-42e4-be1b-c9e37e49d21b-kube-api-access-2kwld\") pod \"designate-operator-controller-manager-9f958b845-wtfj4\" (UID: \"14dd1f12-3943-42e4-be1b-c9e37e49d21b\") " pod="openstack-operators/designate-operator-controller-manager-9f958b845-wtfj4" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.636856 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dd2tc\" (UniqueName: \"kubernetes.io/projected/944c3123-9063-4731-9581-5b683741edc4-kube-api-access-dd2tc\") pod \"heat-operator-controller-manager-5fbc59b5b6-hgmf4\" (UID: \"944c3123-9063-4731-9581-5b683741edc4\") " pod="openstack-operators/heat-operator-controller-manager-5fbc59b5b6-hgmf4" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.636900 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f5dbk\" (UniqueName: \"kubernetes.io/projected/2463afb1-0271-4ab3-ad1a-9329ecaeec4d-kube-api-access-f5dbk\") pod \"horizon-operator-controller-manager-77d5c5b54f-mp5ml\" (UID: \"2463afb1-0271-4ab3-ad1a-9329ecaeec4d\") " pod="openstack-operators/horizon-operator-controller-manager-77d5c5b54f-mp5ml" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.636901 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-77c48c7859-dzn7w"] Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.637055 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h6kj7\" (UniqueName: \"kubernetes.io/projected/b7053a68-ce89-48b6-9b5d-1c6dac813a75-kube-api-access-h6kj7\") pod \"glance-operator-controller-manager-c6994669c-hvq6s\" (UID: \"b7053a68-ce89-48b6-9b5d-1c6dac813a75\") " pod="openstack-operators/glance-operator-controller-manager-c6994669c-hvq6s" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.637649 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-77c48c7859-dzn7w" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.640018 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-ldsjj" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.640131 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.646053 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-77c48c7859-dzn7w"] Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.651557 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-78757b4889-gkwjs"] Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.652215 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-78757b4889-gkwjs" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.656012 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-tz46k" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.660035 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-78757b4889-gkwjs"] Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.699765 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-767fdc4f47-5mhmd"] Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.700613 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-767fdc4f47-5mhmd" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.704333 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-8pwkx" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.713429 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-767fdc4f47-5mhmd"] Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.731761 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-864f6b75bf-c7vsf"] Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.732507 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-864f6b75bf-c7vsf" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.735565 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-864f6b75bf-c7vsf"] Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.739218 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-p2zqb" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.739550 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-c87fff755-d5hn9"] Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.740117 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-c87fff755-d5hn9" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.740813 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h6kj7\" (UniqueName: \"kubernetes.io/projected/b7053a68-ce89-48b6-9b5d-1c6dac813a75-kube-api-access-h6kj7\") pod \"glance-operator-controller-manager-c6994669c-hvq6s\" (UID: \"b7053a68-ce89-48b6-9b5d-1c6dac813a75\") " pod="openstack-operators/glance-operator-controller-manager-c6994669c-hvq6s" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.740841 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h66xm\" (UniqueName: \"kubernetes.io/projected/a25733d4-a5f8-4c97-a897-8d96e637c253-kube-api-access-h66xm\") pod \"cinder-operator-controller-manager-9b68f5989-24r54\" (UID: \"a25733d4-a5f8-4c97-a897-8d96e637c253\") " pod="openstack-operators/cinder-operator-controller-manager-9b68f5989-24r54" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.740888 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vdzth\" (UniqueName: \"kubernetes.io/projected/8ed6bb2c-2770-4785-bf38-9b6fd0bd8472-kube-api-access-vdzth\") pod \"barbican-operator-controller-manager-7ddb5c749-bvptq\" (UID: \"8ed6bb2c-2770-4785-bf38-9b6fd0bd8472\") " pod="openstack-operators/barbican-operator-controller-manager-7ddb5c749-bvptq" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.740915 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2kwld\" (UniqueName: \"kubernetes.io/projected/14dd1f12-3943-42e4-be1b-c9e37e49d21b-kube-api-access-2kwld\") pod \"designate-operator-controller-manager-9f958b845-wtfj4\" (UID: \"14dd1f12-3943-42e4-be1b-c9e37e49d21b\") " pod="openstack-operators/designate-operator-controller-manager-9f958b845-wtfj4" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.740932 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dd2tc\" (UniqueName: \"kubernetes.io/projected/944c3123-9063-4731-9581-5b683741edc4-kube-api-access-dd2tc\") pod \"heat-operator-controller-manager-5fbc59b5b6-hgmf4\" (UID: \"944c3123-9063-4731-9581-5b683741edc4\") " pod="openstack-operators/heat-operator-controller-manager-5fbc59b5b6-hgmf4" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.740946 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f5dbk\" (UniqueName: \"kubernetes.io/projected/2463afb1-0271-4ab3-ad1a-9329ecaeec4d-kube-api-access-f5dbk\") pod \"horizon-operator-controller-manager-77d5c5b54f-mp5ml\" (UID: \"2463afb1-0271-4ab3-ad1a-9329ecaeec4d\") " pod="openstack-operators/horizon-operator-controller-manager-77d5c5b54f-mp5ml" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.748730 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-vdvrx" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.750921 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-c87fff755-d5hn9"] Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.765378 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-cb4666565-7w8p4"] Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.766091 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-cb4666565-7w8p4" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.767507 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-fd68l" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.774223 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2kwld\" (UniqueName: \"kubernetes.io/projected/14dd1f12-3943-42e4-be1b-c9e37e49d21b-kube-api-access-2kwld\") pod \"designate-operator-controller-manager-9f958b845-wtfj4\" (UID: \"14dd1f12-3943-42e4-be1b-c9e37e49d21b\") " pod="openstack-operators/designate-operator-controller-manager-9f958b845-wtfj4" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.775041 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-cb4666565-7w8p4"] Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.775881 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vdzth\" (UniqueName: \"kubernetes.io/projected/8ed6bb2c-2770-4785-bf38-9b6fd0bd8472-kube-api-access-vdzth\") pod \"barbican-operator-controller-manager-7ddb5c749-bvptq\" (UID: \"8ed6bb2c-2770-4785-bf38-9b6fd0bd8472\") " pod="openstack-operators/barbican-operator-controller-manager-7ddb5c749-bvptq" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.778674 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-65849867d6-8ngx6"] Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.780313 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-65849867d6-8ngx6" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.781289 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dd2tc\" (UniqueName: \"kubernetes.io/projected/944c3123-9063-4731-9581-5b683741edc4-kube-api-access-dd2tc\") pod \"heat-operator-controller-manager-5fbc59b5b6-hgmf4\" (UID: \"944c3123-9063-4731-9581-5b683741edc4\") " pod="openstack-operators/heat-operator-controller-manager-5fbc59b5b6-hgmf4" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.781532 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f5dbk\" (UniqueName: \"kubernetes.io/projected/2463afb1-0271-4ab3-ad1a-9329ecaeec4d-kube-api-access-f5dbk\") pod \"horizon-operator-controller-manager-77d5c5b54f-mp5ml\" (UID: \"2463afb1-0271-4ab3-ad1a-9329ecaeec4d\") " pod="openstack-operators/horizon-operator-controller-manager-77d5c5b54f-mp5ml" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.786267 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h6kj7\" (UniqueName: \"kubernetes.io/projected/b7053a68-ce89-48b6-9b5d-1c6dac813a75-kube-api-access-h6kj7\") pod \"glance-operator-controller-manager-c6994669c-hvq6s\" (UID: \"b7053a68-ce89-48b6-9b5d-1c6dac813a75\") " pod="openstack-operators/glance-operator-controller-manager-c6994669c-hvq6s" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.787260 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h66xm\" (UniqueName: \"kubernetes.io/projected/a25733d4-a5f8-4c97-a897-8d96e637c253-kube-api-access-h66xm\") pod \"cinder-operator-controller-manager-9b68f5989-24r54\" (UID: \"a25733d4-a5f8-4c97-a897-8d96e637c253\") " pod="openstack-operators/cinder-operator-controller-manager-9b68f5989-24r54" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.789253 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-r9fdm" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.789575 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-65849867d6-8ngx6"] Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.792249 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-7fc9b76cf6-tncbx"] Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.792879 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-7fc9b76cf6-tncbx" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.794747 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-89gkm" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.797175 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-6b68b8b854zxcbl"] Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.797900 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6b68b8b854zxcbl" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.799818 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.800026 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-ct526" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.808498 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-55db956ddc-78z2q"] Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.809940 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-55db956ddc-78z2q" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.810797 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-7fc9b76cf6-tncbx"] Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.811100 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-sz9vr" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.813749 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-55db956ddc-78z2q"] Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.816747 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-6b68b8b854zxcbl"] Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.822838 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-686df47fcb-c8xvs"] Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.823534 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-686df47fcb-c8xvs" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.824571 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-vlz5r" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.830883 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-686df47fcb-c8xvs"] Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.841576 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-79prb\" (UniqueName: \"kubernetes.io/projected/19b1c150-f5ed-4432-bfc2-717b78eebc7e-kube-api-access-79prb\") pod \"mariadb-operator-controller-manager-c87fff755-d5hn9\" (UID: \"19b1c150-f5ed-4432-bfc2-717b78eebc7e\") " pod="openstack-operators/mariadb-operator-controller-manager-c87fff755-d5hn9" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.841633 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-26h5f\" (UniqueName: \"kubernetes.io/projected/2e5efa9b-40bc-4860-967d-aa37e1838f81-kube-api-access-26h5f\") pod \"keystone-operator-controller-manager-767fdc4f47-5mhmd\" (UID: \"2e5efa9b-40bc-4860-967d-aa37e1838f81\") " pod="openstack-operators/keystone-operator-controller-manager-767fdc4f47-5mhmd" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.841687 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qzdsw\" (UniqueName: \"kubernetes.io/projected/fc911701-7c15-4a20-b89e-709521bdd3a1-kube-api-access-qzdsw\") pod \"ironic-operator-controller-manager-78757b4889-gkwjs\" (UID: \"fc911701-7c15-4a20-b89e-709521bdd3a1\") " pod="openstack-operators/ironic-operator-controller-manager-78757b4889-gkwjs" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.841742 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r2vk8\" (UniqueName: \"kubernetes.io/projected/57e1fe13-b2c8-4711-9efe-1ebbc7e9d12e-kube-api-access-r2vk8\") pod \"infra-operator-controller-manager-77c48c7859-dzn7w\" (UID: \"57e1fe13-b2c8-4711-9efe-1ebbc7e9d12e\") " pod="openstack-operators/infra-operator-controller-manager-77c48c7859-dzn7w" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.841767 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/57e1fe13-b2c8-4711-9efe-1ebbc7e9d12e-cert\") pod \"infra-operator-controller-manager-77c48c7859-dzn7w\" (UID: \"57e1fe13-b2c8-4711-9efe-1ebbc7e9d12e\") " pod="openstack-operators/infra-operator-controller-manager-77c48c7859-dzn7w" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.841792 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vl8hm\" (UniqueName: \"kubernetes.io/projected/18017a19-1676-4fe9-a97e-ae878f423847-kube-api-access-vl8hm\") pod \"manila-operator-controller-manager-864f6b75bf-c7vsf\" (UID: \"18017a19-1676-4fe9-a97e-ae878f423847\") " pod="openstack-operators/manila-operator-controller-manager-864f6b75bf-c7vsf" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.845030 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-85dd56d4cc-wlrdt"] Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.845599 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-85dd56d4cc-wlrdt" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.848619 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-bcx55" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.855628 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-85dd56d4cc-wlrdt"] Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.865378 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-7ddb5c749-bvptq" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.874555 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-9b68f5989-24r54" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.893343 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-9f958b845-wtfj4" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.899616 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-c6994669c-hvq6s" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.922210 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-77d5c5b54f-mp5ml" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.932517 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5fbc59b5b6-hgmf4" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.944223 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mv9nq\" (UniqueName: \"kubernetes.io/projected/8c74d3da-d870-45e9-96b9-70eb2395c977-kube-api-access-mv9nq\") pod \"neutron-operator-controller-manager-cb4666565-7w8p4\" (UID: \"8c74d3da-d870-45e9-96b9-70eb2395c977\") " pod="openstack-operators/neutron-operator-controller-manager-cb4666565-7w8p4" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.944262 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qzdsw\" (UniqueName: \"kubernetes.io/projected/fc911701-7c15-4a20-b89e-709521bdd3a1-kube-api-access-qzdsw\") pod \"ironic-operator-controller-manager-78757b4889-gkwjs\" (UID: \"fc911701-7c15-4a20-b89e-709521bdd3a1\") " pod="openstack-operators/ironic-operator-controller-manager-78757b4889-gkwjs" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.944328 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r2vk8\" (UniqueName: \"kubernetes.io/projected/57e1fe13-b2c8-4711-9efe-1ebbc7e9d12e-kube-api-access-r2vk8\") pod \"infra-operator-controller-manager-77c48c7859-dzn7w\" (UID: \"57e1fe13-b2c8-4711-9efe-1ebbc7e9d12e\") " pod="openstack-operators/infra-operator-controller-manager-77c48c7859-dzn7w" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.944371 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/57e1fe13-b2c8-4711-9efe-1ebbc7e9d12e-cert\") pod \"infra-operator-controller-manager-77c48c7859-dzn7w\" (UID: \"57e1fe13-b2c8-4711-9efe-1ebbc7e9d12e\") " pod="openstack-operators/infra-operator-controller-manager-77c48c7859-dzn7w" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.944400 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vl8hm\" (UniqueName: \"kubernetes.io/projected/18017a19-1676-4fe9-a97e-ae878f423847-kube-api-access-vl8hm\") pod \"manila-operator-controller-manager-864f6b75bf-c7vsf\" (UID: \"18017a19-1676-4fe9-a97e-ae878f423847\") " pod="openstack-operators/manila-operator-controller-manager-864f6b75bf-c7vsf" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.944446 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t2gng\" (UniqueName: \"kubernetes.io/projected/f5639f17-3a5b-4ef4-bd6e-6eeb3274636a-kube-api-access-t2gng\") pod \"ovn-operator-controller-manager-55db956ddc-78z2q\" (UID: \"f5639f17-3a5b-4ef4-bd6e-6eeb3274636a\") " pod="openstack-operators/ovn-operator-controller-manager-55db956ddc-78z2q" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.944478 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-79prb\" (UniqueName: \"kubernetes.io/projected/19b1c150-f5ed-4432-bfc2-717b78eebc7e-kube-api-access-79prb\") pod \"mariadb-operator-controller-manager-c87fff755-d5hn9\" (UID: \"19b1c150-f5ed-4432-bfc2-717b78eebc7e\") " pod="openstack-operators/mariadb-operator-controller-manager-c87fff755-d5hn9" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.944498 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xcm52\" (UniqueName: \"kubernetes.io/projected/3fa95966-3147-4e50-8ed8-f702a87992c4-kube-api-access-xcm52\") pod \"openstack-baremetal-operator-controller-manager-6b68b8b854zxcbl\" (UID: \"3fa95966-3147-4e50-8ed8-f702a87992c4\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6b68b8b854zxcbl" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.944534 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hd9q4\" (UniqueName: \"kubernetes.io/projected/b38ac60b-1643-44d9-a1fe-5f75129247dc-kube-api-access-hd9q4\") pod \"nova-operator-controller-manager-65849867d6-8ngx6\" (UID: \"b38ac60b-1643-44d9-a1fe-5f75129247dc\") " pod="openstack-operators/nova-operator-controller-manager-65849867d6-8ngx6" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.944555 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dk5bv\" (UniqueName: \"kubernetes.io/projected/d8b36106-8f21-4873-890d-c1879ea49068-kube-api-access-dk5bv\") pod \"placement-operator-controller-manager-686df47fcb-c8xvs\" (UID: \"d8b36106-8f21-4873-890d-c1879ea49068\") " pod="openstack-operators/placement-operator-controller-manager-686df47fcb-c8xvs" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.944579 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/3fa95966-3147-4e50-8ed8-f702a87992c4-cert\") pod \"openstack-baremetal-operator-controller-manager-6b68b8b854zxcbl\" (UID: \"3fa95966-3147-4e50-8ed8-f702a87992c4\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6b68b8b854zxcbl" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.944621 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-26h5f\" (UniqueName: \"kubernetes.io/projected/2e5efa9b-40bc-4860-967d-aa37e1838f81-kube-api-access-26h5f\") pod \"keystone-operator-controller-manager-767fdc4f47-5mhmd\" (UID: \"2e5efa9b-40bc-4860-967d-aa37e1838f81\") " pod="openstack-operators/keystone-operator-controller-manager-767fdc4f47-5mhmd" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.944649 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hwdw8\" (UniqueName: \"kubernetes.io/projected/5429da27-a618-44da-8aeb-12b917af8734-kube-api-access-hwdw8\") pod \"octavia-operator-controller-manager-7fc9b76cf6-tncbx\" (UID: \"5429da27-a618-44da-8aeb-12b917af8734\") " pod="openstack-operators/octavia-operator-controller-manager-7fc9b76cf6-tncbx" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.945278 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-5f8f495fcf-6rz5h"] Jan 21 11:23:43 crc kubenswrapper[4824]: E0121 11:23:43.945542 4824 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Jan 21 11:23:43 crc kubenswrapper[4824]: E0121 11:23:43.945582 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/57e1fe13-b2c8-4711-9efe-1ebbc7e9d12e-cert podName:57e1fe13-b2c8-4711-9efe-1ebbc7e9d12e nodeName:}" failed. No retries permitted until 2026-01-21 11:23:44.445568859 +0000 UTC m=+766.738598142 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/57e1fe13-b2c8-4711-9efe-1ebbc7e9d12e-cert") pod "infra-operator-controller-manager-77c48c7859-dzn7w" (UID: "57e1fe13-b2c8-4711-9efe-1ebbc7e9d12e") : secret "infra-operator-webhook-server-cert" not found Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.948499 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-5f8f495fcf-6rz5h" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.951852 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-dptjq" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.970174 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qzdsw\" (UniqueName: \"kubernetes.io/projected/fc911701-7c15-4a20-b89e-709521bdd3a1-kube-api-access-qzdsw\") pod \"ironic-operator-controller-manager-78757b4889-gkwjs\" (UID: \"fc911701-7c15-4a20-b89e-709521bdd3a1\") " pod="openstack-operators/ironic-operator-controller-manager-78757b4889-gkwjs" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.971407 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-79prb\" (UniqueName: \"kubernetes.io/projected/19b1c150-f5ed-4432-bfc2-717b78eebc7e-kube-api-access-79prb\") pod \"mariadb-operator-controller-manager-c87fff755-d5hn9\" (UID: \"19b1c150-f5ed-4432-bfc2-717b78eebc7e\") " pod="openstack-operators/mariadb-operator-controller-manager-c87fff755-d5hn9" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.972329 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vl8hm\" (UniqueName: \"kubernetes.io/projected/18017a19-1676-4fe9-a97e-ae878f423847-kube-api-access-vl8hm\") pod \"manila-operator-controller-manager-864f6b75bf-c7vsf\" (UID: \"18017a19-1676-4fe9-a97e-ae878f423847\") " pod="openstack-operators/manila-operator-controller-manager-864f6b75bf-c7vsf" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.975797 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-78757b4889-gkwjs" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.976059 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r2vk8\" (UniqueName: \"kubernetes.io/projected/57e1fe13-b2c8-4711-9efe-1ebbc7e9d12e-kube-api-access-r2vk8\") pod \"infra-operator-controller-manager-77c48c7859-dzn7w\" (UID: \"57e1fe13-b2c8-4711-9efe-1ebbc7e9d12e\") " pod="openstack-operators/infra-operator-controller-manager-77c48c7859-dzn7w" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.976272 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-26h5f\" (UniqueName: \"kubernetes.io/projected/2e5efa9b-40bc-4860-967d-aa37e1838f81-kube-api-access-26h5f\") pod \"keystone-operator-controller-manager-767fdc4f47-5mhmd\" (UID: \"2e5efa9b-40bc-4860-967d-aa37e1838f81\") " pod="openstack-operators/keystone-operator-controller-manager-767fdc4f47-5mhmd" Jan 21 11:23:43 crc kubenswrapper[4824]: I0121 11:23:43.977377 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-5f8f495fcf-6rz5h"] Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.022128 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-767fdc4f47-5mhmd" Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.046970 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-7cd8bc9dbb-xkp8x"] Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.052933 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t2gng\" (UniqueName: \"kubernetes.io/projected/f5639f17-3a5b-4ef4-bd6e-6eeb3274636a-kube-api-access-t2gng\") pod \"ovn-operator-controller-manager-55db956ddc-78z2q\" (UID: \"f5639f17-3a5b-4ef4-bd6e-6eeb3274636a\") " pod="openstack-operators/ovn-operator-controller-manager-55db956ddc-78z2q" Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.053118 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xcm52\" (UniqueName: \"kubernetes.io/projected/3fa95966-3147-4e50-8ed8-f702a87992c4-kube-api-access-xcm52\") pod \"openstack-baremetal-operator-controller-manager-6b68b8b854zxcbl\" (UID: \"3fa95966-3147-4e50-8ed8-f702a87992c4\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6b68b8b854zxcbl" Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.053171 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hd9q4\" (UniqueName: \"kubernetes.io/projected/b38ac60b-1643-44d9-a1fe-5f75129247dc-kube-api-access-hd9q4\") pod \"nova-operator-controller-manager-65849867d6-8ngx6\" (UID: \"b38ac60b-1643-44d9-a1fe-5f75129247dc\") " pod="openstack-operators/nova-operator-controller-manager-65849867d6-8ngx6" Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.053315 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dk5bv\" (UniqueName: \"kubernetes.io/projected/d8b36106-8f21-4873-890d-c1879ea49068-kube-api-access-dk5bv\") pod \"placement-operator-controller-manager-686df47fcb-c8xvs\" (UID: \"d8b36106-8f21-4873-890d-c1879ea49068\") " pod="openstack-operators/placement-operator-controller-manager-686df47fcb-c8xvs" Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.053389 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/3fa95966-3147-4e50-8ed8-f702a87992c4-cert\") pod \"openstack-baremetal-operator-controller-manager-6b68b8b854zxcbl\" (UID: \"3fa95966-3147-4e50-8ed8-f702a87992c4\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6b68b8b854zxcbl" Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.053548 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hwdw8\" (UniqueName: \"kubernetes.io/projected/5429da27-a618-44da-8aeb-12b917af8734-kube-api-access-hwdw8\") pod \"octavia-operator-controller-manager-7fc9b76cf6-tncbx\" (UID: \"5429da27-a618-44da-8aeb-12b917af8734\") " pod="openstack-operators/octavia-operator-controller-manager-7fc9b76cf6-tncbx" Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.053644 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jvmv5\" (UniqueName: \"kubernetes.io/projected/db84b172-59d8-4995-815b-50b492283bee-kube-api-access-jvmv5\") pod \"telemetry-operator-controller-manager-5f8f495fcf-6rz5h\" (UID: \"db84b172-59d8-4995-815b-50b492283bee\") " pod="openstack-operators/telemetry-operator-controller-manager-5f8f495fcf-6rz5h" Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.053909 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mv9nq\" (UniqueName: \"kubernetes.io/projected/8c74d3da-d870-45e9-96b9-70eb2395c977-kube-api-access-mv9nq\") pod \"neutron-operator-controller-manager-cb4666565-7w8p4\" (UID: \"8c74d3da-d870-45e9-96b9-70eb2395c977\") " pod="openstack-operators/neutron-operator-controller-manager-cb4666565-7w8p4" Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.054576 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kqcnj\" (UniqueName: \"kubernetes.io/projected/b4cb80e6-c720-4875-9848-7d9597a354e4-kube-api-access-kqcnj\") pod \"swift-operator-controller-manager-85dd56d4cc-wlrdt\" (UID: \"b4cb80e6-c720-4875-9848-7d9597a354e4\") " pod="openstack-operators/swift-operator-controller-manager-85dd56d4cc-wlrdt" Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.057043 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-7cd8bc9dbb-xkp8x" Jan 21 11:23:44 crc kubenswrapper[4824]: E0121 11:23:44.057401 4824 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Jan 21 11:23:44 crc kubenswrapper[4824]: E0121 11:23:44.057637 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3fa95966-3147-4e50-8ed8-f702a87992c4-cert podName:3fa95966-3147-4e50-8ed8-f702a87992c4 nodeName:}" failed. No retries permitted until 2026-01-21 11:23:44.557617736 +0000 UTC m=+766.850647028 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/3fa95966-3147-4e50-8ed8-f702a87992c4-cert") pod "openstack-baremetal-operator-controller-manager-6b68b8b854zxcbl" (UID: "3fa95966-3147-4e50-8ed8-f702a87992c4") : secret "openstack-baremetal-operator-webhook-server-cert" not found Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.072097 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-bdjsk" Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.073599 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-c87fff755-d5hn9" Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.083270 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-864f6b75bf-c7vsf" Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.099075 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dk5bv\" (UniqueName: \"kubernetes.io/projected/d8b36106-8f21-4873-890d-c1879ea49068-kube-api-access-dk5bv\") pod \"placement-operator-controller-manager-686df47fcb-c8xvs\" (UID: \"d8b36106-8f21-4873-890d-c1879ea49068\") " pod="openstack-operators/placement-operator-controller-manager-686df47fcb-c8xvs" Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.100307 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hd9q4\" (UniqueName: \"kubernetes.io/projected/b38ac60b-1643-44d9-a1fe-5f75129247dc-kube-api-access-hd9q4\") pod \"nova-operator-controller-manager-65849867d6-8ngx6\" (UID: \"b38ac60b-1643-44d9-a1fe-5f75129247dc\") " pod="openstack-operators/nova-operator-controller-manager-65849867d6-8ngx6" Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.100739 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t2gng\" (UniqueName: \"kubernetes.io/projected/f5639f17-3a5b-4ef4-bd6e-6eeb3274636a-kube-api-access-t2gng\") pod \"ovn-operator-controller-manager-55db956ddc-78z2q\" (UID: \"f5639f17-3a5b-4ef4-bd6e-6eeb3274636a\") " pod="openstack-operators/ovn-operator-controller-manager-55db956ddc-78z2q" Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.101441 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hwdw8\" (UniqueName: \"kubernetes.io/projected/5429da27-a618-44da-8aeb-12b917af8734-kube-api-access-hwdw8\") pod \"octavia-operator-controller-manager-7fc9b76cf6-tncbx\" (UID: \"5429da27-a618-44da-8aeb-12b917af8734\") " pod="openstack-operators/octavia-operator-controller-manager-7fc9b76cf6-tncbx" Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.112429 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mv9nq\" (UniqueName: \"kubernetes.io/projected/8c74d3da-d870-45e9-96b9-70eb2395c977-kube-api-access-mv9nq\") pod \"neutron-operator-controller-manager-cb4666565-7w8p4\" (UID: \"8c74d3da-d870-45e9-96b9-70eb2395c977\") " pod="openstack-operators/neutron-operator-controller-manager-cb4666565-7w8p4" Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.113526 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xcm52\" (UniqueName: \"kubernetes.io/projected/3fa95966-3147-4e50-8ed8-f702a87992c4-kube-api-access-xcm52\") pod \"openstack-baremetal-operator-controller-manager-6b68b8b854zxcbl\" (UID: \"3fa95966-3147-4e50-8ed8-f702a87992c4\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6b68b8b854zxcbl" Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.124537 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-cb4666565-7w8p4" Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.125813 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-7cd8bc9dbb-xkp8x"] Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.132424 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-65849867d6-8ngx6" Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.143256 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-7fc9b76cf6-tncbx" Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.155410 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-64cd966744-9w4ch"] Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.156230 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-64cd966744-9w4ch" Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.157526 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jvmv5\" (UniqueName: \"kubernetes.io/projected/db84b172-59d8-4995-815b-50b492283bee-kube-api-access-jvmv5\") pod \"telemetry-operator-controller-manager-5f8f495fcf-6rz5h\" (UID: \"db84b172-59d8-4995-815b-50b492283bee\") " pod="openstack-operators/telemetry-operator-controller-manager-5f8f495fcf-6rz5h" Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.157612 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kqcnj\" (UniqueName: \"kubernetes.io/projected/b4cb80e6-c720-4875-9848-7d9597a354e4-kube-api-access-kqcnj\") pod \"swift-operator-controller-manager-85dd56d4cc-wlrdt\" (UID: \"b4cb80e6-c720-4875-9848-7d9597a354e4\") " pod="openstack-operators/swift-operator-controller-manager-85dd56d4cc-wlrdt" Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.157784 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r4547\" (UniqueName: \"kubernetes.io/projected/52491e7b-cd8c-403a-942c-312debd6f1c6-kube-api-access-r4547\") pod \"test-operator-controller-manager-7cd8bc9dbb-xkp8x\" (UID: \"52491e7b-cd8c-403a-942c-312debd6f1c6\") " pod="openstack-operators/test-operator-controller-manager-7cd8bc9dbb-xkp8x" Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.160777 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-qtdgh" Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.161812 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-55db956ddc-78z2q" Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.166622 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-64cd966744-9w4ch"] Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.168104 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-686df47fcb-c8xvs" Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.182179 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jvmv5\" (UniqueName: \"kubernetes.io/projected/db84b172-59d8-4995-815b-50b492283bee-kube-api-access-jvmv5\") pod \"telemetry-operator-controller-manager-5f8f495fcf-6rz5h\" (UID: \"db84b172-59d8-4995-815b-50b492283bee\") " pod="openstack-operators/telemetry-operator-controller-manager-5f8f495fcf-6rz5h" Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.186506 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kqcnj\" (UniqueName: \"kubernetes.io/projected/b4cb80e6-c720-4875-9848-7d9597a354e4-kube-api-access-kqcnj\") pod \"swift-operator-controller-manager-85dd56d4cc-wlrdt\" (UID: \"b4cb80e6-c720-4875-9848-7d9597a354e4\") " pod="openstack-operators/swift-operator-controller-manager-85dd56d4cc-wlrdt" Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.258668 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r4547\" (UniqueName: \"kubernetes.io/projected/52491e7b-cd8c-403a-942c-312debd6f1c6-kube-api-access-r4547\") pod \"test-operator-controller-manager-7cd8bc9dbb-xkp8x\" (UID: \"52491e7b-cd8c-403a-942c-312debd6f1c6\") " pod="openstack-operators/test-operator-controller-manager-7cd8bc9dbb-xkp8x" Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.258712 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pqc8b\" (UniqueName: \"kubernetes.io/projected/0715aa91-2ae5-4e07-8388-c37def4c7743-kube-api-access-pqc8b\") pod \"watcher-operator-controller-manager-64cd966744-9w4ch\" (UID: \"0715aa91-2ae5-4e07-8388-c37def4c7743\") " pod="openstack-operators/watcher-operator-controller-manager-64cd966744-9w4ch" Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.283914 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-74d84f758c-vjgps"] Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.284682 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-74d84f758c-vjgps" Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.286693 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-5f8f495fcf-6rz5h" Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.289232 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"metrics-server-cert" Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.295274 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.295432 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-xjvvw" Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.302812 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r4547\" (UniqueName: \"kubernetes.io/projected/52491e7b-cd8c-403a-942c-312debd6f1c6-kube-api-access-r4547\") pod \"test-operator-controller-manager-7cd8bc9dbb-xkp8x\" (UID: \"52491e7b-cd8c-403a-942c-312debd6f1c6\") " pod="openstack-operators/test-operator-controller-manager-7cd8bc9dbb-xkp8x" Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.325906 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-74d84f758c-vjgps"] Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.359668 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pqc8b\" (UniqueName: \"kubernetes.io/projected/0715aa91-2ae5-4e07-8388-c37def4c7743-kube-api-access-pqc8b\") pod \"watcher-operator-controller-manager-64cd966744-9w4ch\" (UID: \"0715aa91-2ae5-4e07-8388-c37def4c7743\") " pod="openstack-operators/watcher-operator-controller-manager-64cd966744-9w4ch" Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.359752 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c1128e01-8b7c-4742-b1da-dbf06aa8b4d8-metrics-certs\") pod \"openstack-operator-controller-manager-74d84f758c-vjgps\" (UID: \"c1128e01-8b7c-4742-b1da-dbf06aa8b4d8\") " pod="openstack-operators/openstack-operator-controller-manager-74d84f758c-vjgps" Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.359785 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/c1128e01-8b7c-4742-b1da-dbf06aa8b4d8-webhook-certs\") pod \"openstack-operator-controller-manager-74d84f758c-vjgps\" (UID: \"c1128e01-8b7c-4742-b1da-dbf06aa8b4d8\") " pod="openstack-operators/openstack-operator-controller-manager-74d84f758c-vjgps" Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.359867 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6spbp\" (UniqueName: \"kubernetes.io/projected/c1128e01-8b7c-4742-b1da-dbf06aa8b4d8-kube-api-access-6spbp\") pod \"openstack-operator-controller-manager-74d84f758c-vjgps\" (UID: \"c1128e01-8b7c-4742-b1da-dbf06aa8b4d8\") " pod="openstack-operators/openstack-operator-controller-manager-74d84f758c-vjgps" Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.382359 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pqc8b\" (UniqueName: \"kubernetes.io/projected/0715aa91-2ae5-4e07-8388-c37def4c7743-kube-api-access-pqc8b\") pod \"watcher-operator-controller-manager-64cd966744-9w4ch\" (UID: \"0715aa91-2ae5-4e07-8388-c37def4c7743\") " pod="openstack-operators/watcher-operator-controller-manager-64cd966744-9w4ch" Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.394064 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7ddb5c749-bvptq"] Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.398702 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-94lqx"] Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.399498 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-94lqx" Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.406501 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-w4k22" Jan 21 11:23:44 crc kubenswrapper[4824]: W0121 11:23:44.409735 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8ed6bb2c_2770_4785_bf38_9b6fd0bd8472.slice/crio-b6219845aa9a1771a156cc32960916cee0673df44cdd949c240384b6968c356d WatchSource:0}: Error finding container b6219845aa9a1771a156cc32960916cee0673df44cdd949c240384b6968c356d: Status 404 returned error can't find the container with id b6219845aa9a1771a156cc32960916cee0673df44cdd949c240384b6968c356d Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.419163 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-94lqx"] Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.428857 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-7cd8bc9dbb-xkp8x" Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.429183 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-9b68f5989-24r54"] Jan 21 11:23:44 crc kubenswrapper[4824]: W0121 11:23:44.434848 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda25733d4_a5f8_4c97_a897_8d96e637c253.slice/crio-f69ed3b642e0e14f2ed7c1d550fcce820181b38e087c18d787fe936ecdb84588 WatchSource:0}: Error finding container f69ed3b642e0e14f2ed7c1d550fcce820181b38e087c18d787fe936ecdb84588: Status 404 returned error can't find the container with id f69ed3b642e0e14f2ed7c1d550fcce820181b38e087c18d787fe936ecdb84588 Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.436135 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-9f958b845-wtfj4"] Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.460669 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c1128e01-8b7c-4742-b1da-dbf06aa8b4d8-metrics-certs\") pod \"openstack-operator-controller-manager-74d84f758c-vjgps\" (UID: \"c1128e01-8b7c-4742-b1da-dbf06aa8b4d8\") " pod="openstack-operators/openstack-operator-controller-manager-74d84f758c-vjgps" Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.460708 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/c1128e01-8b7c-4742-b1da-dbf06aa8b4d8-webhook-certs\") pod \"openstack-operator-controller-manager-74d84f758c-vjgps\" (UID: \"c1128e01-8b7c-4742-b1da-dbf06aa8b4d8\") " pod="openstack-operators/openstack-operator-controller-manager-74d84f758c-vjgps" Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.460793 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-77d5c5b54f-mp5ml"] Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.460804 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6spbp\" (UniqueName: \"kubernetes.io/projected/c1128e01-8b7c-4742-b1da-dbf06aa8b4d8-kube-api-access-6spbp\") pod \"openstack-operator-controller-manager-74d84f758c-vjgps\" (UID: \"c1128e01-8b7c-4742-b1da-dbf06aa8b4d8\") " pod="openstack-operators/openstack-operator-controller-manager-74d84f758c-vjgps" Jan 21 11:23:44 crc kubenswrapper[4824]: E0121 11:23:44.460804 4824 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Jan 21 11:23:44 crc kubenswrapper[4824]: E0121 11:23:44.460915 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c1128e01-8b7c-4742-b1da-dbf06aa8b4d8-metrics-certs podName:c1128e01-8b7c-4742-b1da-dbf06aa8b4d8 nodeName:}" failed. No retries permitted until 2026-01-21 11:23:44.960902243 +0000 UTC m=+767.253931536 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/c1128e01-8b7c-4742-b1da-dbf06aa8b4d8-metrics-certs") pod "openstack-operator-controller-manager-74d84f758c-vjgps" (UID: "c1128e01-8b7c-4742-b1da-dbf06aa8b4d8") : secret "metrics-server-cert" not found Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.460867 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/57e1fe13-b2c8-4711-9efe-1ebbc7e9d12e-cert\") pod \"infra-operator-controller-manager-77c48c7859-dzn7w\" (UID: \"57e1fe13-b2c8-4711-9efe-1ebbc7e9d12e\") " pod="openstack-operators/infra-operator-controller-manager-77c48c7859-dzn7w" Jan 21 11:23:44 crc kubenswrapper[4824]: E0121 11:23:44.460866 4824 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Jan 21 11:23:44 crc kubenswrapper[4824]: E0121 11:23:44.460943 4824 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.460951 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5v855\" (UniqueName: \"kubernetes.io/projected/782093ab-751c-4e38-bba3-c5f76308c82e-kube-api-access-5v855\") pod \"rabbitmq-cluster-operator-manager-668c99d594-94lqx\" (UID: \"782093ab-751c-4e38-bba3-c5f76308c82e\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-94lqx" Jan 21 11:23:44 crc kubenswrapper[4824]: E0121 11:23:44.461026 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/57e1fe13-b2c8-4711-9efe-1ebbc7e9d12e-cert podName:57e1fe13-b2c8-4711-9efe-1ebbc7e9d12e nodeName:}" failed. No retries permitted until 2026-01-21 11:23:45.461014074 +0000 UTC m=+767.754043366 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/57e1fe13-b2c8-4711-9efe-1ebbc7e9d12e-cert") pod "infra-operator-controller-manager-77c48c7859-dzn7w" (UID: "57e1fe13-b2c8-4711-9efe-1ebbc7e9d12e") : secret "infra-operator-webhook-server-cert" not found Jan 21 11:23:44 crc kubenswrapper[4824]: E0121 11:23:44.461045 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c1128e01-8b7c-4742-b1da-dbf06aa8b4d8-webhook-certs podName:c1128e01-8b7c-4742-b1da-dbf06aa8b4d8 nodeName:}" failed. No retries permitted until 2026-01-21 11:23:44.96103846 +0000 UTC m=+767.254067752 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/c1128e01-8b7c-4742-b1da-dbf06aa8b4d8-webhook-certs") pod "openstack-operator-controller-manager-74d84f758c-vjgps" (UID: "c1128e01-8b7c-4742-b1da-dbf06aa8b4d8") : secret "webhook-server-cert" not found Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.475207 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-85dd56d4cc-wlrdt" Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.482017 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6spbp\" (UniqueName: \"kubernetes.io/projected/c1128e01-8b7c-4742-b1da-dbf06aa8b4d8-kube-api-access-6spbp\") pod \"openstack-operator-controller-manager-74d84f758c-vjgps\" (UID: \"c1128e01-8b7c-4742-b1da-dbf06aa8b4d8\") " pod="openstack-operators/openstack-operator-controller-manager-74d84f758c-vjgps" Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.483570 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-64cd966744-9w4ch" Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.561675 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5v855\" (UniqueName: \"kubernetes.io/projected/782093ab-751c-4e38-bba3-c5f76308c82e-kube-api-access-5v855\") pod \"rabbitmq-cluster-operator-manager-668c99d594-94lqx\" (UID: \"782093ab-751c-4e38-bba3-c5f76308c82e\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-94lqx" Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.561761 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/3fa95966-3147-4e50-8ed8-f702a87992c4-cert\") pod \"openstack-baremetal-operator-controller-manager-6b68b8b854zxcbl\" (UID: \"3fa95966-3147-4e50-8ed8-f702a87992c4\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6b68b8b854zxcbl" Jan 21 11:23:44 crc kubenswrapper[4824]: E0121 11:23:44.561867 4824 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Jan 21 11:23:44 crc kubenswrapper[4824]: E0121 11:23:44.561903 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3fa95966-3147-4e50-8ed8-f702a87992c4-cert podName:3fa95966-3147-4e50-8ed8-f702a87992c4 nodeName:}" failed. No retries permitted until 2026-01-21 11:23:45.561892363 +0000 UTC m=+767.854921655 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/3fa95966-3147-4e50-8ed8-f702a87992c4-cert") pod "openstack-baremetal-operator-controller-manager-6b68b8b854zxcbl" (UID: "3fa95966-3147-4e50-8ed8-f702a87992c4") : secret "openstack-baremetal-operator-webhook-server-cert" not found Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.579542 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5v855\" (UniqueName: \"kubernetes.io/projected/782093ab-751c-4e38-bba3-c5f76308c82e-kube-api-access-5v855\") pod \"rabbitmq-cluster-operator-manager-668c99d594-94lqx\" (UID: \"782093ab-751c-4e38-bba3-c5f76308c82e\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-94lqx" Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.639824 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5fbc59b5b6-hgmf4"] Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.653408 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-78757b4889-gkwjs"] Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.673983 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-c6994669c-hvq6s"] Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.733448 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-94lqx" Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.775342 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-767fdc4f47-5mhmd"] Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.790102 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-77d5c5b54f-mp5ml" event={"ID":"2463afb1-0271-4ab3-ad1a-9329ecaeec4d","Type":"ContainerStarted","Data":"9b62e3ccc2517416d91485038728c16ef411f1e88c5143cd21730db6b19f1309"} Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.793571 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-9b68f5989-24r54" event={"ID":"a25733d4-a5f8-4c97-a897-8d96e637c253","Type":"ContainerStarted","Data":"f69ed3b642e0e14f2ed7c1d550fcce820181b38e087c18d787fe936ecdb84588"} Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.794615 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-9f958b845-wtfj4" event={"ID":"14dd1f12-3943-42e4-be1b-c9e37e49d21b","Type":"ContainerStarted","Data":"fb96fc845fe4d6b80ac3f8cd82ccfab77cf45f6adcda11a32c14793a6e34f919"} Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.796725 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-c6994669c-hvq6s" event={"ID":"b7053a68-ce89-48b6-9b5d-1c6dac813a75","Type":"ContainerStarted","Data":"806f3e2a06a85342fe737a02653b5b4dc955ad63bc0aa934a4f3c1f3cb9984e5"} Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.801161 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-78757b4889-gkwjs" event={"ID":"fc911701-7c15-4a20-b89e-709521bdd3a1","Type":"ContainerStarted","Data":"fa06726926078f65ef97f8f89be9c291d385d655fc3dba940c74219d1c5360af"} Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.805630 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7ddb5c749-bvptq" event={"ID":"8ed6bb2c-2770-4785-bf38-9b6fd0bd8472","Type":"ContainerStarted","Data":"b6219845aa9a1771a156cc32960916cee0673df44cdd949c240384b6968c356d"} Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.806504 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5fbc59b5b6-hgmf4" event={"ID":"944c3123-9063-4731-9581-5b683741edc4","Type":"ContainerStarted","Data":"edc274cec7fd06e3df02f7ceb526561402e0899480a5e381bf4312fa7c43216d"} Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.883168 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-686df47fcb-c8xvs"] Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.912005 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-c87fff755-d5hn9"] Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.932013 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-864f6b75bf-c7vsf"] Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.969065 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c1128e01-8b7c-4742-b1da-dbf06aa8b4d8-metrics-certs\") pod \"openstack-operator-controller-manager-74d84f758c-vjgps\" (UID: \"c1128e01-8b7c-4742-b1da-dbf06aa8b4d8\") " pod="openstack-operators/openstack-operator-controller-manager-74d84f758c-vjgps" Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.969116 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/c1128e01-8b7c-4742-b1da-dbf06aa8b4d8-webhook-certs\") pod \"openstack-operator-controller-manager-74d84f758c-vjgps\" (UID: \"c1128e01-8b7c-4742-b1da-dbf06aa8b4d8\") " pod="openstack-operators/openstack-operator-controller-manager-74d84f758c-vjgps" Jan 21 11:23:44 crc kubenswrapper[4824]: E0121 11:23:44.969290 4824 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Jan 21 11:23:44 crc kubenswrapper[4824]: E0121 11:23:44.969333 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c1128e01-8b7c-4742-b1da-dbf06aa8b4d8-webhook-certs podName:c1128e01-8b7c-4742-b1da-dbf06aa8b4d8 nodeName:}" failed. No retries permitted until 2026-01-21 11:23:45.969320952 +0000 UTC m=+768.262350243 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/c1128e01-8b7c-4742-b1da-dbf06aa8b4d8-webhook-certs") pod "openstack-operator-controller-manager-74d84f758c-vjgps" (UID: "c1128e01-8b7c-4742-b1da-dbf06aa8b4d8") : secret "webhook-server-cert" not found Jan 21 11:23:44 crc kubenswrapper[4824]: E0121 11:23:44.969602 4824 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Jan 21 11:23:44 crc kubenswrapper[4824]: E0121 11:23:44.969642 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c1128e01-8b7c-4742-b1da-dbf06aa8b4d8-metrics-certs podName:c1128e01-8b7c-4742-b1da-dbf06aa8b4d8 nodeName:}" failed. No retries permitted until 2026-01-21 11:23:45.969633801 +0000 UTC m=+768.262663093 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/c1128e01-8b7c-4742-b1da-dbf06aa8b4d8-metrics-certs") pod "openstack-operator-controller-manager-74d84f758c-vjgps" (UID: "c1128e01-8b7c-4742-b1da-dbf06aa8b4d8") : secret "metrics-server-cert" not found Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.985892 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-cb4666565-7w8p4"] Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.988975 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-7fc9b76cf6-tncbx"] Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.992557 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-5f8f495fcf-6rz5h"] Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.995925 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-65849867d6-8ngx6"] Jan 21 11:23:44 crc kubenswrapper[4824]: I0121 11:23:44.999080 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-55db956ddc-78z2q"] Jan 21 11:23:45 crc kubenswrapper[4824]: E0121 11:23:45.003492 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/nova-operator@sha256:6defa56fc6a5bfbd5b27d28ff7b1c7bc89b24b2ef956e2a6d97b2726f668a231,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-hd9q4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-operator-controller-manager-65849867d6-8ngx6_openstack-operators(b38ac60b-1643-44d9-a1fe-5f75129247dc): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Jan 21 11:23:45 crc kubenswrapper[4824]: W0121 11:23:45.003705 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddb84b172_59d8_4995_815b_50b492283bee.slice/crio-dea14bc3dbf16dc26f2bf74918823842c7ee5b7d971485978ab0114af36bdf44 WatchSource:0}: Error finding container dea14bc3dbf16dc26f2bf74918823842c7ee5b7d971485978ab0114af36bdf44: Status 404 returned error can't find the container with id dea14bc3dbf16dc26f2bf74918823842c7ee5b7d971485978ab0114af36bdf44 Jan 21 11:23:45 crc kubenswrapper[4824]: E0121 11:23:45.004677 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/nova-operator-controller-manager-65849867d6-8ngx6" podUID="b38ac60b-1643-44d9-a1fe-5f75129247dc" Jan 21 11:23:45 crc kubenswrapper[4824]: E0121 11:23:45.005849 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/telemetry-operator@sha256:2e89109f5db66abf1afd15ef59bda35a53db40c5e59e020579ac5aa0acea1843,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-jvmv5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-5f8f495fcf-6rz5h_openstack-operators(db84b172-59d8-4995-815b-50b492283bee): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Jan 21 11:23:45 crc kubenswrapper[4824]: E0121 11:23:45.006979 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/telemetry-operator-controller-manager-5f8f495fcf-6rz5h" podUID="db84b172-59d8-4995-815b-50b492283bee" Jan 21 11:23:45 crc kubenswrapper[4824]: W0121 11:23:45.008180 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8c74d3da_d870_45e9_96b9_70eb2395c977.slice/crio-cc100a8bfa18f73c4ececc9bf8a9ded089e2a55ee7ce7b6379f51ec545977f6f WatchSource:0}: Error finding container cc100a8bfa18f73c4ececc9bf8a9ded089e2a55ee7ce7b6379f51ec545977f6f: Status 404 returned error can't find the container with id cc100a8bfa18f73c4ececc9bf8a9ded089e2a55ee7ce7b6379f51ec545977f6f Jan 21 11:23:45 crc kubenswrapper[4824]: E0121 11:23:45.008362 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ovn-operator@sha256:8b3bfb9e86618b7ac69443939b0968fae28a22cd62ea1e429b599ff9f8a5f8cf,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-t2gng,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-operator-controller-manager-55db956ddc-78z2q_openstack-operators(f5639f17-3a5b-4ef4-bd6e-6eeb3274636a): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Jan 21 11:23:45 crc kubenswrapper[4824]: E0121 11:23:45.009581 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/ovn-operator-controller-manager-55db956ddc-78z2q" podUID="f5639f17-3a5b-4ef4-bd6e-6eeb3274636a" Jan 21 11:23:45 crc kubenswrapper[4824]: E0121 11:23:45.010178 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/neutron-operator@sha256:0f440bf7dc937ce0135bdd328716686fd2f1320f453a9ac4e11e96383148ad6c,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-mv9nq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod neutron-operator-controller-manager-cb4666565-7w8p4_openstack-operators(8c74d3da-d870-45e9-96b9-70eb2395c977): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Jan 21 11:23:45 crc kubenswrapper[4824]: E0121 11:23:45.012297 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/neutron-operator-controller-manager-cb4666565-7w8p4" podUID="8c74d3da-d870-45e9-96b9-70eb2395c977" Jan 21 11:23:45 crc kubenswrapper[4824]: I0121 11:23:45.098941 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-85dd56d4cc-wlrdt"] Jan 21 11:23:45 crc kubenswrapper[4824]: I0121 11:23:45.103302 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-64cd966744-9w4ch"] Jan 21 11:23:45 crc kubenswrapper[4824]: I0121 11:23:45.108240 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-7cd8bc9dbb-xkp8x"] Jan 21 11:23:45 crc kubenswrapper[4824]: W0121 11:23:45.110773 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb4cb80e6_c720_4875_9848_7d9597a354e4.slice/crio-90eab79beadafe7c6d034fbad8cc7e89f7c0052842ddc3b3c0acf01963d3c3ab WatchSource:0}: Error finding container 90eab79beadafe7c6d034fbad8cc7e89f7c0052842ddc3b3c0acf01963d3c3ab: Status 404 returned error can't find the container with id 90eab79beadafe7c6d034fbad8cc7e89f7c0052842ddc3b3c0acf01963d3c3ab Jan 21 11:23:45 crc kubenswrapper[4824]: W0121 11:23:45.112691 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod52491e7b_cd8c_403a_942c_312debd6f1c6.slice/crio-512c9c60611f1470288b1906809cbd0a00165c21927d0d39083bb9fb1f9bca05 WatchSource:0}: Error finding container 512c9c60611f1470288b1906809cbd0a00165c21927d0d39083bb9fb1f9bca05: Status 404 returned error can't find the container with id 512c9c60611f1470288b1906809cbd0a00165c21927d0d39083bb9fb1f9bca05 Jan 21 11:23:45 crc kubenswrapper[4824]: E0121 11:23:45.115121 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:244a4906353b84899db16a89e1ebb64491c9f85e69327cb2a72b6da0142a6e5e,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-r4547,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-7cd8bc9dbb-xkp8x_openstack-operators(52491e7b-cd8c-403a-942c-312debd6f1c6): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Jan 21 11:23:45 crc kubenswrapper[4824]: E0121 11:23:45.116651 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/test-operator-controller-manager-7cd8bc9dbb-xkp8x" podUID="52491e7b-cd8c-403a-942c-312debd6f1c6" Jan 21 11:23:45 crc kubenswrapper[4824]: W0121 11:23:45.117576 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0715aa91_2ae5_4e07_8388_c37def4c7743.slice/crio-bbbd6d582790918c7a352ccd6df64b933c752de05fd23d6abbeeefaf76e5fd9d WatchSource:0}: Error finding container bbbd6d582790918c7a352ccd6df64b933c752de05fd23d6abbeeefaf76e5fd9d: Status 404 returned error can't find the container with id bbbd6d582790918c7a352ccd6df64b933c752de05fd23d6abbeeefaf76e5fd9d Jan 21 11:23:45 crc kubenswrapper[4824]: E0121 11:23:45.122534 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/watcher-operator@sha256:d687150a46d97eb382dcd8305a2a611943af74771debe1fa9cc13a21e51c69ad,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-pqc8b,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-64cd966744-9w4ch_openstack-operators(0715aa91-2ae5-4e07-8388-c37def4c7743): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Jan 21 11:23:45 crc kubenswrapper[4824]: E0121 11:23:45.123742 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/watcher-operator-controller-manager-64cd966744-9w4ch" podUID="0715aa91-2ae5-4e07-8388-c37def4c7743" Jan 21 11:23:45 crc kubenswrapper[4824]: I0121 11:23:45.168530 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-94lqx"] Jan 21 11:23:45 crc kubenswrapper[4824]: W0121 11:23:45.175662 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod782093ab_751c_4e38_bba3_c5f76308c82e.slice/crio-82026058c15748e5962d53b16adc2de013ffb9ac4f2a91d6edf426b7e3b96462 WatchSource:0}: Error finding container 82026058c15748e5962d53b16adc2de013ffb9ac4f2a91d6edf426b7e3b96462: Status 404 returned error can't find the container with id 82026058c15748e5962d53b16adc2de013ffb9ac4f2a91d6edf426b7e3b96462 Jan 21 11:23:45 crc kubenswrapper[4824]: E0121 11:23:45.178558 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-5v855,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-668c99d594-94lqx_openstack-operators(782093ab-751c-4e38-bba3-c5f76308c82e): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Jan 21 11:23:45 crc kubenswrapper[4824]: E0121 11:23:45.179694 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-94lqx" podUID="782093ab-751c-4e38-bba3-c5f76308c82e" Jan 21 11:23:45 crc kubenswrapper[4824]: I0121 11:23:45.476255 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/57e1fe13-b2c8-4711-9efe-1ebbc7e9d12e-cert\") pod \"infra-operator-controller-manager-77c48c7859-dzn7w\" (UID: \"57e1fe13-b2c8-4711-9efe-1ebbc7e9d12e\") " pod="openstack-operators/infra-operator-controller-manager-77c48c7859-dzn7w" Jan 21 11:23:45 crc kubenswrapper[4824]: E0121 11:23:45.476450 4824 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Jan 21 11:23:45 crc kubenswrapper[4824]: E0121 11:23:45.476495 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/57e1fe13-b2c8-4711-9efe-1ebbc7e9d12e-cert podName:57e1fe13-b2c8-4711-9efe-1ebbc7e9d12e nodeName:}" failed. No retries permitted until 2026-01-21 11:23:47.476480036 +0000 UTC m=+769.769509328 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/57e1fe13-b2c8-4711-9efe-1ebbc7e9d12e-cert") pod "infra-operator-controller-manager-77c48c7859-dzn7w" (UID: "57e1fe13-b2c8-4711-9efe-1ebbc7e9d12e") : secret "infra-operator-webhook-server-cert" not found Jan 21 11:23:45 crc kubenswrapper[4824]: I0121 11:23:45.579863 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/3fa95966-3147-4e50-8ed8-f702a87992c4-cert\") pod \"openstack-baremetal-operator-controller-manager-6b68b8b854zxcbl\" (UID: \"3fa95966-3147-4e50-8ed8-f702a87992c4\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6b68b8b854zxcbl" Jan 21 11:23:45 crc kubenswrapper[4824]: E0121 11:23:45.580052 4824 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Jan 21 11:23:45 crc kubenswrapper[4824]: E0121 11:23:45.580445 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3fa95966-3147-4e50-8ed8-f702a87992c4-cert podName:3fa95966-3147-4e50-8ed8-f702a87992c4 nodeName:}" failed. No retries permitted until 2026-01-21 11:23:47.580420518 +0000 UTC m=+769.873449810 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/3fa95966-3147-4e50-8ed8-f702a87992c4-cert") pod "openstack-baremetal-operator-controller-manager-6b68b8b854zxcbl" (UID: "3fa95966-3147-4e50-8ed8-f702a87992c4") : secret "openstack-baremetal-operator-webhook-server-cert" not found Jan 21 11:23:45 crc kubenswrapper[4824]: I0121 11:23:45.830934 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-94lqx" event={"ID":"782093ab-751c-4e38-bba3-c5f76308c82e","Type":"ContainerStarted","Data":"82026058c15748e5962d53b16adc2de013ffb9ac4f2a91d6edf426b7e3b96462"} Jan 21 11:23:45 crc kubenswrapper[4824]: I0121 11:23:45.832325 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-686df47fcb-c8xvs" event={"ID":"d8b36106-8f21-4873-890d-c1879ea49068","Type":"ContainerStarted","Data":"db42dc90ff1dd4318f10e5f4c30a19442bab1d9aec8ed8b3ed68b96466cda44b"} Jan 21 11:23:45 crc kubenswrapper[4824]: E0121 11:23:45.833272 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-94lqx" podUID="782093ab-751c-4e38-bba3-c5f76308c82e" Jan 21 11:23:45 crc kubenswrapper[4824]: I0121 11:23:45.836762 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-65849867d6-8ngx6" event={"ID":"b38ac60b-1643-44d9-a1fe-5f75129247dc","Type":"ContainerStarted","Data":"e12efdc25c6f7aebda5d72120c94def99db956f5056672b90c777eeb9cac66d9"} Jan 21 11:23:45 crc kubenswrapper[4824]: E0121 11:23:45.841398 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/nova-operator@sha256:6defa56fc6a5bfbd5b27d28ff7b1c7bc89b24b2ef956e2a6d97b2726f668a231\\\"\"" pod="openstack-operators/nova-operator-controller-manager-65849867d6-8ngx6" podUID="b38ac60b-1643-44d9-a1fe-5f75129247dc" Jan 21 11:23:45 crc kubenswrapper[4824]: I0121 11:23:45.841650 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-85dd56d4cc-wlrdt" event={"ID":"b4cb80e6-c720-4875-9848-7d9597a354e4","Type":"ContainerStarted","Data":"90eab79beadafe7c6d034fbad8cc7e89f7c0052842ddc3b3c0acf01963d3c3ab"} Jan 21 11:23:45 crc kubenswrapper[4824]: I0121 11:23:45.845122 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-c87fff755-d5hn9" event={"ID":"19b1c150-f5ed-4432-bfc2-717b78eebc7e","Type":"ContainerStarted","Data":"5b02d188e4a18fcba9a3114467d976279e50e34fe21b0c02722c3abf1a3e1f52"} Jan 21 11:23:45 crc kubenswrapper[4824]: I0121 11:23:45.848989 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-7fc9b76cf6-tncbx" event={"ID":"5429da27-a618-44da-8aeb-12b917af8734","Type":"ContainerStarted","Data":"103b9e266f3a6c0be180c01b6c107e061eae09a140d6736420e12e6117cc9206"} Jan 21 11:23:45 crc kubenswrapper[4824]: I0121 11:23:45.850385 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-64cd966744-9w4ch" event={"ID":"0715aa91-2ae5-4e07-8388-c37def4c7743","Type":"ContainerStarted","Data":"bbbd6d582790918c7a352ccd6df64b933c752de05fd23d6abbeeefaf76e5fd9d"} Jan 21 11:23:45 crc kubenswrapper[4824]: E0121 11:23:45.851994 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:d687150a46d97eb382dcd8305a2a611943af74771debe1fa9cc13a21e51c69ad\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-64cd966744-9w4ch" podUID="0715aa91-2ae5-4e07-8388-c37def4c7743" Jan 21 11:23:45 crc kubenswrapper[4824]: I0121 11:23:45.852485 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-7cd8bc9dbb-xkp8x" event={"ID":"52491e7b-cd8c-403a-942c-312debd6f1c6","Type":"ContainerStarted","Data":"512c9c60611f1470288b1906809cbd0a00165c21927d0d39083bb9fb1f9bca05"} Jan 21 11:23:45 crc kubenswrapper[4824]: I0121 11:23:45.853465 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-cb4666565-7w8p4" event={"ID":"8c74d3da-d870-45e9-96b9-70eb2395c977","Type":"ContainerStarted","Data":"cc100a8bfa18f73c4ececc9bf8a9ded089e2a55ee7ce7b6379f51ec545977f6f"} Jan 21 11:23:45 crc kubenswrapper[4824]: E0121 11:23:45.856947 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/neutron-operator@sha256:0f440bf7dc937ce0135bdd328716686fd2f1320f453a9ac4e11e96383148ad6c\\\"\"" pod="openstack-operators/neutron-operator-controller-manager-cb4666565-7w8p4" podUID="8c74d3da-d870-45e9-96b9-70eb2395c977" Jan 21 11:23:45 crc kubenswrapper[4824]: E0121 11:23:45.857153 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:244a4906353b84899db16a89e1ebb64491c9f85e69327cb2a72b6da0142a6e5e\\\"\"" pod="openstack-operators/test-operator-controller-manager-7cd8bc9dbb-xkp8x" podUID="52491e7b-cd8c-403a-942c-312debd6f1c6" Jan 21 11:23:45 crc kubenswrapper[4824]: I0121 11:23:45.865070 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-767fdc4f47-5mhmd" event={"ID":"2e5efa9b-40bc-4860-967d-aa37e1838f81","Type":"ContainerStarted","Data":"4c8979cc26d3eca0078286f0469ec6b079f68b32dc1981d65f9cdd1f7d5745db"} Jan 21 11:23:45 crc kubenswrapper[4824]: I0121 11:23:45.867416 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-55db956ddc-78z2q" event={"ID":"f5639f17-3a5b-4ef4-bd6e-6eeb3274636a","Type":"ContainerStarted","Data":"1b41b5017b8d393ecb73e4a44909daf1af252b297e8f80ffd5ee58fbcb0af69b"} Jan 21 11:23:45 crc kubenswrapper[4824]: E0121 11:23:45.868753 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:8b3bfb9e86618b7ac69443939b0968fae28a22cd62ea1e429b599ff9f8a5f8cf\\\"\"" pod="openstack-operators/ovn-operator-controller-manager-55db956ddc-78z2q" podUID="f5639f17-3a5b-4ef4-bd6e-6eeb3274636a" Jan 21 11:23:45 crc kubenswrapper[4824]: I0121 11:23:45.870819 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-5f8f495fcf-6rz5h" event={"ID":"db84b172-59d8-4995-815b-50b492283bee","Type":"ContainerStarted","Data":"dea14bc3dbf16dc26f2bf74918823842c7ee5b7d971485978ab0114af36bdf44"} Jan 21 11:23:45 crc kubenswrapper[4824]: E0121 11:23:45.872522 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:2e89109f5db66abf1afd15ef59bda35a53db40c5e59e020579ac5aa0acea1843\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-5f8f495fcf-6rz5h" podUID="db84b172-59d8-4995-815b-50b492283bee" Jan 21 11:23:45 crc kubenswrapper[4824]: I0121 11:23:45.875597 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-864f6b75bf-c7vsf" event={"ID":"18017a19-1676-4fe9-a97e-ae878f423847","Type":"ContainerStarted","Data":"126e7933e900edc5e407c8d545e0d546bd94935c101798e19e751069062d75fe"} Jan 21 11:23:45 crc kubenswrapper[4824]: I0121 11:23:45.984940 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c1128e01-8b7c-4742-b1da-dbf06aa8b4d8-metrics-certs\") pod \"openstack-operator-controller-manager-74d84f758c-vjgps\" (UID: \"c1128e01-8b7c-4742-b1da-dbf06aa8b4d8\") " pod="openstack-operators/openstack-operator-controller-manager-74d84f758c-vjgps" Jan 21 11:23:45 crc kubenswrapper[4824]: I0121 11:23:45.985029 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/c1128e01-8b7c-4742-b1da-dbf06aa8b4d8-webhook-certs\") pod \"openstack-operator-controller-manager-74d84f758c-vjgps\" (UID: \"c1128e01-8b7c-4742-b1da-dbf06aa8b4d8\") " pod="openstack-operators/openstack-operator-controller-manager-74d84f758c-vjgps" Jan 21 11:23:45 crc kubenswrapper[4824]: E0121 11:23:45.985180 4824 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Jan 21 11:23:45 crc kubenswrapper[4824]: E0121 11:23:45.985337 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c1128e01-8b7c-4742-b1da-dbf06aa8b4d8-metrics-certs podName:c1128e01-8b7c-4742-b1da-dbf06aa8b4d8 nodeName:}" failed. No retries permitted until 2026-01-21 11:23:47.985324487 +0000 UTC m=+770.278353779 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/c1128e01-8b7c-4742-b1da-dbf06aa8b4d8-metrics-certs") pod "openstack-operator-controller-manager-74d84f758c-vjgps" (UID: "c1128e01-8b7c-4742-b1da-dbf06aa8b4d8") : secret "metrics-server-cert" not found Jan 21 11:23:45 crc kubenswrapper[4824]: E0121 11:23:45.985196 4824 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Jan 21 11:23:45 crc kubenswrapper[4824]: E0121 11:23:45.985445 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c1128e01-8b7c-4742-b1da-dbf06aa8b4d8-webhook-certs podName:c1128e01-8b7c-4742-b1da-dbf06aa8b4d8 nodeName:}" failed. No retries permitted until 2026-01-21 11:23:47.985428463 +0000 UTC m=+770.278457755 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/c1128e01-8b7c-4742-b1da-dbf06aa8b4d8-webhook-certs") pod "openstack-operator-controller-manager-74d84f758c-vjgps" (UID: "c1128e01-8b7c-4742-b1da-dbf06aa8b4d8") : secret "webhook-server-cert" not found Jan 21 11:23:46 crc kubenswrapper[4824]: I0121 11:23:46.065365 4824 patch_prober.go:28] interesting pod/machine-config-daemon-zqjg9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 21 11:23:46 crc kubenswrapper[4824]: I0121 11:23:46.065408 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 21 11:23:46 crc kubenswrapper[4824]: I0121 11:23:46.065470 4824 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" Jan 21 11:23:46 crc kubenswrapper[4824]: I0121 11:23:46.065863 4824 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"0c03e9f8dae3ff83ead8b2e38560a1562de223bd29105640743d2bd8e0cdca44"} pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 21 11:23:46 crc kubenswrapper[4824]: I0121 11:23:46.065912 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" containerName="machine-config-daemon" containerID="cri-o://0c03e9f8dae3ff83ead8b2e38560a1562de223bd29105640743d2bd8e0cdca44" gracePeriod=600 Jan 21 11:23:46 crc kubenswrapper[4824]: I0121 11:23:46.894386 4824 generic.go:334] "Generic (PLEG): container finished" podID="33f3d922-4ffe-409b-a49a-d88c85898260" containerID="0c03e9f8dae3ff83ead8b2e38560a1562de223bd29105640743d2bd8e0cdca44" exitCode=0 Jan 21 11:23:46 crc kubenswrapper[4824]: I0121 11:23:46.896208 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" event={"ID":"33f3d922-4ffe-409b-a49a-d88c85898260","Type":"ContainerDied","Data":"0c03e9f8dae3ff83ead8b2e38560a1562de223bd29105640743d2bd8e0cdca44"} Jan 21 11:23:46 crc kubenswrapper[4824]: I0121 11:23:46.898250 4824 scope.go:117] "RemoveContainer" containerID="0494d41399c844b692156a76450e15e5791154f06b98545910f6ac8899bc8509" Jan 21 11:23:46 crc kubenswrapper[4824]: E0121 11:23:46.898372 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/nova-operator@sha256:6defa56fc6a5bfbd5b27d28ff7b1c7bc89b24b2ef956e2a6d97b2726f668a231\\\"\"" pod="openstack-operators/nova-operator-controller-manager-65849867d6-8ngx6" podUID="b38ac60b-1643-44d9-a1fe-5f75129247dc" Jan 21 11:23:46 crc kubenswrapper[4824]: E0121 11:23:46.899190 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/neutron-operator@sha256:0f440bf7dc937ce0135bdd328716686fd2f1320f453a9ac4e11e96383148ad6c\\\"\"" pod="openstack-operators/neutron-operator-controller-manager-cb4666565-7w8p4" podUID="8c74d3da-d870-45e9-96b9-70eb2395c977" Jan 21 11:23:46 crc kubenswrapper[4824]: E0121 11:23:46.909084 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:8b3bfb9e86618b7ac69443939b0968fae28a22cd62ea1e429b599ff9f8a5f8cf\\\"\"" pod="openstack-operators/ovn-operator-controller-manager-55db956ddc-78z2q" podUID="f5639f17-3a5b-4ef4-bd6e-6eeb3274636a" Jan 21 11:23:46 crc kubenswrapper[4824]: E0121 11:23:46.909180 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:d687150a46d97eb382dcd8305a2a611943af74771debe1fa9cc13a21e51c69ad\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-64cd966744-9w4ch" podUID="0715aa91-2ae5-4e07-8388-c37def4c7743" Jan 21 11:23:46 crc kubenswrapper[4824]: E0121 11:23:46.909447 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:244a4906353b84899db16a89e1ebb64491c9f85e69327cb2a72b6da0142a6e5e\\\"\"" pod="openstack-operators/test-operator-controller-manager-7cd8bc9dbb-xkp8x" podUID="52491e7b-cd8c-403a-942c-312debd6f1c6" Jan 21 11:23:46 crc kubenswrapper[4824]: E0121 11:23:46.909508 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:2e89109f5db66abf1afd15ef59bda35a53db40c5e59e020579ac5aa0acea1843\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-5f8f495fcf-6rz5h" podUID="db84b172-59d8-4995-815b-50b492283bee" Jan 21 11:23:46 crc kubenswrapper[4824]: E0121 11:23:46.910639 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-94lqx" podUID="782093ab-751c-4e38-bba3-c5f76308c82e" Jan 21 11:23:47 crc kubenswrapper[4824]: I0121 11:23:47.521577 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/57e1fe13-b2c8-4711-9efe-1ebbc7e9d12e-cert\") pod \"infra-operator-controller-manager-77c48c7859-dzn7w\" (UID: \"57e1fe13-b2c8-4711-9efe-1ebbc7e9d12e\") " pod="openstack-operators/infra-operator-controller-manager-77c48c7859-dzn7w" Jan 21 11:23:47 crc kubenswrapper[4824]: E0121 11:23:47.522046 4824 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Jan 21 11:23:47 crc kubenswrapper[4824]: E0121 11:23:47.522099 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/57e1fe13-b2c8-4711-9efe-1ebbc7e9d12e-cert podName:57e1fe13-b2c8-4711-9efe-1ebbc7e9d12e nodeName:}" failed. No retries permitted until 2026-01-21 11:23:51.522085071 +0000 UTC m=+773.815114364 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/57e1fe13-b2c8-4711-9efe-1ebbc7e9d12e-cert") pod "infra-operator-controller-manager-77c48c7859-dzn7w" (UID: "57e1fe13-b2c8-4711-9efe-1ebbc7e9d12e") : secret "infra-operator-webhook-server-cert" not found Jan 21 11:23:47 crc kubenswrapper[4824]: I0121 11:23:47.622768 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/3fa95966-3147-4e50-8ed8-f702a87992c4-cert\") pod \"openstack-baremetal-operator-controller-manager-6b68b8b854zxcbl\" (UID: \"3fa95966-3147-4e50-8ed8-f702a87992c4\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6b68b8b854zxcbl" Jan 21 11:23:47 crc kubenswrapper[4824]: E0121 11:23:47.624187 4824 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Jan 21 11:23:47 crc kubenswrapper[4824]: E0121 11:23:47.624262 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3fa95966-3147-4e50-8ed8-f702a87992c4-cert podName:3fa95966-3147-4e50-8ed8-f702a87992c4 nodeName:}" failed. No retries permitted until 2026-01-21 11:23:51.624246459 +0000 UTC m=+773.917275752 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/3fa95966-3147-4e50-8ed8-f702a87992c4-cert") pod "openstack-baremetal-operator-controller-manager-6b68b8b854zxcbl" (UID: "3fa95966-3147-4e50-8ed8-f702a87992c4") : secret "openstack-baremetal-operator-webhook-server-cert" not found Jan 21 11:23:48 crc kubenswrapper[4824]: I0121 11:23:48.028418 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c1128e01-8b7c-4742-b1da-dbf06aa8b4d8-metrics-certs\") pod \"openstack-operator-controller-manager-74d84f758c-vjgps\" (UID: \"c1128e01-8b7c-4742-b1da-dbf06aa8b4d8\") " pod="openstack-operators/openstack-operator-controller-manager-74d84f758c-vjgps" Jan 21 11:23:48 crc kubenswrapper[4824]: I0121 11:23:48.028478 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/c1128e01-8b7c-4742-b1da-dbf06aa8b4d8-webhook-certs\") pod \"openstack-operator-controller-manager-74d84f758c-vjgps\" (UID: \"c1128e01-8b7c-4742-b1da-dbf06aa8b4d8\") " pod="openstack-operators/openstack-operator-controller-manager-74d84f758c-vjgps" Jan 21 11:23:48 crc kubenswrapper[4824]: E0121 11:23:48.028565 4824 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Jan 21 11:23:48 crc kubenswrapper[4824]: E0121 11:23:48.028595 4824 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Jan 21 11:23:48 crc kubenswrapper[4824]: E0121 11:23:48.028652 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c1128e01-8b7c-4742-b1da-dbf06aa8b4d8-metrics-certs podName:c1128e01-8b7c-4742-b1da-dbf06aa8b4d8 nodeName:}" failed. No retries permitted until 2026-01-21 11:23:52.02863651 +0000 UTC m=+774.321665802 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/c1128e01-8b7c-4742-b1da-dbf06aa8b4d8-metrics-certs") pod "openstack-operator-controller-manager-74d84f758c-vjgps" (UID: "c1128e01-8b7c-4742-b1da-dbf06aa8b4d8") : secret "metrics-server-cert" not found Jan 21 11:23:48 crc kubenswrapper[4824]: E0121 11:23:48.028687 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c1128e01-8b7c-4742-b1da-dbf06aa8b4d8-webhook-certs podName:c1128e01-8b7c-4742-b1da-dbf06aa8b4d8 nodeName:}" failed. No retries permitted until 2026-01-21 11:23:52.028680323 +0000 UTC m=+774.321709615 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/c1128e01-8b7c-4742-b1da-dbf06aa8b4d8-webhook-certs") pod "openstack-operator-controller-manager-74d84f758c-vjgps" (UID: "c1128e01-8b7c-4742-b1da-dbf06aa8b4d8") : secret "webhook-server-cert" not found Jan 21 11:23:51 crc kubenswrapper[4824]: I0121 11:23:51.571932 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/57e1fe13-b2c8-4711-9efe-1ebbc7e9d12e-cert\") pod \"infra-operator-controller-manager-77c48c7859-dzn7w\" (UID: \"57e1fe13-b2c8-4711-9efe-1ebbc7e9d12e\") " pod="openstack-operators/infra-operator-controller-manager-77c48c7859-dzn7w" Jan 21 11:23:51 crc kubenswrapper[4824]: E0121 11:23:51.572106 4824 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Jan 21 11:23:51 crc kubenswrapper[4824]: E0121 11:23:51.572740 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/57e1fe13-b2c8-4711-9efe-1ebbc7e9d12e-cert podName:57e1fe13-b2c8-4711-9efe-1ebbc7e9d12e nodeName:}" failed. No retries permitted until 2026-01-21 11:23:59.572712107 +0000 UTC m=+781.865741399 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/57e1fe13-b2c8-4711-9efe-1ebbc7e9d12e-cert") pod "infra-operator-controller-manager-77c48c7859-dzn7w" (UID: "57e1fe13-b2c8-4711-9efe-1ebbc7e9d12e") : secret "infra-operator-webhook-server-cert" not found Jan 21 11:23:51 crc kubenswrapper[4824]: I0121 11:23:51.673870 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/3fa95966-3147-4e50-8ed8-f702a87992c4-cert\") pod \"openstack-baremetal-operator-controller-manager-6b68b8b854zxcbl\" (UID: \"3fa95966-3147-4e50-8ed8-f702a87992c4\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6b68b8b854zxcbl" Jan 21 11:23:51 crc kubenswrapper[4824]: E0121 11:23:51.674058 4824 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Jan 21 11:23:51 crc kubenswrapper[4824]: E0121 11:23:51.674183 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3fa95966-3147-4e50-8ed8-f702a87992c4-cert podName:3fa95966-3147-4e50-8ed8-f702a87992c4 nodeName:}" failed. No retries permitted until 2026-01-21 11:23:59.67416568 +0000 UTC m=+781.967194973 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/3fa95966-3147-4e50-8ed8-f702a87992c4-cert") pod "openstack-baremetal-operator-controller-manager-6b68b8b854zxcbl" (UID: "3fa95966-3147-4e50-8ed8-f702a87992c4") : secret "openstack-baremetal-operator-webhook-server-cert" not found Jan 21 11:23:52 crc kubenswrapper[4824]: I0121 11:23:52.079221 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c1128e01-8b7c-4742-b1da-dbf06aa8b4d8-metrics-certs\") pod \"openstack-operator-controller-manager-74d84f758c-vjgps\" (UID: \"c1128e01-8b7c-4742-b1da-dbf06aa8b4d8\") " pod="openstack-operators/openstack-operator-controller-manager-74d84f758c-vjgps" Jan 21 11:23:52 crc kubenswrapper[4824]: I0121 11:23:52.079280 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/c1128e01-8b7c-4742-b1da-dbf06aa8b4d8-webhook-certs\") pod \"openstack-operator-controller-manager-74d84f758c-vjgps\" (UID: \"c1128e01-8b7c-4742-b1da-dbf06aa8b4d8\") " pod="openstack-operators/openstack-operator-controller-manager-74d84f758c-vjgps" Jan 21 11:23:52 crc kubenswrapper[4824]: E0121 11:23:52.079345 4824 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Jan 21 11:23:52 crc kubenswrapper[4824]: E0121 11:23:52.079420 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c1128e01-8b7c-4742-b1da-dbf06aa8b4d8-metrics-certs podName:c1128e01-8b7c-4742-b1da-dbf06aa8b4d8 nodeName:}" failed. No retries permitted until 2026-01-21 11:24:00.079404682 +0000 UTC m=+782.372433974 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/c1128e01-8b7c-4742-b1da-dbf06aa8b4d8-metrics-certs") pod "openstack-operator-controller-manager-74d84f758c-vjgps" (UID: "c1128e01-8b7c-4742-b1da-dbf06aa8b4d8") : secret "metrics-server-cert" not found Jan 21 11:23:52 crc kubenswrapper[4824]: E0121 11:23:52.079437 4824 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Jan 21 11:23:52 crc kubenswrapper[4824]: E0121 11:23:52.079510 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c1128e01-8b7c-4742-b1da-dbf06aa8b4d8-webhook-certs podName:c1128e01-8b7c-4742-b1da-dbf06aa8b4d8 nodeName:}" failed. No retries permitted until 2026-01-21 11:24:00.079495513 +0000 UTC m=+782.372524805 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/c1128e01-8b7c-4742-b1da-dbf06aa8b4d8-webhook-certs") pod "openstack-operator-controller-manager-74d84f758c-vjgps" (UID: "c1128e01-8b7c-4742-b1da-dbf06aa8b4d8") : secret "webhook-server-cert" not found Jan 21 11:23:53 crc kubenswrapper[4824]: I0121 11:23:53.935016 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" event={"ID":"33f3d922-4ffe-409b-a49a-d88c85898260","Type":"ContainerStarted","Data":"365550d435931ba3a56d32c848ae2c97f8d7721cf74faadf03f5fc5d15e5b5a5"} Jan 21 11:23:54 crc kubenswrapper[4824]: I0121 11:23:54.941275 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-9f958b845-wtfj4" event={"ID":"14dd1f12-3943-42e4-be1b-c9e37e49d21b","Type":"ContainerStarted","Data":"0144a260ffe05018da1da674ab01673a24253513ff57e1761613aabaa45aa079"} Jan 21 11:23:54 crc kubenswrapper[4824]: I0121 11:23:54.942646 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-9f958b845-wtfj4" Jan 21 11:23:54 crc kubenswrapper[4824]: I0121 11:23:54.943873 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-c6994669c-hvq6s" event={"ID":"b7053a68-ce89-48b6-9b5d-1c6dac813a75","Type":"ContainerStarted","Data":"cdd57094082d6f2ee787569412482c5180a1b01f0caadb654d941d17800f292d"} Jan 21 11:23:54 crc kubenswrapper[4824]: I0121 11:23:54.944304 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-c6994669c-hvq6s" Jan 21 11:23:54 crc kubenswrapper[4824]: I0121 11:23:54.945538 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-85dd56d4cc-wlrdt" event={"ID":"b4cb80e6-c720-4875-9848-7d9597a354e4","Type":"ContainerStarted","Data":"b4597e6eec1e1001b39c732a9a2649db63b32895745dbe0933eff0a933c598ca"} Jan 21 11:23:54 crc kubenswrapper[4824]: I0121 11:23:54.945902 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-85dd56d4cc-wlrdt" Jan 21 11:23:54 crc kubenswrapper[4824]: I0121 11:23:54.952091 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-c87fff755-d5hn9" event={"ID":"19b1c150-f5ed-4432-bfc2-717b78eebc7e","Type":"ContainerStarted","Data":"01d9913b9bf0a1793eef82f6062ebc37cd27c2977db578f223abb27c36c5c629"} Jan 21 11:23:54 crc kubenswrapper[4824]: I0121 11:23:54.952208 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-c87fff755-d5hn9" Jan 21 11:23:54 crc kubenswrapper[4824]: I0121 11:23:54.955578 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-9f958b845-wtfj4" podStartSLOduration=3.134284007 podStartE2EDuration="11.955569696s" podCreationTimestamp="2026-01-21 11:23:43 +0000 UTC" firstStartedPulling="2026-01-21 11:23:44.459904592 +0000 UTC m=+766.752933884" lastFinishedPulling="2026-01-21 11:23:53.281190282 +0000 UTC m=+775.574219573" observedRunningTime="2026-01-21 11:23:54.954534193 +0000 UTC m=+777.247563486" watchObservedRunningTime="2026-01-21 11:23:54.955569696 +0000 UTC m=+777.248598988" Jan 21 11:23:54 crc kubenswrapper[4824]: I0121 11:23:54.956351 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-7fc9b76cf6-tncbx" event={"ID":"5429da27-a618-44da-8aeb-12b917af8734","Type":"ContainerStarted","Data":"7413edd5c109e85ef6bac3e1732ee07186fda80dd976c15abc1f710b14a3e5d3"} Jan 21 11:23:54 crc kubenswrapper[4824]: I0121 11:23:54.956414 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-7fc9b76cf6-tncbx" Jan 21 11:23:54 crc kubenswrapper[4824]: I0121 11:23:54.957492 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-9b68f5989-24r54" event={"ID":"a25733d4-a5f8-4c97-a897-8d96e637c253","Type":"ContainerStarted","Data":"fbc4fa10919d81e093dfc0ac0717ea73b800f34bdefab338d33815f6ba05d1cf"} Jan 21 11:23:54 crc kubenswrapper[4824]: I0121 11:23:54.957914 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-9b68f5989-24r54" Jan 21 11:23:54 crc kubenswrapper[4824]: I0121 11:23:54.959061 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-78757b4889-gkwjs" event={"ID":"fc911701-7c15-4a20-b89e-709521bdd3a1","Type":"ContainerStarted","Data":"3f749911d19fa1accdb0221d5a5e56f9cb511c2178b4f5396dcd7a3cef2360fc"} Jan 21 11:23:54 crc kubenswrapper[4824]: I0121 11:23:54.959412 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-78757b4889-gkwjs" Jan 21 11:23:54 crc kubenswrapper[4824]: I0121 11:23:54.961026 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7ddb5c749-bvptq" event={"ID":"8ed6bb2c-2770-4785-bf38-9b6fd0bd8472","Type":"ContainerStarted","Data":"c551da34bdb35bedc6b251ecd69aeb95220f8fbbd35cab2305ae053fe28ccdaa"} Jan 21 11:23:54 crc kubenswrapper[4824]: I0121 11:23:54.961139 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-7ddb5c749-bvptq" Jan 21 11:23:54 crc kubenswrapper[4824]: I0121 11:23:54.962916 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-686df47fcb-c8xvs" event={"ID":"d8b36106-8f21-4873-890d-c1879ea49068","Type":"ContainerStarted","Data":"44941f67f8f6a783521e8647554e0de989687e5da65df0495c30f6faec8d8be4"} Jan 21 11:23:54 crc kubenswrapper[4824]: I0121 11:23:54.963358 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-686df47fcb-c8xvs" Jan 21 11:23:54 crc kubenswrapper[4824]: I0121 11:23:54.964830 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5fbc59b5b6-hgmf4" event={"ID":"944c3123-9063-4731-9581-5b683741edc4","Type":"ContainerStarted","Data":"d117636860c96b58d7ff7b5b600b9fa9a0115fe5fd3291c15cd27ab7fef0b75b"} Jan 21 11:23:54 crc kubenswrapper[4824]: I0121 11:23:54.965183 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-5fbc59b5b6-hgmf4" Jan 21 11:23:54 crc kubenswrapper[4824]: I0121 11:23:54.966397 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-864f6b75bf-c7vsf" event={"ID":"18017a19-1676-4fe9-a97e-ae878f423847","Type":"ContainerStarted","Data":"8098b147d4a88222e038826914e86df6eb153544fedaeaa10b9c65d308469e86"} Jan 21 11:23:54 crc kubenswrapper[4824]: I0121 11:23:54.967015 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-864f6b75bf-c7vsf" Jan 21 11:23:54 crc kubenswrapper[4824]: I0121 11:23:54.968533 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-77d5c5b54f-mp5ml" event={"ID":"2463afb1-0271-4ab3-ad1a-9329ecaeec4d","Type":"ContainerStarted","Data":"7a27f2bda3a9bcde11636fb20ef32586d8b71da26ed80ee6abdef4ed31a6ef22"} Jan 21 11:23:54 crc kubenswrapper[4824]: I0121 11:23:54.968917 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-77d5c5b54f-mp5ml" Jan 21 11:23:54 crc kubenswrapper[4824]: I0121 11:23:54.971459 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-c6994669c-hvq6s" podStartSLOduration=2.910498731 podStartE2EDuration="11.971448419s" podCreationTimestamp="2026-01-21 11:23:43 +0000 UTC" firstStartedPulling="2026-01-21 11:23:44.68150545 +0000 UTC m=+766.974534743" lastFinishedPulling="2026-01-21 11:23:53.742455139 +0000 UTC m=+776.035484431" observedRunningTime="2026-01-21 11:23:54.966301076 +0000 UTC m=+777.259330369" watchObservedRunningTime="2026-01-21 11:23:54.971448419 +0000 UTC m=+777.264477711" Jan 21 11:23:54 crc kubenswrapper[4824]: I0121 11:23:54.976996 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-767fdc4f47-5mhmd" event={"ID":"2e5efa9b-40bc-4860-967d-aa37e1838f81","Type":"ContainerStarted","Data":"28d6b089a1084e2e9c3b6649718b08f0745f019a6d9501a71029677fabfb5b08"} Jan 21 11:23:54 crc kubenswrapper[4824]: I0121 11:23:54.977029 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-767fdc4f47-5mhmd" Jan 21 11:23:54 crc kubenswrapper[4824]: I0121 11:23:54.999516 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-85dd56d4cc-wlrdt" podStartSLOduration=3.370033267 podStartE2EDuration="11.999503578s" podCreationTimestamp="2026-01-21 11:23:43 +0000 UTC" firstStartedPulling="2026-01-21 11:23:45.113891367 +0000 UTC m=+767.406920659" lastFinishedPulling="2026-01-21 11:23:53.743361678 +0000 UTC m=+776.036390970" observedRunningTime="2026-01-21 11:23:54.997885567 +0000 UTC m=+777.290914859" watchObservedRunningTime="2026-01-21 11:23:54.999503578 +0000 UTC m=+777.292532859" Jan 21 11:23:55 crc kubenswrapper[4824]: I0121 11:23:55.023614 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-c87fff755-d5hn9" podStartSLOduration=3.676806274 podStartE2EDuration="12.023602667s" podCreationTimestamp="2026-01-21 11:23:43 +0000 UTC" firstStartedPulling="2026-01-21 11:23:44.934582986 +0000 UTC m=+767.227612277" lastFinishedPulling="2026-01-21 11:23:53.281379378 +0000 UTC m=+775.574408670" observedRunningTime="2026-01-21 11:23:55.019018385 +0000 UTC m=+777.312047677" watchObservedRunningTime="2026-01-21 11:23:55.023602667 +0000 UTC m=+777.316631959" Jan 21 11:23:55 crc kubenswrapper[4824]: I0121 11:23:55.041738 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-77d5c5b54f-mp5ml" podStartSLOduration=3.705731862 podStartE2EDuration="12.041709461s" podCreationTimestamp="2026-01-21 11:23:43 +0000 UTC" firstStartedPulling="2026-01-21 11:23:44.530070946 +0000 UTC m=+766.823100237" lastFinishedPulling="2026-01-21 11:23:52.866048543 +0000 UTC m=+775.159077836" observedRunningTime="2026-01-21 11:23:55.038025376 +0000 UTC m=+777.331054668" watchObservedRunningTime="2026-01-21 11:23:55.041709461 +0000 UTC m=+777.334738753" Jan 21 11:23:55 crc kubenswrapper[4824]: I0121 11:23:55.060681 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-864f6b75bf-c7vsf" podStartSLOduration=3.244062133 podStartE2EDuration="12.060669373s" podCreationTimestamp="2026-01-21 11:23:43 +0000 UTC" firstStartedPulling="2026-01-21 11:23:44.931192154 +0000 UTC m=+767.224221446" lastFinishedPulling="2026-01-21 11:23:53.747799394 +0000 UTC m=+776.040828686" observedRunningTime="2026-01-21 11:23:55.056754693 +0000 UTC m=+777.349783985" watchObservedRunningTime="2026-01-21 11:23:55.060669373 +0000 UTC m=+777.353698666" Jan 21 11:23:55 crc kubenswrapper[4824]: I0121 11:23:55.081815 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-7fc9b76cf6-tncbx" podStartSLOduration=3.28721898 podStartE2EDuration="12.08180101s" podCreationTimestamp="2026-01-21 11:23:43 +0000 UTC" firstStartedPulling="2026-01-21 11:23:45.000416381 +0000 UTC m=+767.293445683" lastFinishedPulling="2026-01-21 11:23:53.79499842 +0000 UTC m=+776.088027713" observedRunningTime="2026-01-21 11:23:55.077589991 +0000 UTC m=+777.370619283" watchObservedRunningTime="2026-01-21 11:23:55.08180101 +0000 UTC m=+777.374830302" Jan 21 11:23:55 crc kubenswrapper[4824]: I0121 11:23:55.101536 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-686df47fcb-c8xvs" podStartSLOduration=3.257009882 podStartE2EDuration="12.101523719s" podCreationTimestamp="2026-01-21 11:23:43 +0000 UTC" firstStartedPulling="2026-01-21 11:23:44.897869365 +0000 UTC m=+767.190898658" lastFinishedPulling="2026-01-21 11:23:53.742383203 +0000 UTC m=+776.035412495" observedRunningTime="2026-01-21 11:23:55.097224364 +0000 UTC m=+777.390253656" watchObservedRunningTime="2026-01-21 11:23:55.101523719 +0000 UTC m=+777.394553011" Jan 21 11:23:55 crc kubenswrapper[4824]: I0121 11:23:55.115423 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-78757b4889-gkwjs" podStartSLOduration=3.049022588 podStartE2EDuration="12.115412041s" podCreationTimestamp="2026-01-21 11:23:43 +0000 UTC" firstStartedPulling="2026-01-21 11:23:44.675498967 +0000 UTC m=+766.968528260" lastFinishedPulling="2026-01-21 11:23:53.741888421 +0000 UTC m=+776.034917713" observedRunningTime="2026-01-21 11:23:55.111886905 +0000 UTC m=+777.404916197" watchObservedRunningTime="2026-01-21 11:23:55.115412041 +0000 UTC m=+777.408441332" Jan 21 11:23:55 crc kubenswrapper[4824]: I0121 11:23:55.130841 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-7ddb5c749-bvptq" podStartSLOduration=2.817660107 podStartE2EDuration="12.13082743s" podCreationTimestamp="2026-01-21 11:23:43 +0000 UTC" firstStartedPulling="2026-01-21 11:23:44.428566245 +0000 UTC m=+766.721595538" lastFinishedPulling="2026-01-21 11:23:53.741733569 +0000 UTC m=+776.034762861" observedRunningTime="2026-01-21 11:23:55.129733519 +0000 UTC m=+777.422762810" watchObservedRunningTime="2026-01-21 11:23:55.13082743 +0000 UTC m=+777.423856723" Jan 21 11:23:55 crc kubenswrapper[4824]: I0121 11:23:55.141927 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-9b68f5989-24r54" podStartSLOduration=3.741001391 podStartE2EDuration="12.141915312s" podCreationTimestamp="2026-01-21 11:23:43 +0000 UTC" firstStartedPulling="2026-01-21 11:23:44.453782671 +0000 UTC m=+766.746811964" lastFinishedPulling="2026-01-21 11:23:52.854696593 +0000 UTC m=+775.147725885" observedRunningTime="2026-01-21 11:23:55.141682444 +0000 UTC m=+777.434711735" watchObservedRunningTime="2026-01-21 11:23:55.141915312 +0000 UTC m=+777.434944604" Jan 21 11:23:55 crc kubenswrapper[4824]: I0121 11:23:55.160906 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-5fbc59b5b6-hgmf4" podStartSLOduration=3.032802018 podStartE2EDuration="12.16089421s" podCreationTimestamp="2026-01-21 11:23:43 +0000 UTC" firstStartedPulling="2026-01-21 11:23:44.675232235 +0000 UTC m=+766.968261516" lastFinishedPulling="2026-01-21 11:23:53.803324416 +0000 UTC m=+776.096353708" observedRunningTime="2026-01-21 11:23:55.160131944 +0000 UTC m=+777.453161235" watchObservedRunningTime="2026-01-21 11:23:55.16089421 +0000 UTC m=+777.453923502" Jan 21 11:23:55 crc kubenswrapper[4824]: I0121 11:23:55.173347 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-767fdc4f47-5mhmd" podStartSLOduration=3.207844527 podStartE2EDuration="12.173335624s" podCreationTimestamp="2026-01-21 11:23:43 +0000 UTC" firstStartedPulling="2026-01-21 11:23:44.79080212 +0000 UTC m=+767.083831402" lastFinishedPulling="2026-01-21 11:23:53.756293206 +0000 UTC m=+776.049322499" observedRunningTime="2026-01-21 11:23:55.17291466 +0000 UTC m=+777.465943962" watchObservedRunningTime="2026-01-21 11:23:55.173335624 +0000 UTC m=+777.466364916" Jan 21 11:23:59 crc kubenswrapper[4824]: I0121 11:23:59.584329 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/57e1fe13-b2c8-4711-9efe-1ebbc7e9d12e-cert\") pod \"infra-operator-controller-manager-77c48c7859-dzn7w\" (UID: \"57e1fe13-b2c8-4711-9efe-1ebbc7e9d12e\") " pod="openstack-operators/infra-operator-controller-manager-77c48c7859-dzn7w" Jan 21 11:23:59 crc kubenswrapper[4824]: I0121 11:23:59.589251 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/57e1fe13-b2c8-4711-9efe-1ebbc7e9d12e-cert\") pod \"infra-operator-controller-manager-77c48c7859-dzn7w\" (UID: \"57e1fe13-b2c8-4711-9efe-1ebbc7e9d12e\") " pod="openstack-operators/infra-operator-controller-manager-77c48c7859-dzn7w" Jan 21 11:23:59 crc kubenswrapper[4824]: I0121 11:23:59.686198 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/3fa95966-3147-4e50-8ed8-f702a87992c4-cert\") pod \"openstack-baremetal-operator-controller-manager-6b68b8b854zxcbl\" (UID: \"3fa95966-3147-4e50-8ed8-f702a87992c4\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6b68b8b854zxcbl" Jan 21 11:23:59 crc kubenswrapper[4824]: I0121 11:23:59.689586 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/3fa95966-3147-4e50-8ed8-f702a87992c4-cert\") pod \"openstack-baremetal-operator-controller-manager-6b68b8b854zxcbl\" (UID: \"3fa95966-3147-4e50-8ed8-f702a87992c4\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6b68b8b854zxcbl" Jan 21 11:23:59 crc kubenswrapper[4824]: I0121 11:23:59.753970 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-ct526" Jan 21 11:23:59 crc kubenswrapper[4824]: I0121 11:23:59.763003 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6b68b8b854zxcbl" Jan 21 11:23:59 crc kubenswrapper[4824]: I0121 11:23:59.858989 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-ldsjj" Jan 21 11:23:59 crc kubenswrapper[4824]: I0121 11:23:59.867169 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-77c48c7859-dzn7w" Jan 21 11:24:00 crc kubenswrapper[4824]: I0121 11:24:00.004369 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-5f8f495fcf-6rz5h" event={"ID":"db84b172-59d8-4995-815b-50b492283bee","Type":"ContainerStarted","Data":"d2c19ea693d44ba91ea907492ee970b2352043e1ba248842c926563c68d5ba97"} Jan 21 11:24:00 crc kubenswrapper[4824]: I0121 11:24:00.004767 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-5f8f495fcf-6rz5h" Jan 21 11:24:00 crc kubenswrapper[4824]: I0121 11:24:00.020062 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-5f8f495fcf-6rz5h" podStartSLOduration=2.593054925 podStartE2EDuration="17.020049241s" podCreationTimestamp="2026-01-21 11:23:43 +0000 UTC" firstStartedPulling="2026-01-21 11:23:45.005765555 +0000 UTC m=+767.298794847" lastFinishedPulling="2026-01-21 11:23:59.432759872 +0000 UTC m=+781.725789163" observedRunningTime="2026-01-21 11:24:00.015951146 +0000 UTC m=+782.308980438" watchObservedRunningTime="2026-01-21 11:24:00.020049241 +0000 UTC m=+782.313078533" Jan 21 11:24:00 crc kubenswrapper[4824]: I0121 11:24:00.090475 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c1128e01-8b7c-4742-b1da-dbf06aa8b4d8-metrics-certs\") pod \"openstack-operator-controller-manager-74d84f758c-vjgps\" (UID: \"c1128e01-8b7c-4742-b1da-dbf06aa8b4d8\") " pod="openstack-operators/openstack-operator-controller-manager-74d84f758c-vjgps" Jan 21 11:24:00 crc kubenswrapper[4824]: I0121 11:24:00.090530 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/c1128e01-8b7c-4742-b1da-dbf06aa8b4d8-webhook-certs\") pod \"openstack-operator-controller-manager-74d84f758c-vjgps\" (UID: \"c1128e01-8b7c-4742-b1da-dbf06aa8b4d8\") " pod="openstack-operators/openstack-operator-controller-manager-74d84f758c-vjgps" Jan 21 11:24:00 crc kubenswrapper[4824]: E0121 11:24:00.090631 4824 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Jan 21 11:24:00 crc kubenswrapper[4824]: E0121 11:24:00.090688 4824 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Jan 21 11:24:00 crc kubenswrapper[4824]: E0121 11:24:00.090698 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c1128e01-8b7c-4742-b1da-dbf06aa8b4d8-metrics-certs podName:c1128e01-8b7c-4742-b1da-dbf06aa8b4d8 nodeName:}" failed. No retries permitted until 2026-01-21 11:24:16.090667363 +0000 UTC m=+798.383696655 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/c1128e01-8b7c-4742-b1da-dbf06aa8b4d8-metrics-certs") pod "openstack-operator-controller-manager-74d84f758c-vjgps" (UID: "c1128e01-8b7c-4742-b1da-dbf06aa8b4d8") : secret "metrics-server-cert" not found Jan 21 11:24:00 crc kubenswrapper[4824]: E0121 11:24:00.090798 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c1128e01-8b7c-4742-b1da-dbf06aa8b4d8-webhook-certs podName:c1128e01-8b7c-4742-b1da-dbf06aa8b4d8 nodeName:}" failed. No retries permitted until 2026-01-21 11:24:16.090766269 +0000 UTC m=+798.383795561 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/c1128e01-8b7c-4742-b1da-dbf06aa8b4d8-webhook-certs") pod "openstack-operator-controller-manager-74d84f758c-vjgps" (UID: "c1128e01-8b7c-4742-b1da-dbf06aa8b4d8") : secret "webhook-server-cert" not found Jan 21 11:24:00 crc kubenswrapper[4824]: I0121 11:24:00.134732 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-6b68b8b854zxcbl"] Jan 21 11:24:00 crc kubenswrapper[4824]: W0121 11:24:00.140728 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3fa95966_3147_4e50_8ed8_f702a87992c4.slice/crio-eed19a8e42f8cee6fa76d0869d23edca90ebfc224ff7958bd7850ad2a22a3286 WatchSource:0}: Error finding container eed19a8e42f8cee6fa76d0869d23edca90ebfc224ff7958bd7850ad2a22a3286: Status 404 returned error can't find the container with id eed19a8e42f8cee6fa76d0869d23edca90ebfc224ff7958bd7850ad2a22a3286 Jan 21 11:24:00 crc kubenswrapper[4824]: I0121 11:24:00.228752 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-77c48c7859-dzn7w"] Jan 21 11:24:00 crc kubenswrapper[4824]: W0121 11:24:00.232626 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod57e1fe13_b2c8_4711_9efe_1ebbc7e9d12e.slice/crio-1fdcea483970f40088fb49b697cb5f962853a54895fc1319cad82d4dfde02a2f WatchSource:0}: Error finding container 1fdcea483970f40088fb49b697cb5f962853a54895fc1319cad82d4dfde02a2f: Status 404 returned error can't find the container with id 1fdcea483970f40088fb49b697cb5f962853a54895fc1319cad82d4dfde02a2f Jan 21 11:24:01 crc kubenswrapper[4824]: I0121 11:24:01.014197 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6b68b8b854zxcbl" event={"ID":"3fa95966-3147-4e50-8ed8-f702a87992c4","Type":"ContainerStarted","Data":"eed19a8e42f8cee6fa76d0869d23edca90ebfc224ff7958bd7850ad2a22a3286"} Jan 21 11:24:01 crc kubenswrapper[4824]: I0121 11:24:01.017881 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-77c48c7859-dzn7w" event={"ID":"57e1fe13-b2c8-4711-9efe-1ebbc7e9d12e","Type":"ContainerStarted","Data":"1fdcea483970f40088fb49b697cb5f962853a54895fc1319cad82d4dfde02a2f"} Jan 21 11:24:01 crc kubenswrapper[4824]: I0121 11:24:01.019892 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-65849867d6-8ngx6" event={"ID":"b38ac60b-1643-44d9-a1fe-5f75129247dc","Type":"ContainerStarted","Data":"dab60b48ed9b2e82be6ad6709abeac7a43f8f3c164163b277dd4c72e10eff626"} Jan 21 11:24:01 crc kubenswrapper[4824]: I0121 11:24:01.021433 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-65849867d6-8ngx6" Jan 21 11:24:01 crc kubenswrapper[4824]: I0121 11:24:01.034045 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-65849867d6-8ngx6" podStartSLOduration=2.377611408 podStartE2EDuration="18.034029617s" podCreationTimestamp="2026-01-21 11:23:43 +0000 UTC" firstStartedPulling="2026-01-21 11:23:45.003380289 +0000 UTC m=+767.296409581" lastFinishedPulling="2026-01-21 11:24:00.659798498 +0000 UTC m=+782.952827790" observedRunningTime="2026-01-21 11:24:01.033572825 +0000 UTC m=+783.326602108" watchObservedRunningTime="2026-01-21 11:24:01.034029617 +0000 UTC m=+783.327058908" Jan 21 11:24:03 crc kubenswrapper[4824]: I0121 11:24:03.039127 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-77c48c7859-dzn7w" event={"ID":"57e1fe13-b2c8-4711-9efe-1ebbc7e9d12e","Type":"ContainerStarted","Data":"6f712b0e06905ad53625e13daa538506c186e3d2f91f50bf50f6d5e64d5de21d"} Jan 21 11:24:03 crc kubenswrapper[4824]: I0121 11:24:03.040287 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-77c48c7859-dzn7w" Jan 21 11:24:03 crc kubenswrapper[4824]: I0121 11:24:03.043709 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-55db956ddc-78z2q" event={"ID":"f5639f17-3a5b-4ef4-bd6e-6eeb3274636a","Type":"ContainerStarted","Data":"88e327fb8ddda307734a971cf283fd5be51c0aecdfa49eb1437c82a96f083bf9"} Jan 21 11:24:03 crc kubenswrapper[4824]: I0121 11:24:03.044093 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-55db956ddc-78z2q" Jan 21 11:24:03 crc kubenswrapper[4824]: I0121 11:24:03.044981 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6b68b8b854zxcbl" event={"ID":"3fa95966-3147-4e50-8ed8-f702a87992c4","Type":"ContainerStarted","Data":"4b62ac54cccbebffb3336a2120bc91a3453bdc07a6c7b190cbcaa9142b158d2e"} Jan 21 11:24:03 crc kubenswrapper[4824]: I0121 11:24:03.045304 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6b68b8b854zxcbl" Jan 21 11:24:03 crc kubenswrapper[4824]: I0121 11:24:03.057305 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-77c48c7859-dzn7w" podStartSLOduration=17.588435571 podStartE2EDuration="20.057294877s" podCreationTimestamp="2026-01-21 11:23:43 +0000 UTC" firstStartedPulling="2026-01-21 11:24:00.234661875 +0000 UTC m=+782.527691167" lastFinishedPulling="2026-01-21 11:24:02.70352118 +0000 UTC m=+784.996550473" observedRunningTime="2026-01-21 11:24:03.05209682 +0000 UTC m=+785.345126111" watchObservedRunningTime="2026-01-21 11:24:03.057294877 +0000 UTC m=+785.350324169" Jan 21 11:24:03 crc kubenswrapper[4824]: I0121 11:24:03.075226 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-55db956ddc-78z2q" podStartSLOduration=2.380820997 podStartE2EDuration="20.075210496s" podCreationTimestamp="2026-01-21 11:23:43 +0000 UTC" firstStartedPulling="2026-01-21 11:23:45.008271108 +0000 UTC m=+767.301300401" lastFinishedPulling="2026-01-21 11:24:02.702660608 +0000 UTC m=+784.995689900" observedRunningTime="2026-01-21 11:24:03.064194763 +0000 UTC m=+785.357224055" watchObservedRunningTime="2026-01-21 11:24:03.075210496 +0000 UTC m=+785.368239789" Jan 21 11:24:03 crc kubenswrapper[4824]: I0121 11:24:03.090178 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6b68b8b854zxcbl" podStartSLOduration=17.509007375 podStartE2EDuration="20.090159786s" podCreationTimestamp="2026-01-21 11:23:43 +0000 UTC" firstStartedPulling="2026-01-21 11:24:00.145979694 +0000 UTC m=+782.439008986" lastFinishedPulling="2026-01-21 11:24:02.727132105 +0000 UTC m=+785.020161397" observedRunningTime="2026-01-21 11:24:03.084979351 +0000 UTC m=+785.378008643" watchObservedRunningTime="2026-01-21 11:24:03.090159786 +0000 UTC m=+785.383189078" Jan 21 11:24:03 crc kubenswrapper[4824]: I0121 11:24:03.868645 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-7ddb5c749-bvptq" Jan 21 11:24:03 crc kubenswrapper[4824]: I0121 11:24:03.878531 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-9b68f5989-24r54" Jan 21 11:24:03 crc kubenswrapper[4824]: I0121 11:24:03.896648 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-9f958b845-wtfj4" Jan 21 11:24:03 crc kubenswrapper[4824]: I0121 11:24:03.906583 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-c6994669c-hvq6s" Jan 21 11:24:03 crc kubenswrapper[4824]: I0121 11:24:03.925027 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-77d5c5b54f-mp5ml" Jan 21 11:24:03 crc kubenswrapper[4824]: I0121 11:24:03.935606 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-5fbc59b5b6-hgmf4" Jan 21 11:24:03 crc kubenswrapper[4824]: I0121 11:24:03.978655 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-78757b4889-gkwjs" Jan 21 11:24:04 crc kubenswrapper[4824]: I0121 11:24:04.024109 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-767fdc4f47-5mhmd" Jan 21 11:24:04 crc kubenswrapper[4824]: I0121 11:24:04.077535 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-c87fff755-d5hn9" Jan 21 11:24:04 crc kubenswrapper[4824]: I0121 11:24:04.085784 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-864f6b75bf-c7vsf" Jan 21 11:24:04 crc kubenswrapper[4824]: I0121 11:24:04.146274 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-7fc9b76cf6-tncbx" Jan 21 11:24:04 crc kubenswrapper[4824]: I0121 11:24:04.172459 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-686df47fcb-c8xvs" Jan 21 11:24:04 crc kubenswrapper[4824]: I0121 11:24:04.293915 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-5f8f495fcf-6rz5h" Jan 21 11:24:04 crc kubenswrapper[4824]: I0121 11:24:04.483297 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-85dd56d4cc-wlrdt" Jan 21 11:24:05 crc kubenswrapper[4824]: I0121 11:24:05.057516 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-7cd8bc9dbb-xkp8x" event={"ID":"52491e7b-cd8c-403a-942c-312debd6f1c6","Type":"ContainerStarted","Data":"1dc24aac38479c7e94c84527a6d3167ce426dc8f5e9c76700dd730953f534c97"} Jan 21 11:24:05 crc kubenswrapper[4824]: I0121 11:24:05.057647 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-7cd8bc9dbb-xkp8x" Jan 21 11:24:05 crc kubenswrapper[4824]: I0121 11:24:05.058811 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-64cd966744-9w4ch" event={"ID":"0715aa91-2ae5-4e07-8388-c37def4c7743","Type":"ContainerStarted","Data":"1f098a0e7196ad5933f6d0820f01f162fefc145fc5caad398e23c286893eae93"} Jan 21 11:24:05 crc kubenswrapper[4824]: I0121 11:24:05.059110 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-64cd966744-9w4ch" Jan 21 11:24:05 crc kubenswrapper[4824]: I0121 11:24:05.072154 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-7cd8bc9dbb-xkp8x" podStartSLOduration=1.7818485069999999 podStartE2EDuration="21.072138393s" podCreationTimestamp="2026-01-21 11:23:44 +0000 UTC" firstStartedPulling="2026-01-21 11:23:45.115013422 +0000 UTC m=+767.408042715" lastFinishedPulling="2026-01-21 11:24:04.40530331 +0000 UTC m=+786.698332601" observedRunningTime="2026-01-21 11:24:05.068497329 +0000 UTC m=+787.361526611" watchObservedRunningTime="2026-01-21 11:24:05.072138393 +0000 UTC m=+787.365167685" Jan 21 11:24:09 crc kubenswrapper[4824]: E0121 11:24:09.500904 4824 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = parsing image configuration: Get \"https://cdn01.quay.io/quayio-production-s3/sha256/1d/1dd2737718dec9f0ef230eebb751bd30ae264b25cf038f0f2c5932675c31963c?X-Amz-Algorithm=AWS4-HMAC-SHA256&X-Amz-Credential=AKIATAAF2YHTGR23ZTE6%2F20260121%2Fus-east-1%2Fs3%2Faws4_request&X-Amz-Date=20260121T112359Z&X-Amz-Expires=600&X-Amz-SignedHeaders=host&X-Amz-Signature=2908eeb17e9386508e401967adadfeeb1a8c24ec763c4f2a1f640fe7038c9d90®ion=us-east-1&namespace=openstack-k8s-operators&username=openshift-release-dev+ocm_access_1b89217552bc42d1be3fb06a1aed001a&repo_name=rabbitmq-cluster-operator&akamai_signature=exp=1768995539~hmac=f620f0ae3fbab2bcd94414a12289b90d9494338380b60c29155025ae5b7b0e0b\": net/http: TLS handshake timeout" image="quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2" Jan 21 11:24:09 crc kubenswrapper[4824]: E0121 11:24:09.501501 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-5v855,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-668c99d594-94lqx_openstack-operators(782093ab-751c-4e38-bba3-c5f76308c82e): ErrImagePull: parsing image configuration: Get \"https://cdn01.quay.io/quayio-production-s3/sha256/1d/1dd2737718dec9f0ef230eebb751bd30ae264b25cf038f0f2c5932675c31963c?X-Amz-Algorithm=AWS4-HMAC-SHA256&X-Amz-Credential=AKIATAAF2YHTGR23ZTE6%2F20260121%2Fus-east-1%2Fs3%2Faws4_request&X-Amz-Date=20260121T112359Z&X-Amz-Expires=600&X-Amz-SignedHeaders=host&X-Amz-Signature=2908eeb17e9386508e401967adadfeeb1a8c24ec763c4f2a1f640fe7038c9d90®ion=us-east-1&namespace=openstack-k8s-operators&username=openshift-release-dev+ocm_access_1b89217552bc42d1be3fb06a1aed001a&repo_name=rabbitmq-cluster-operator&akamai_signature=exp=1768995539~hmac=f620f0ae3fbab2bcd94414a12289b90d9494338380b60c29155025ae5b7b0e0b\": net/http: TLS handshake timeout" logger="UnhandledError" Jan 21 11:24:09 crc kubenswrapper[4824]: E0121 11:24:09.502707 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"parsing image configuration: Get \\\"https://cdn01.quay.io/quayio-production-s3/sha256/1d/1dd2737718dec9f0ef230eebb751bd30ae264b25cf038f0f2c5932675c31963c?X-Amz-Algorithm=AWS4-HMAC-SHA256&X-Amz-Credential=AKIATAAF2YHTGR23ZTE6%2F20260121%2Fus-east-1%2Fs3%2Faws4_request&X-Amz-Date=20260121T112359Z&X-Amz-Expires=600&X-Amz-SignedHeaders=host&X-Amz-Signature=2908eeb17e9386508e401967adadfeeb1a8c24ec763c4f2a1f640fe7038c9d90®ion=us-east-1&namespace=openstack-k8s-operators&username=openshift-release-dev+ocm_access_1b89217552bc42d1be3fb06a1aed001a&repo_name=rabbitmq-cluster-operator&akamai_signature=exp=1768995539~hmac=f620f0ae3fbab2bcd94414a12289b90d9494338380b60c29155025ae5b7b0e0b\\\": net/http: TLS handshake timeout\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-94lqx" podUID="782093ab-751c-4e38-bba3-c5f76308c82e" Jan 21 11:24:09 crc kubenswrapper[4824]: I0121 11:24:09.769659 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6b68b8b854zxcbl" Jan 21 11:24:09 crc kubenswrapper[4824]: I0121 11:24:09.790027 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-64cd966744-9w4ch" podStartSLOduration=6.50495367 podStartE2EDuration="25.790012526s" podCreationTimestamp="2026-01-21 11:23:44 +0000 UTC" firstStartedPulling="2026-01-21 11:23:45.122431636 +0000 UTC m=+767.415460929" lastFinishedPulling="2026-01-21 11:24:04.407490493 +0000 UTC m=+786.700519785" observedRunningTime="2026-01-21 11:24:05.087530425 +0000 UTC m=+787.380559717" watchObservedRunningTime="2026-01-21 11:24:09.790012526 +0000 UTC m=+792.083041818" Jan 21 11:24:09 crc kubenswrapper[4824]: I0121 11:24:09.872704 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-77c48c7859-dzn7w" Jan 21 11:24:11 crc kubenswrapper[4824]: E0121 11:24:11.502498 4824 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = parsing image configuration: Get \"https://cdn01.quay.io/quayio-production-s3/sha256/af/afdc062de42921db29b4b435d761747aae5b8aa0eceee742dd2e647442dbf521?X-Amz-Algorithm=AWS4-HMAC-SHA256&X-Amz-Credential=AKIATAAF2YHTGR23ZTE6%2F20260121%2Fus-east-1%2Fs3%2Faws4_request&X-Amz-Date=20260121T112401Z&X-Amz-Expires=600&X-Amz-SignedHeaders=host&X-Amz-Signature=837dd8e842ca293dd58f711fd3a985c4f6012b6b591e06d0baa4bdc0876561ac®ion=us-east-1&namespace=openstack-k8s-operators&username=openshift-release-dev+ocm_access_1b89217552bc42d1be3fb06a1aed001a&repo_name=neutron-operator&akamai_signature=exp=1768995541~hmac=2f4c7f6edf21ae166ff1bf15ce0fa3c5672c3d4860a1962c6454928ba7c14a2c\": net/http: TLS handshake timeout" image="quay.io/openstack-k8s-operators/neutron-operator@sha256:0f440bf7dc937ce0135bdd328716686fd2f1320f453a9ac4e11e96383148ad6c" Jan 21 11:24:11 crc kubenswrapper[4824]: E0121 11:24:11.503025 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/neutron-operator@sha256:0f440bf7dc937ce0135bdd328716686fd2f1320f453a9ac4e11e96383148ad6c,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-mv9nq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod neutron-operator-controller-manager-cb4666565-7w8p4_openstack-operators(8c74d3da-d870-45e9-96b9-70eb2395c977): ErrImagePull: parsing image configuration: Get \"https://cdn01.quay.io/quayio-production-s3/sha256/af/afdc062de42921db29b4b435d761747aae5b8aa0eceee742dd2e647442dbf521?X-Amz-Algorithm=AWS4-HMAC-SHA256&X-Amz-Credential=AKIATAAF2YHTGR23ZTE6%2F20260121%2Fus-east-1%2Fs3%2Faws4_request&X-Amz-Date=20260121T112401Z&X-Amz-Expires=600&X-Amz-SignedHeaders=host&X-Amz-Signature=837dd8e842ca293dd58f711fd3a985c4f6012b6b591e06d0baa4bdc0876561ac®ion=us-east-1&namespace=openstack-k8s-operators&username=openshift-release-dev+ocm_access_1b89217552bc42d1be3fb06a1aed001a&repo_name=neutron-operator&akamai_signature=exp=1768995541~hmac=2f4c7f6edf21ae166ff1bf15ce0fa3c5672c3d4860a1962c6454928ba7c14a2c\": net/http: TLS handshake timeout" logger="UnhandledError" Jan 21 11:24:11 crc kubenswrapper[4824]: E0121 11:24:11.504185 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"parsing image configuration: Get \\\"https://cdn01.quay.io/quayio-production-s3/sha256/af/afdc062de42921db29b4b435d761747aae5b8aa0eceee742dd2e647442dbf521?X-Amz-Algorithm=AWS4-HMAC-SHA256&X-Amz-Credential=AKIATAAF2YHTGR23ZTE6%2F20260121%2Fus-east-1%2Fs3%2Faws4_request&X-Amz-Date=20260121T112401Z&X-Amz-Expires=600&X-Amz-SignedHeaders=host&X-Amz-Signature=837dd8e842ca293dd58f711fd3a985c4f6012b6b591e06d0baa4bdc0876561ac®ion=us-east-1&namespace=openstack-k8s-operators&username=openshift-release-dev+ocm_access_1b89217552bc42d1be3fb06a1aed001a&repo_name=neutron-operator&akamai_signature=exp=1768995541~hmac=2f4c7f6edf21ae166ff1bf15ce0fa3c5672c3d4860a1962c6454928ba7c14a2c\\\": net/http: TLS handshake timeout\"" pod="openstack-operators/neutron-operator-controller-manager-cb4666565-7w8p4" podUID="8c74d3da-d870-45e9-96b9-70eb2395c977" Jan 21 11:24:14 crc kubenswrapper[4824]: I0121 11:24:14.136419 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-65849867d6-8ngx6" Jan 21 11:24:14 crc kubenswrapper[4824]: I0121 11:24:14.164533 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-55db956ddc-78z2q" Jan 21 11:24:14 crc kubenswrapper[4824]: I0121 11:24:14.432135 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-7cd8bc9dbb-xkp8x" Jan 21 11:24:14 crc kubenswrapper[4824]: I0121 11:24:14.486001 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-64cd966744-9w4ch" Jan 21 11:24:16 crc kubenswrapper[4824]: I0121 11:24:16.191950 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c1128e01-8b7c-4742-b1da-dbf06aa8b4d8-metrics-certs\") pod \"openstack-operator-controller-manager-74d84f758c-vjgps\" (UID: \"c1128e01-8b7c-4742-b1da-dbf06aa8b4d8\") " pod="openstack-operators/openstack-operator-controller-manager-74d84f758c-vjgps" Jan 21 11:24:16 crc kubenswrapper[4824]: I0121 11:24:16.192268 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/c1128e01-8b7c-4742-b1da-dbf06aa8b4d8-webhook-certs\") pod \"openstack-operator-controller-manager-74d84f758c-vjgps\" (UID: \"c1128e01-8b7c-4742-b1da-dbf06aa8b4d8\") " pod="openstack-operators/openstack-operator-controller-manager-74d84f758c-vjgps" Jan 21 11:24:16 crc kubenswrapper[4824]: I0121 11:24:16.197197 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c1128e01-8b7c-4742-b1da-dbf06aa8b4d8-metrics-certs\") pod \"openstack-operator-controller-manager-74d84f758c-vjgps\" (UID: \"c1128e01-8b7c-4742-b1da-dbf06aa8b4d8\") " pod="openstack-operators/openstack-operator-controller-manager-74d84f758c-vjgps" Jan 21 11:24:16 crc kubenswrapper[4824]: I0121 11:24:16.197314 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/c1128e01-8b7c-4742-b1da-dbf06aa8b4d8-webhook-certs\") pod \"openstack-operator-controller-manager-74d84f758c-vjgps\" (UID: \"c1128e01-8b7c-4742-b1da-dbf06aa8b4d8\") " pod="openstack-operators/openstack-operator-controller-manager-74d84f758c-vjgps" Jan 21 11:24:16 crc kubenswrapper[4824]: I0121 11:24:16.476687 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-xjvvw" Jan 21 11:24:16 crc kubenswrapper[4824]: I0121 11:24:16.485625 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-74d84f758c-vjgps" Jan 21 11:24:16 crc kubenswrapper[4824]: I0121 11:24:16.838425 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-74d84f758c-vjgps"] Jan 21 11:24:17 crc kubenswrapper[4824]: I0121 11:24:17.133986 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-74d84f758c-vjgps" event={"ID":"c1128e01-8b7c-4742-b1da-dbf06aa8b4d8","Type":"ContainerStarted","Data":"9d0b177e548eb2e34aa7bd44e026feaa41ad53dd5eac90d7991be6b44e6233bc"} Jan 21 11:24:17 crc kubenswrapper[4824]: I0121 11:24:17.134302 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-74d84f758c-vjgps" Jan 21 11:24:17 crc kubenswrapper[4824]: I0121 11:24:17.134315 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-74d84f758c-vjgps" event={"ID":"c1128e01-8b7c-4742-b1da-dbf06aa8b4d8","Type":"ContainerStarted","Data":"dfdacc7e7292e1d386dfca28ca5044d5c10632f2eb094375d8a700c43363fbf0"} Jan 21 11:24:17 crc kubenswrapper[4824]: I0121 11:24:17.154865 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-74d84f758c-vjgps" podStartSLOduration=33.154843049 podStartE2EDuration="33.154843049s" podCreationTimestamp="2026-01-21 11:23:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:24:17.153971295 +0000 UTC m=+799.447000587" watchObservedRunningTime="2026-01-21 11:24:17.154843049 +0000 UTC m=+799.447872341" Jan 21 11:24:20 crc kubenswrapper[4824]: E0121 11:24:20.050590 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-94lqx" podUID="782093ab-751c-4e38-bba3-c5f76308c82e" Jan 21 11:24:25 crc kubenswrapper[4824]: E0121 11:24:25.050441 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/neutron-operator@sha256:0f440bf7dc937ce0135bdd328716686fd2f1320f453a9ac4e11e96383148ad6c\\\"\"" pod="openstack-operators/neutron-operator-controller-manager-cb4666565-7w8p4" podUID="8c74d3da-d870-45e9-96b9-70eb2395c977" Jan 21 11:24:26 crc kubenswrapper[4824]: I0121 11:24:26.490457 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-74d84f758c-vjgps" Jan 21 11:24:39 crc kubenswrapper[4824]: I0121 11:24:39.248390 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-94lqx" event={"ID":"782093ab-751c-4e38-bba3-c5f76308c82e","Type":"ContainerStarted","Data":"4c12156aea277683ef2931fbe886c4ae11e0b16b83a9ffb882ec8ab5f8cf54f7"} Jan 21 11:24:39 crc kubenswrapper[4824]: I0121 11:24:39.260734 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-94lqx" podStartSLOduration=2.158120752 podStartE2EDuration="55.260720716s" podCreationTimestamp="2026-01-21 11:23:44 +0000 UTC" firstStartedPulling="2026-01-21 11:23:45.178437605 +0000 UTC m=+767.471466897" lastFinishedPulling="2026-01-21 11:24:38.281037569 +0000 UTC m=+820.574066861" observedRunningTime="2026-01-21 11:24:39.259497109 +0000 UTC m=+821.552526401" watchObservedRunningTime="2026-01-21 11:24:39.260720716 +0000 UTC m=+821.553750009" Jan 21 11:24:41 crc kubenswrapper[4824]: I0121 11:24:41.258187 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-cb4666565-7w8p4" event={"ID":"8c74d3da-d870-45e9-96b9-70eb2395c977","Type":"ContainerStarted","Data":"0bc7d1544e430cce051bfb7bf230c2897a71d350529bafdf4faae5c98df4f629"} Jan 21 11:24:41 crc kubenswrapper[4824]: I0121 11:24:41.258723 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-cb4666565-7w8p4" Jan 21 11:24:41 crc kubenswrapper[4824]: I0121 11:24:41.272209 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-cb4666565-7w8p4" podStartSLOduration=2.704576031 podStartE2EDuration="58.272189561s" podCreationTimestamp="2026-01-21 11:23:43 +0000 UTC" firstStartedPulling="2026-01-21 11:23:45.010099676 +0000 UTC m=+767.303128968" lastFinishedPulling="2026-01-21 11:24:40.577713205 +0000 UTC m=+822.870742498" observedRunningTime="2026-01-21 11:24:41.268641813 +0000 UTC m=+823.561671105" watchObservedRunningTime="2026-01-21 11:24:41.272189561 +0000 UTC m=+823.565218853" Jan 21 11:24:54 crc kubenswrapper[4824]: I0121 11:24:54.127888 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-cb4666565-7w8p4" Jan 21 11:25:06 crc kubenswrapper[4824]: I0121 11:25:06.913985 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-lmtkp"] Jan 21 11:25:06 crc kubenswrapper[4824]: I0121 11:25:06.915520 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-lmtkp" Jan 21 11:25:06 crc kubenswrapper[4824]: I0121 11:25:06.917301 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Jan 21 11:25:06 crc kubenswrapper[4824]: I0121 11:25:06.917657 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Jan 21 11:25:06 crc kubenswrapper[4824]: I0121 11:25:06.917676 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Jan 21 11:25:06 crc kubenswrapper[4824]: I0121 11:25:06.917950 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-r6gf5" Jan 21 11:25:06 crc kubenswrapper[4824]: I0121 11:25:06.925103 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-lmtkp"] Jan 21 11:25:06 crc kubenswrapper[4824]: I0121 11:25:06.971391 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-tmr2j"] Jan 21 11:25:06 crc kubenswrapper[4824]: I0121 11:25:06.972421 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-tmr2j" Jan 21 11:25:06 crc kubenswrapper[4824]: I0121 11:25:06.974290 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Jan 21 11:25:06 crc kubenswrapper[4824]: I0121 11:25:06.980288 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-tmr2j"] Jan 21 11:25:07 crc kubenswrapper[4824]: I0121 11:25:07.032123 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e8aca65b-76cd-45f9-8b20-da2e34a41fc2-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-tmr2j\" (UID: \"e8aca65b-76cd-45f9-8b20-da2e34a41fc2\") " pod="openstack/dnsmasq-dns-78dd6ddcc-tmr2j" Jan 21 11:25:07 crc kubenswrapper[4824]: I0121 11:25:07.032181 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b48ab67f-88cb-4921-b6a3-38b77b534bf9-config\") pod \"dnsmasq-dns-675f4bcbfc-lmtkp\" (UID: \"b48ab67f-88cb-4921-b6a3-38b77b534bf9\") " pod="openstack/dnsmasq-dns-675f4bcbfc-lmtkp" Jan 21 11:25:07 crc kubenswrapper[4824]: I0121 11:25:07.032233 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jt49s\" (UniqueName: \"kubernetes.io/projected/e8aca65b-76cd-45f9-8b20-da2e34a41fc2-kube-api-access-jt49s\") pod \"dnsmasq-dns-78dd6ddcc-tmr2j\" (UID: \"e8aca65b-76cd-45f9-8b20-da2e34a41fc2\") " pod="openstack/dnsmasq-dns-78dd6ddcc-tmr2j" Jan 21 11:25:07 crc kubenswrapper[4824]: I0121 11:25:07.032277 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8q4rz\" (UniqueName: \"kubernetes.io/projected/b48ab67f-88cb-4921-b6a3-38b77b534bf9-kube-api-access-8q4rz\") pod \"dnsmasq-dns-675f4bcbfc-lmtkp\" (UID: \"b48ab67f-88cb-4921-b6a3-38b77b534bf9\") " pod="openstack/dnsmasq-dns-675f4bcbfc-lmtkp" Jan 21 11:25:07 crc kubenswrapper[4824]: I0121 11:25:07.032301 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e8aca65b-76cd-45f9-8b20-da2e34a41fc2-config\") pod \"dnsmasq-dns-78dd6ddcc-tmr2j\" (UID: \"e8aca65b-76cd-45f9-8b20-da2e34a41fc2\") " pod="openstack/dnsmasq-dns-78dd6ddcc-tmr2j" Jan 21 11:25:07 crc kubenswrapper[4824]: I0121 11:25:07.133566 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b48ab67f-88cb-4921-b6a3-38b77b534bf9-config\") pod \"dnsmasq-dns-675f4bcbfc-lmtkp\" (UID: \"b48ab67f-88cb-4921-b6a3-38b77b534bf9\") " pod="openstack/dnsmasq-dns-675f4bcbfc-lmtkp" Jan 21 11:25:07 crc kubenswrapper[4824]: I0121 11:25:07.133648 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jt49s\" (UniqueName: \"kubernetes.io/projected/e8aca65b-76cd-45f9-8b20-da2e34a41fc2-kube-api-access-jt49s\") pod \"dnsmasq-dns-78dd6ddcc-tmr2j\" (UID: \"e8aca65b-76cd-45f9-8b20-da2e34a41fc2\") " pod="openstack/dnsmasq-dns-78dd6ddcc-tmr2j" Jan 21 11:25:07 crc kubenswrapper[4824]: I0121 11:25:07.133725 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8q4rz\" (UniqueName: \"kubernetes.io/projected/b48ab67f-88cb-4921-b6a3-38b77b534bf9-kube-api-access-8q4rz\") pod \"dnsmasq-dns-675f4bcbfc-lmtkp\" (UID: \"b48ab67f-88cb-4921-b6a3-38b77b534bf9\") " pod="openstack/dnsmasq-dns-675f4bcbfc-lmtkp" Jan 21 11:25:07 crc kubenswrapper[4824]: I0121 11:25:07.133927 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e8aca65b-76cd-45f9-8b20-da2e34a41fc2-config\") pod \"dnsmasq-dns-78dd6ddcc-tmr2j\" (UID: \"e8aca65b-76cd-45f9-8b20-da2e34a41fc2\") " pod="openstack/dnsmasq-dns-78dd6ddcc-tmr2j" Jan 21 11:25:07 crc kubenswrapper[4824]: I0121 11:25:07.134289 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e8aca65b-76cd-45f9-8b20-da2e34a41fc2-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-tmr2j\" (UID: \"e8aca65b-76cd-45f9-8b20-da2e34a41fc2\") " pod="openstack/dnsmasq-dns-78dd6ddcc-tmr2j" Jan 21 11:25:07 crc kubenswrapper[4824]: I0121 11:25:07.134593 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e8aca65b-76cd-45f9-8b20-da2e34a41fc2-config\") pod \"dnsmasq-dns-78dd6ddcc-tmr2j\" (UID: \"e8aca65b-76cd-45f9-8b20-da2e34a41fc2\") " pod="openstack/dnsmasq-dns-78dd6ddcc-tmr2j" Jan 21 11:25:07 crc kubenswrapper[4824]: I0121 11:25:07.134857 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e8aca65b-76cd-45f9-8b20-da2e34a41fc2-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-tmr2j\" (UID: \"e8aca65b-76cd-45f9-8b20-da2e34a41fc2\") " pod="openstack/dnsmasq-dns-78dd6ddcc-tmr2j" Jan 21 11:25:07 crc kubenswrapper[4824]: I0121 11:25:07.134985 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b48ab67f-88cb-4921-b6a3-38b77b534bf9-config\") pod \"dnsmasq-dns-675f4bcbfc-lmtkp\" (UID: \"b48ab67f-88cb-4921-b6a3-38b77b534bf9\") " pod="openstack/dnsmasq-dns-675f4bcbfc-lmtkp" Jan 21 11:25:07 crc kubenswrapper[4824]: I0121 11:25:07.148564 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8q4rz\" (UniqueName: \"kubernetes.io/projected/b48ab67f-88cb-4921-b6a3-38b77b534bf9-kube-api-access-8q4rz\") pod \"dnsmasq-dns-675f4bcbfc-lmtkp\" (UID: \"b48ab67f-88cb-4921-b6a3-38b77b534bf9\") " pod="openstack/dnsmasq-dns-675f4bcbfc-lmtkp" Jan 21 11:25:07 crc kubenswrapper[4824]: I0121 11:25:07.149879 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jt49s\" (UniqueName: \"kubernetes.io/projected/e8aca65b-76cd-45f9-8b20-da2e34a41fc2-kube-api-access-jt49s\") pod \"dnsmasq-dns-78dd6ddcc-tmr2j\" (UID: \"e8aca65b-76cd-45f9-8b20-da2e34a41fc2\") " pod="openstack/dnsmasq-dns-78dd6ddcc-tmr2j" Jan 21 11:25:07 crc kubenswrapper[4824]: I0121 11:25:07.230986 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-lmtkp" Jan 21 11:25:07 crc kubenswrapper[4824]: I0121 11:25:07.283816 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-tmr2j" Jan 21 11:25:07 crc kubenswrapper[4824]: I0121 11:25:07.613165 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-lmtkp"] Jan 21 11:25:07 crc kubenswrapper[4824]: I0121 11:25:07.668518 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-tmr2j"] Jan 21 11:25:07 crc kubenswrapper[4824]: W0121 11:25:07.672199 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode8aca65b_76cd_45f9_8b20_da2e34a41fc2.slice/crio-8c626b420278544fc13bdb44a0fc27116c48e8b3b2e62a1f886c8aa8d6e59139 WatchSource:0}: Error finding container 8c626b420278544fc13bdb44a0fc27116c48e8b3b2e62a1f886c8aa8d6e59139: Status 404 returned error can't find the container with id 8c626b420278544fc13bdb44a0fc27116c48e8b3b2e62a1f886c8aa8d6e59139 Jan 21 11:25:08 crc kubenswrapper[4824]: I0121 11:25:08.385778 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-lmtkp" event={"ID":"b48ab67f-88cb-4921-b6a3-38b77b534bf9","Type":"ContainerStarted","Data":"c935cf67bf3bb60e71d9e8fa32ea33c10f87bb841cf838b50e9a195acf0ce5f3"} Jan 21 11:25:08 crc kubenswrapper[4824]: I0121 11:25:08.387466 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-tmr2j" event={"ID":"e8aca65b-76cd-45f9-8b20-da2e34a41fc2","Type":"ContainerStarted","Data":"8c626b420278544fc13bdb44a0fc27116c48e8b3b2e62a1f886c8aa8d6e59139"} Jan 21 11:25:09 crc kubenswrapper[4824]: I0121 11:25:09.887148 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-lmtkp"] Jan 21 11:25:09 crc kubenswrapper[4824]: I0121 11:25:09.911040 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-zspkp"] Jan 21 11:25:09 crc kubenswrapper[4824]: I0121 11:25:09.920300 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-zspkp"] Jan 21 11:25:09 crc kubenswrapper[4824]: I0121 11:25:09.920382 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-zspkp" Jan 21 11:25:09 crc kubenswrapper[4824]: I0121 11:25:09.979839 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f67dfc52-dc0e-404e-af22-75c6a3d10721-dns-svc\") pod \"dnsmasq-dns-666b6646f7-zspkp\" (UID: \"f67dfc52-dc0e-404e-af22-75c6a3d10721\") " pod="openstack/dnsmasq-dns-666b6646f7-zspkp" Jan 21 11:25:09 crc kubenswrapper[4824]: I0121 11:25:09.979967 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f67dfc52-dc0e-404e-af22-75c6a3d10721-config\") pod \"dnsmasq-dns-666b6646f7-zspkp\" (UID: \"f67dfc52-dc0e-404e-af22-75c6a3d10721\") " pod="openstack/dnsmasq-dns-666b6646f7-zspkp" Jan 21 11:25:09 crc kubenswrapper[4824]: I0121 11:25:09.980034 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-msx57\" (UniqueName: \"kubernetes.io/projected/f67dfc52-dc0e-404e-af22-75c6a3d10721-kube-api-access-msx57\") pod \"dnsmasq-dns-666b6646f7-zspkp\" (UID: \"f67dfc52-dc0e-404e-af22-75c6a3d10721\") " pod="openstack/dnsmasq-dns-666b6646f7-zspkp" Jan 21 11:25:10 crc kubenswrapper[4824]: I0121 11:25:10.086472 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-msx57\" (UniqueName: \"kubernetes.io/projected/f67dfc52-dc0e-404e-af22-75c6a3d10721-kube-api-access-msx57\") pod \"dnsmasq-dns-666b6646f7-zspkp\" (UID: \"f67dfc52-dc0e-404e-af22-75c6a3d10721\") " pod="openstack/dnsmasq-dns-666b6646f7-zspkp" Jan 21 11:25:10 crc kubenswrapper[4824]: I0121 11:25:10.086670 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f67dfc52-dc0e-404e-af22-75c6a3d10721-dns-svc\") pod \"dnsmasq-dns-666b6646f7-zspkp\" (UID: \"f67dfc52-dc0e-404e-af22-75c6a3d10721\") " pod="openstack/dnsmasq-dns-666b6646f7-zspkp" Jan 21 11:25:10 crc kubenswrapper[4824]: I0121 11:25:10.089973 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f67dfc52-dc0e-404e-af22-75c6a3d10721-dns-svc\") pod \"dnsmasq-dns-666b6646f7-zspkp\" (UID: \"f67dfc52-dc0e-404e-af22-75c6a3d10721\") " pod="openstack/dnsmasq-dns-666b6646f7-zspkp" Jan 21 11:25:10 crc kubenswrapper[4824]: I0121 11:25:10.090080 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f67dfc52-dc0e-404e-af22-75c6a3d10721-config\") pod \"dnsmasq-dns-666b6646f7-zspkp\" (UID: \"f67dfc52-dc0e-404e-af22-75c6a3d10721\") " pod="openstack/dnsmasq-dns-666b6646f7-zspkp" Jan 21 11:25:10 crc kubenswrapper[4824]: I0121 11:25:10.091478 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f67dfc52-dc0e-404e-af22-75c6a3d10721-config\") pod \"dnsmasq-dns-666b6646f7-zspkp\" (UID: \"f67dfc52-dc0e-404e-af22-75c6a3d10721\") " pod="openstack/dnsmasq-dns-666b6646f7-zspkp" Jan 21 11:25:10 crc kubenswrapper[4824]: I0121 11:25:10.111987 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-msx57\" (UniqueName: \"kubernetes.io/projected/f67dfc52-dc0e-404e-af22-75c6a3d10721-kube-api-access-msx57\") pod \"dnsmasq-dns-666b6646f7-zspkp\" (UID: \"f67dfc52-dc0e-404e-af22-75c6a3d10721\") " pod="openstack/dnsmasq-dns-666b6646f7-zspkp" Jan 21 11:25:10 crc kubenswrapper[4824]: I0121 11:25:10.154905 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-tmr2j"] Jan 21 11:25:10 crc kubenswrapper[4824]: I0121 11:25:10.174007 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-nrqwx"] Jan 21 11:25:10 crc kubenswrapper[4824]: I0121 11:25:10.175237 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-nrqwx" Jan 21 11:25:10 crc kubenswrapper[4824]: I0121 11:25:10.187988 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-nrqwx"] Jan 21 11:25:10 crc kubenswrapper[4824]: I0121 11:25:10.240471 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-zspkp" Jan 21 11:25:10 crc kubenswrapper[4824]: I0121 11:25:10.295677 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/71b408e9-2ba6-4b0e-9b4c-22faf4ef4fd4-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-nrqwx\" (UID: \"71b408e9-2ba6-4b0e-9b4c-22faf4ef4fd4\") " pod="openstack/dnsmasq-dns-57d769cc4f-nrqwx" Jan 21 11:25:10 crc kubenswrapper[4824]: I0121 11:25:10.297412 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/71b408e9-2ba6-4b0e-9b4c-22faf4ef4fd4-config\") pod \"dnsmasq-dns-57d769cc4f-nrqwx\" (UID: \"71b408e9-2ba6-4b0e-9b4c-22faf4ef4fd4\") " pod="openstack/dnsmasq-dns-57d769cc4f-nrqwx" Jan 21 11:25:10 crc kubenswrapper[4824]: I0121 11:25:10.297585 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wcgfn\" (UniqueName: \"kubernetes.io/projected/71b408e9-2ba6-4b0e-9b4c-22faf4ef4fd4-kube-api-access-wcgfn\") pod \"dnsmasq-dns-57d769cc4f-nrqwx\" (UID: \"71b408e9-2ba6-4b0e-9b4c-22faf4ef4fd4\") " pod="openstack/dnsmasq-dns-57d769cc4f-nrqwx" Jan 21 11:25:10 crc kubenswrapper[4824]: I0121 11:25:10.398886 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/71b408e9-2ba6-4b0e-9b4c-22faf4ef4fd4-config\") pod \"dnsmasq-dns-57d769cc4f-nrqwx\" (UID: \"71b408e9-2ba6-4b0e-9b4c-22faf4ef4fd4\") " pod="openstack/dnsmasq-dns-57d769cc4f-nrqwx" Jan 21 11:25:10 crc kubenswrapper[4824]: I0121 11:25:10.399094 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wcgfn\" (UniqueName: \"kubernetes.io/projected/71b408e9-2ba6-4b0e-9b4c-22faf4ef4fd4-kube-api-access-wcgfn\") pod \"dnsmasq-dns-57d769cc4f-nrqwx\" (UID: \"71b408e9-2ba6-4b0e-9b4c-22faf4ef4fd4\") " pod="openstack/dnsmasq-dns-57d769cc4f-nrqwx" Jan 21 11:25:10 crc kubenswrapper[4824]: I0121 11:25:10.399288 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/71b408e9-2ba6-4b0e-9b4c-22faf4ef4fd4-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-nrqwx\" (UID: \"71b408e9-2ba6-4b0e-9b4c-22faf4ef4fd4\") " pod="openstack/dnsmasq-dns-57d769cc4f-nrqwx" Jan 21 11:25:10 crc kubenswrapper[4824]: I0121 11:25:10.399752 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/71b408e9-2ba6-4b0e-9b4c-22faf4ef4fd4-config\") pod \"dnsmasq-dns-57d769cc4f-nrqwx\" (UID: \"71b408e9-2ba6-4b0e-9b4c-22faf4ef4fd4\") " pod="openstack/dnsmasq-dns-57d769cc4f-nrqwx" Jan 21 11:25:10 crc kubenswrapper[4824]: I0121 11:25:10.401449 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/71b408e9-2ba6-4b0e-9b4c-22faf4ef4fd4-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-nrqwx\" (UID: \"71b408e9-2ba6-4b0e-9b4c-22faf4ef4fd4\") " pod="openstack/dnsmasq-dns-57d769cc4f-nrqwx" Jan 21 11:25:10 crc kubenswrapper[4824]: I0121 11:25:10.416739 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wcgfn\" (UniqueName: \"kubernetes.io/projected/71b408e9-2ba6-4b0e-9b4c-22faf4ef4fd4-kube-api-access-wcgfn\") pod \"dnsmasq-dns-57d769cc4f-nrqwx\" (UID: \"71b408e9-2ba6-4b0e-9b4c-22faf4ef4fd4\") " pod="openstack/dnsmasq-dns-57d769cc4f-nrqwx" Jan 21 11:25:10 crc kubenswrapper[4824]: I0121 11:25:10.495899 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-nrqwx" Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.047060 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.049320 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.051741 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.052103 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.052321 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.052468 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-22jwj" Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.052527 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.052619 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.052738 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.065722 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.109008 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/84e74d7b-18a0-4a3f-8680-6246ac538a6e-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"84e74d7b-18a0-4a3f-8680-6246ac538a6e\") " pod="openstack/rabbitmq-server-0" Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.109049 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/84e74d7b-18a0-4a3f-8680-6246ac538a6e-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"84e74d7b-18a0-4a3f-8680-6246ac538a6e\") " pod="openstack/rabbitmq-server-0" Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.109108 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/84e74d7b-18a0-4a3f-8680-6246ac538a6e-server-conf\") pod \"rabbitmq-server-0\" (UID: \"84e74d7b-18a0-4a3f-8680-6246ac538a6e\") " pod="openstack/rabbitmq-server-0" Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.109234 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/84e74d7b-18a0-4a3f-8680-6246ac538a6e-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"84e74d7b-18a0-4a3f-8680-6246ac538a6e\") " pod="openstack/rabbitmq-server-0" Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.109280 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/84e74d7b-18a0-4a3f-8680-6246ac538a6e-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"84e74d7b-18a0-4a3f-8680-6246ac538a6e\") " pod="openstack/rabbitmq-server-0" Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.109299 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/84e74d7b-18a0-4a3f-8680-6246ac538a6e-config-data\") pod \"rabbitmq-server-0\" (UID: \"84e74d7b-18a0-4a3f-8680-6246ac538a6e\") " pod="openstack/rabbitmq-server-0" Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.109313 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/84e74d7b-18a0-4a3f-8680-6246ac538a6e-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"84e74d7b-18a0-4a3f-8680-6246ac538a6e\") " pod="openstack/rabbitmq-server-0" Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.109343 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/84e74d7b-18a0-4a3f-8680-6246ac538a6e-pod-info\") pod \"rabbitmq-server-0\" (UID: \"84e74d7b-18a0-4a3f-8680-6246ac538a6e\") " pod="openstack/rabbitmq-server-0" Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.109432 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-server-0\" (UID: \"84e74d7b-18a0-4a3f-8680-6246ac538a6e\") " pod="openstack/rabbitmq-server-0" Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.109464 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/84e74d7b-18a0-4a3f-8680-6246ac538a6e-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"84e74d7b-18a0-4a3f-8680-6246ac538a6e\") " pod="openstack/rabbitmq-server-0" Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.109519 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zkwkl\" (UniqueName: \"kubernetes.io/projected/84e74d7b-18a0-4a3f-8680-6246ac538a6e-kube-api-access-zkwkl\") pod \"rabbitmq-server-0\" (UID: \"84e74d7b-18a0-4a3f-8680-6246ac538a6e\") " pod="openstack/rabbitmq-server-0" Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.211439 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zkwkl\" (UniqueName: \"kubernetes.io/projected/84e74d7b-18a0-4a3f-8680-6246ac538a6e-kube-api-access-zkwkl\") pod \"rabbitmq-server-0\" (UID: \"84e74d7b-18a0-4a3f-8680-6246ac538a6e\") " pod="openstack/rabbitmq-server-0" Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.211530 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/84e74d7b-18a0-4a3f-8680-6246ac538a6e-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"84e74d7b-18a0-4a3f-8680-6246ac538a6e\") " pod="openstack/rabbitmq-server-0" Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.211552 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/84e74d7b-18a0-4a3f-8680-6246ac538a6e-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"84e74d7b-18a0-4a3f-8680-6246ac538a6e\") " pod="openstack/rabbitmq-server-0" Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.211571 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/84e74d7b-18a0-4a3f-8680-6246ac538a6e-server-conf\") pod \"rabbitmq-server-0\" (UID: \"84e74d7b-18a0-4a3f-8680-6246ac538a6e\") " pod="openstack/rabbitmq-server-0" Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.211611 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/84e74d7b-18a0-4a3f-8680-6246ac538a6e-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"84e74d7b-18a0-4a3f-8680-6246ac538a6e\") " pod="openstack/rabbitmq-server-0" Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.211636 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/84e74d7b-18a0-4a3f-8680-6246ac538a6e-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"84e74d7b-18a0-4a3f-8680-6246ac538a6e\") " pod="openstack/rabbitmq-server-0" Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.211652 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/84e74d7b-18a0-4a3f-8680-6246ac538a6e-config-data\") pod \"rabbitmq-server-0\" (UID: \"84e74d7b-18a0-4a3f-8680-6246ac538a6e\") " pod="openstack/rabbitmq-server-0" Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.211665 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/84e74d7b-18a0-4a3f-8680-6246ac538a6e-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"84e74d7b-18a0-4a3f-8680-6246ac538a6e\") " pod="openstack/rabbitmq-server-0" Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.211703 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/84e74d7b-18a0-4a3f-8680-6246ac538a6e-pod-info\") pod \"rabbitmq-server-0\" (UID: \"84e74d7b-18a0-4a3f-8680-6246ac538a6e\") " pod="openstack/rabbitmq-server-0" Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.211748 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-server-0\" (UID: \"84e74d7b-18a0-4a3f-8680-6246ac538a6e\") " pod="openstack/rabbitmq-server-0" Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.211767 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/84e74d7b-18a0-4a3f-8680-6246ac538a6e-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"84e74d7b-18a0-4a3f-8680-6246ac538a6e\") " pod="openstack/rabbitmq-server-0" Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.212719 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/84e74d7b-18a0-4a3f-8680-6246ac538a6e-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"84e74d7b-18a0-4a3f-8680-6246ac538a6e\") " pod="openstack/rabbitmq-server-0" Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.213164 4824 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-server-0\" (UID: \"84e74d7b-18a0-4a3f-8680-6246ac538a6e\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/rabbitmq-server-0" Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.214482 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/84e74d7b-18a0-4a3f-8680-6246ac538a6e-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"84e74d7b-18a0-4a3f-8680-6246ac538a6e\") " pod="openstack/rabbitmq-server-0" Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.214564 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/84e74d7b-18a0-4a3f-8680-6246ac538a6e-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"84e74d7b-18a0-4a3f-8680-6246ac538a6e\") " pod="openstack/rabbitmq-server-0" Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.215373 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/84e74d7b-18a0-4a3f-8680-6246ac538a6e-server-conf\") pod \"rabbitmq-server-0\" (UID: \"84e74d7b-18a0-4a3f-8680-6246ac538a6e\") " pod="openstack/rabbitmq-server-0" Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.217217 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/84e74d7b-18a0-4a3f-8680-6246ac538a6e-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"84e74d7b-18a0-4a3f-8680-6246ac538a6e\") " pod="openstack/rabbitmq-server-0" Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.217309 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/84e74d7b-18a0-4a3f-8680-6246ac538a6e-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"84e74d7b-18a0-4a3f-8680-6246ac538a6e\") " pod="openstack/rabbitmq-server-0" Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.217363 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/84e74d7b-18a0-4a3f-8680-6246ac538a6e-config-data\") pod \"rabbitmq-server-0\" (UID: \"84e74d7b-18a0-4a3f-8680-6246ac538a6e\") " pod="openstack/rabbitmq-server-0" Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.224829 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/84e74d7b-18a0-4a3f-8680-6246ac538a6e-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"84e74d7b-18a0-4a3f-8680-6246ac538a6e\") " pod="openstack/rabbitmq-server-0" Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.225348 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/84e74d7b-18a0-4a3f-8680-6246ac538a6e-pod-info\") pod \"rabbitmq-server-0\" (UID: \"84e74d7b-18a0-4a3f-8680-6246ac538a6e\") " pod="openstack/rabbitmq-server-0" Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.226608 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zkwkl\" (UniqueName: \"kubernetes.io/projected/84e74d7b-18a0-4a3f-8680-6246ac538a6e-kube-api-access-zkwkl\") pod \"rabbitmq-server-0\" (UID: \"84e74d7b-18a0-4a3f-8680-6246ac538a6e\") " pod="openstack/rabbitmq-server-0" Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.234922 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-server-0\" (UID: \"84e74d7b-18a0-4a3f-8680-6246ac538a6e\") " pod="openstack/rabbitmq-server-0" Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.306244 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.307358 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.309442 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.309739 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.309858 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.309932 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.309997 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.311777 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-zjdwv" Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.311783 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.312529 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.372007 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.413495 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"156fec50-b486-4e84-a7bf-b40491a863c7\") " pod="openstack/rabbitmq-cell1-server-0" Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.413539 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/156fec50-b486-4e84-a7bf-b40491a863c7-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"156fec50-b486-4e84-a7bf-b40491a863c7\") " pod="openstack/rabbitmq-cell1-server-0" Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.413558 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/156fec50-b486-4e84-a7bf-b40491a863c7-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"156fec50-b486-4e84-a7bf-b40491a863c7\") " pod="openstack/rabbitmq-cell1-server-0" Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.413616 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/156fec50-b486-4e84-a7bf-b40491a863c7-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"156fec50-b486-4e84-a7bf-b40491a863c7\") " pod="openstack/rabbitmq-cell1-server-0" Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.413641 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/156fec50-b486-4e84-a7bf-b40491a863c7-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"156fec50-b486-4e84-a7bf-b40491a863c7\") " pod="openstack/rabbitmq-cell1-server-0" Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.413667 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/156fec50-b486-4e84-a7bf-b40491a863c7-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"156fec50-b486-4e84-a7bf-b40491a863c7\") " pod="openstack/rabbitmq-cell1-server-0" Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.413711 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/156fec50-b486-4e84-a7bf-b40491a863c7-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"156fec50-b486-4e84-a7bf-b40491a863c7\") " pod="openstack/rabbitmq-cell1-server-0" Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.413725 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/156fec50-b486-4e84-a7bf-b40491a863c7-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"156fec50-b486-4e84-a7bf-b40491a863c7\") " pod="openstack/rabbitmq-cell1-server-0" Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.413742 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/156fec50-b486-4e84-a7bf-b40491a863c7-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"156fec50-b486-4e84-a7bf-b40491a863c7\") " pod="openstack/rabbitmq-cell1-server-0" Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.413759 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/156fec50-b486-4e84-a7bf-b40491a863c7-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"156fec50-b486-4e84-a7bf-b40491a863c7\") " pod="openstack/rabbitmq-cell1-server-0" Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.413786 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j9nc5\" (UniqueName: \"kubernetes.io/projected/156fec50-b486-4e84-a7bf-b40491a863c7-kube-api-access-j9nc5\") pod \"rabbitmq-cell1-server-0\" (UID: \"156fec50-b486-4e84-a7bf-b40491a863c7\") " pod="openstack/rabbitmq-cell1-server-0" Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.515162 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/156fec50-b486-4e84-a7bf-b40491a863c7-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"156fec50-b486-4e84-a7bf-b40491a863c7\") " pod="openstack/rabbitmq-cell1-server-0" Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.515296 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/156fec50-b486-4e84-a7bf-b40491a863c7-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"156fec50-b486-4e84-a7bf-b40491a863c7\") " pod="openstack/rabbitmq-cell1-server-0" Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.515315 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/156fec50-b486-4e84-a7bf-b40491a863c7-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"156fec50-b486-4e84-a7bf-b40491a863c7\") " pod="openstack/rabbitmq-cell1-server-0" Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.515333 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/156fec50-b486-4e84-a7bf-b40491a863c7-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"156fec50-b486-4e84-a7bf-b40491a863c7\") " pod="openstack/rabbitmq-cell1-server-0" Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.515349 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/156fec50-b486-4e84-a7bf-b40491a863c7-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"156fec50-b486-4e84-a7bf-b40491a863c7\") " pod="openstack/rabbitmq-cell1-server-0" Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.515381 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j9nc5\" (UniqueName: \"kubernetes.io/projected/156fec50-b486-4e84-a7bf-b40491a863c7-kube-api-access-j9nc5\") pod \"rabbitmq-cell1-server-0\" (UID: \"156fec50-b486-4e84-a7bf-b40491a863c7\") " pod="openstack/rabbitmq-cell1-server-0" Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.515434 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"156fec50-b486-4e84-a7bf-b40491a863c7\") " pod="openstack/rabbitmq-cell1-server-0" Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.515457 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/156fec50-b486-4e84-a7bf-b40491a863c7-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"156fec50-b486-4e84-a7bf-b40491a863c7\") " pod="openstack/rabbitmq-cell1-server-0" Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.515474 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/156fec50-b486-4e84-a7bf-b40491a863c7-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"156fec50-b486-4e84-a7bf-b40491a863c7\") " pod="openstack/rabbitmq-cell1-server-0" Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.515547 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/156fec50-b486-4e84-a7bf-b40491a863c7-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"156fec50-b486-4e84-a7bf-b40491a863c7\") " pod="openstack/rabbitmq-cell1-server-0" Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.515575 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/156fec50-b486-4e84-a7bf-b40491a863c7-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"156fec50-b486-4e84-a7bf-b40491a863c7\") " pod="openstack/rabbitmq-cell1-server-0" Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.515602 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/156fec50-b486-4e84-a7bf-b40491a863c7-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"156fec50-b486-4e84-a7bf-b40491a863c7\") " pod="openstack/rabbitmq-cell1-server-0" Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.515811 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/156fec50-b486-4e84-a7bf-b40491a863c7-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"156fec50-b486-4e84-a7bf-b40491a863c7\") " pod="openstack/rabbitmq-cell1-server-0" Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.515995 4824 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"156fec50-b486-4e84-a7bf-b40491a863c7\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/rabbitmq-cell1-server-0" Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.516734 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/156fec50-b486-4e84-a7bf-b40491a863c7-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"156fec50-b486-4e84-a7bf-b40491a863c7\") " pod="openstack/rabbitmq-cell1-server-0" Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.517429 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/156fec50-b486-4e84-a7bf-b40491a863c7-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"156fec50-b486-4e84-a7bf-b40491a863c7\") " pod="openstack/rabbitmq-cell1-server-0" Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.518773 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/156fec50-b486-4e84-a7bf-b40491a863c7-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"156fec50-b486-4e84-a7bf-b40491a863c7\") " pod="openstack/rabbitmq-cell1-server-0" Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.519310 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/156fec50-b486-4e84-a7bf-b40491a863c7-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"156fec50-b486-4e84-a7bf-b40491a863c7\") " pod="openstack/rabbitmq-cell1-server-0" Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.519554 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/156fec50-b486-4e84-a7bf-b40491a863c7-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"156fec50-b486-4e84-a7bf-b40491a863c7\") " pod="openstack/rabbitmq-cell1-server-0" Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.519851 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/156fec50-b486-4e84-a7bf-b40491a863c7-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"156fec50-b486-4e84-a7bf-b40491a863c7\") " pod="openstack/rabbitmq-cell1-server-0" Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.522441 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/156fec50-b486-4e84-a7bf-b40491a863c7-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"156fec50-b486-4e84-a7bf-b40491a863c7\") " pod="openstack/rabbitmq-cell1-server-0" Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.530712 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j9nc5\" (UniqueName: \"kubernetes.io/projected/156fec50-b486-4e84-a7bf-b40491a863c7-kube-api-access-j9nc5\") pod \"rabbitmq-cell1-server-0\" (UID: \"156fec50-b486-4e84-a7bf-b40491a863c7\") " pod="openstack/rabbitmq-cell1-server-0" Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.530938 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"156fec50-b486-4e84-a7bf-b40491a863c7\") " pod="openstack/rabbitmq-cell1-server-0" Jan 21 11:25:11 crc kubenswrapper[4824]: I0121 11:25:11.630706 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Jan 21 11:25:12 crc kubenswrapper[4824]: I0121 11:25:12.647872 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Jan 21 11:25:12 crc kubenswrapper[4824]: I0121 11:25:12.658821 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Jan 21 11:25:12 crc kubenswrapper[4824]: I0121 11:25:12.658910 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Jan 21 11:25:12 crc kubenswrapper[4824]: I0121 11:25:12.667322 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Jan 21 11:25:12 crc kubenswrapper[4824]: I0121 11:25:12.668033 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Jan 21 11:25:12 crc kubenswrapper[4824]: I0121 11:25:12.668145 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-wrlfw" Jan 21 11:25:12 crc kubenswrapper[4824]: I0121 11:25:12.668745 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Jan 21 11:25:12 crc kubenswrapper[4824]: I0121 11:25:12.670499 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Jan 21 11:25:12 crc kubenswrapper[4824]: I0121 11:25:12.833627 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/218aca21-4c37-4862-b891-fc228baa5348-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"218aca21-4c37-4862-b891-fc228baa5348\") " pod="openstack/openstack-galera-0" Jan 21 11:25:12 crc kubenswrapper[4824]: I0121 11:25:12.833708 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/218aca21-4c37-4862-b891-fc228baa5348-config-data-default\") pod \"openstack-galera-0\" (UID: \"218aca21-4c37-4862-b891-fc228baa5348\") " pod="openstack/openstack-galera-0" Jan 21 11:25:12 crc kubenswrapper[4824]: I0121 11:25:12.833733 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/218aca21-4c37-4862-b891-fc228baa5348-config-data-generated\") pod \"openstack-galera-0\" (UID: \"218aca21-4c37-4862-b891-fc228baa5348\") " pod="openstack/openstack-galera-0" Jan 21 11:25:12 crc kubenswrapper[4824]: I0121 11:25:12.833751 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/218aca21-4c37-4862-b891-fc228baa5348-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"218aca21-4c37-4862-b891-fc228baa5348\") " pod="openstack/openstack-galera-0" Jan 21 11:25:12 crc kubenswrapper[4824]: I0121 11:25:12.833774 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/218aca21-4c37-4862-b891-fc228baa5348-kolla-config\") pod \"openstack-galera-0\" (UID: \"218aca21-4c37-4862-b891-fc228baa5348\") " pod="openstack/openstack-galera-0" Jan 21 11:25:12 crc kubenswrapper[4824]: I0121 11:25:12.833816 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-galera-0\" (UID: \"218aca21-4c37-4862-b891-fc228baa5348\") " pod="openstack/openstack-galera-0" Jan 21 11:25:12 crc kubenswrapper[4824]: I0121 11:25:12.833861 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/218aca21-4c37-4862-b891-fc228baa5348-operator-scripts\") pod \"openstack-galera-0\" (UID: \"218aca21-4c37-4862-b891-fc228baa5348\") " pod="openstack/openstack-galera-0" Jan 21 11:25:12 crc kubenswrapper[4824]: I0121 11:25:12.833876 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bkxx7\" (UniqueName: \"kubernetes.io/projected/218aca21-4c37-4862-b891-fc228baa5348-kube-api-access-bkxx7\") pod \"openstack-galera-0\" (UID: \"218aca21-4c37-4862-b891-fc228baa5348\") " pod="openstack/openstack-galera-0" Jan 21 11:25:12 crc kubenswrapper[4824]: I0121 11:25:12.934933 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/218aca21-4c37-4862-b891-fc228baa5348-operator-scripts\") pod \"openstack-galera-0\" (UID: \"218aca21-4c37-4862-b891-fc228baa5348\") " pod="openstack/openstack-galera-0" Jan 21 11:25:12 crc kubenswrapper[4824]: I0121 11:25:12.934989 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bkxx7\" (UniqueName: \"kubernetes.io/projected/218aca21-4c37-4862-b891-fc228baa5348-kube-api-access-bkxx7\") pod \"openstack-galera-0\" (UID: \"218aca21-4c37-4862-b891-fc228baa5348\") " pod="openstack/openstack-galera-0" Jan 21 11:25:12 crc kubenswrapper[4824]: I0121 11:25:12.935029 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/218aca21-4c37-4862-b891-fc228baa5348-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"218aca21-4c37-4862-b891-fc228baa5348\") " pod="openstack/openstack-galera-0" Jan 21 11:25:12 crc kubenswrapper[4824]: I0121 11:25:12.935102 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/218aca21-4c37-4862-b891-fc228baa5348-config-data-default\") pod \"openstack-galera-0\" (UID: \"218aca21-4c37-4862-b891-fc228baa5348\") " pod="openstack/openstack-galera-0" Jan 21 11:25:12 crc kubenswrapper[4824]: I0121 11:25:12.935123 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/218aca21-4c37-4862-b891-fc228baa5348-config-data-generated\") pod \"openstack-galera-0\" (UID: \"218aca21-4c37-4862-b891-fc228baa5348\") " pod="openstack/openstack-galera-0" Jan 21 11:25:12 crc kubenswrapper[4824]: I0121 11:25:12.935147 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/218aca21-4c37-4862-b891-fc228baa5348-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"218aca21-4c37-4862-b891-fc228baa5348\") " pod="openstack/openstack-galera-0" Jan 21 11:25:12 crc kubenswrapper[4824]: I0121 11:25:12.935179 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/218aca21-4c37-4862-b891-fc228baa5348-kolla-config\") pod \"openstack-galera-0\" (UID: \"218aca21-4c37-4862-b891-fc228baa5348\") " pod="openstack/openstack-galera-0" Jan 21 11:25:12 crc kubenswrapper[4824]: I0121 11:25:12.935203 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-galera-0\" (UID: \"218aca21-4c37-4862-b891-fc228baa5348\") " pod="openstack/openstack-galera-0" Jan 21 11:25:12 crc kubenswrapper[4824]: I0121 11:25:12.935439 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/218aca21-4c37-4862-b891-fc228baa5348-config-data-generated\") pod \"openstack-galera-0\" (UID: \"218aca21-4c37-4862-b891-fc228baa5348\") " pod="openstack/openstack-galera-0" Jan 21 11:25:12 crc kubenswrapper[4824]: I0121 11:25:12.937156 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/218aca21-4c37-4862-b891-fc228baa5348-config-data-default\") pod \"openstack-galera-0\" (UID: \"218aca21-4c37-4862-b891-fc228baa5348\") " pod="openstack/openstack-galera-0" Jan 21 11:25:12 crc kubenswrapper[4824]: I0121 11:25:12.937236 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/218aca21-4c37-4862-b891-fc228baa5348-kolla-config\") pod \"openstack-galera-0\" (UID: \"218aca21-4c37-4862-b891-fc228baa5348\") " pod="openstack/openstack-galera-0" Jan 21 11:25:12 crc kubenswrapper[4824]: I0121 11:25:12.937315 4824 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-galera-0\" (UID: \"218aca21-4c37-4862-b891-fc228baa5348\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/openstack-galera-0" Jan 21 11:25:12 crc kubenswrapper[4824]: I0121 11:25:12.938469 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/218aca21-4c37-4862-b891-fc228baa5348-operator-scripts\") pod \"openstack-galera-0\" (UID: \"218aca21-4c37-4862-b891-fc228baa5348\") " pod="openstack/openstack-galera-0" Jan 21 11:25:12 crc kubenswrapper[4824]: I0121 11:25:12.948244 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/218aca21-4c37-4862-b891-fc228baa5348-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"218aca21-4c37-4862-b891-fc228baa5348\") " pod="openstack/openstack-galera-0" Jan 21 11:25:12 crc kubenswrapper[4824]: I0121 11:25:12.950854 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/218aca21-4c37-4862-b891-fc228baa5348-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"218aca21-4c37-4862-b891-fc228baa5348\") " pod="openstack/openstack-galera-0" Jan 21 11:25:12 crc kubenswrapper[4824]: I0121 11:25:12.961307 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bkxx7\" (UniqueName: \"kubernetes.io/projected/218aca21-4c37-4862-b891-fc228baa5348-kube-api-access-bkxx7\") pod \"openstack-galera-0\" (UID: \"218aca21-4c37-4862-b891-fc228baa5348\") " pod="openstack/openstack-galera-0" Jan 21 11:25:12 crc kubenswrapper[4824]: I0121 11:25:12.965464 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-galera-0\" (UID: \"218aca21-4c37-4862-b891-fc228baa5348\") " pod="openstack/openstack-galera-0" Jan 21 11:25:12 crc kubenswrapper[4824]: I0121 11:25:12.975443 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Jan 21 11:25:14 crc kubenswrapper[4824]: I0121 11:25:14.001298 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Jan 21 11:25:14 crc kubenswrapper[4824]: I0121 11:25:14.005440 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Jan 21 11:25:14 crc kubenswrapper[4824]: I0121 11:25:14.007247 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Jan 21 11:25:14 crc kubenswrapper[4824]: I0121 11:25:14.007925 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Jan 21 11:25:14 crc kubenswrapper[4824]: I0121 11:25:14.008065 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-4hq4q" Jan 21 11:25:14 crc kubenswrapper[4824]: I0121 11:25:14.008242 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Jan 21 11:25:14 crc kubenswrapper[4824]: I0121 11:25:14.019421 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Jan 21 11:25:14 crc kubenswrapper[4824]: I0121 11:25:14.054570 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tt9bn\" (UniqueName: \"kubernetes.io/projected/7cd8d62b-ab74-4a1b-95f5-7253ed6c0346-kube-api-access-tt9bn\") pod \"openstack-cell1-galera-0\" (UID: \"7cd8d62b-ab74-4a1b-95f5-7253ed6c0346\") " pod="openstack/openstack-cell1-galera-0" Jan 21 11:25:14 crc kubenswrapper[4824]: I0121 11:25:14.054814 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/7cd8d62b-ab74-4a1b-95f5-7253ed6c0346-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"7cd8d62b-ab74-4a1b-95f5-7253ed6c0346\") " pod="openstack/openstack-cell1-galera-0" Jan 21 11:25:14 crc kubenswrapper[4824]: I0121 11:25:14.054861 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7cd8d62b-ab74-4a1b-95f5-7253ed6c0346-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"7cd8d62b-ab74-4a1b-95f5-7253ed6c0346\") " pod="openstack/openstack-cell1-galera-0" Jan 21 11:25:14 crc kubenswrapper[4824]: I0121 11:25:14.054898 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/7cd8d62b-ab74-4a1b-95f5-7253ed6c0346-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"7cd8d62b-ab74-4a1b-95f5-7253ed6c0346\") " pod="openstack/openstack-cell1-galera-0" Jan 21 11:25:14 crc kubenswrapper[4824]: I0121 11:25:14.054974 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"openstack-cell1-galera-0\" (UID: \"7cd8d62b-ab74-4a1b-95f5-7253ed6c0346\") " pod="openstack/openstack-cell1-galera-0" Jan 21 11:25:14 crc kubenswrapper[4824]: I0121 11:25:14.055002 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7cd8d62b-ab74-4a1b-95f5-7253ed6c0346-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"7cd8d62b-ab74-4a1b-95f5-7253ed6c0346\") " pod="openstack/openstack-cell1-galera-0" Jan 21 11:25:14 crc kubenswrapper[4824]: I0121 11:25:14.055047 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/7cd8d62b-ab74-4a1b-95f5-7253ed6c0346-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"7cd8d62b-ab74-4a1b-95f5-7253ed6c0346\") " pod="openstack/openstack-cell1-galera-0" Jan 21 11:25:14 crc kubenswrapper[4824]: I0121 11:25:14.055067 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/7cd8d62b-ab74-4a1b-95f5-7253ed6c0346-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"7cd8d62b-ab74-4a1b-95f5-7253ed6c0346\") " pod="openstack/openstack-cell1-galera-0" Jan 21 11:25:14 crc kubenswrapper[4824]: I0121 11:25:14.156597 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/7cd8d62b-ab74-4a1b-95f5-7253ed6c0346-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"7cd8d62b-ab74-4a1b-95f5-7253ed6c0346\") " pod="openstack/openstack-cell1-galera-0" Jan 21 11:25:14 crc kubenswrapper[4824]: I0121 11:25:14.156640 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/7cd8d62b-ab74-4a1b-95f5-7253ed6c0346-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"7cd8d62b-ab74-4a1b-95f5-7253ed6c0346\") " pod="openstack/openstack-cell1-galera-0" Jan 21 11:25:14 crc kubenswrapper[4824]: I0121 11:25:14.156745 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tt9bn\" (UniqueName: \"kubernetes.io/projected/7cd8d62b-ab74-4a1b-95f5-7253ed6c0346-kube-api-access-tt9bn\") pod \"openstack-cell1-galera-0\" (UID: \"7cd8d62b-ab74-4a1b-95f5-7253ed6c0346\") " pod="openstack/openstack-cell1-galera-0" Jan 21 11:25:14 crc kubenswrapper[4824]: I0121 11:25:14.156796 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/7cd8d62b-ab74-4a1b-95f5-7253ed6c0346-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"7cd8d62b-ab74-4a1b-95f5-7253ed6c0346\") " pod="openstack/openstack-cell1-galera-0" Jan 21 11:25:14 crc kubenswrapper[4824]: I0121 11:25:14.156821 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7cd8d62b-ab74-4a1b-95f5-7253ed6c0346-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"7cd8d62b-ab74-4a1b-95f5-7253ed6c0346\") " pod="openstack/openstack-cell1-galera-0" Jan 21 11:25:14 crc kubenswrapper[4824]: I0121 11:25:14.156845 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/7cd8d62b-ab74-4a1b-95f5-7253ed6c0346-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"7cd8d62b-ab74-4a1b-95f5-7253ed6c0346\") " pod="openstack/openstack-cell1-galera-0" Jan 21 11:25:14 crc kubenswrapper[4824]: I0121 11:25:14.156895 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"openstack-cell1-galera-0\" (UID: \"7cd8d62b-ab74-4a1b-95f5-7253ed6c0346\") " pod="openstack/openstack-cell1-galera-0" Jan 21 11:25:14 crc kubenswrapper[4824]: I0121 11:25:14.156918 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7cd8d62b-ab74-4a1b-95f5-7253ed6c0346-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"7cd8d62b-ab74-4a1b-95f5-7253ed6c0346\") " pod="openstack/openstack-cell1-galera-0" Jan 21 11:25:14 crc kubenswrapper[4824]: I0121 11:25:14.157051 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/7cd8d62b-ab74-4a1b-95f5-7253ed6c0346-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"7cd8d62b-ab74-4a1b-95f5-7253ed6c0346\") " pod="openstack/openstack-cell1-galera-0" Jan 21 11:25:14 crc kubenswrapper[4824]: I0121 11:25:14.157580 4824 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"openstack-cell1-galera-0\" (UID: \"7cd8d62b-ab74-4a1b-95f5-7253ed6c0346\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/openstack-cell1-galera-0" Jan 21 11:25:14 crc kubenswrapper[4824]: I0121 11:25:14.157642 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/7cd8d62b-ab74-4a1b-95f5-7253ed6c0346-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"7cd8d62b-ab74-4a1b-95f5-7253ed6c0346\") " pod="openstack/openstack-cell1-galera-0" Jan 21 11:25:14 crc kubenswrapper[4824]: I0121 11:25:14.158020 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/7cd8d62b-ab74-4a1b-95f5-7253ed6c0346-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"7cd8d62b-ab74-4a1b-95f5-7253ed6c0346\") " pod="openstack/openstack-cell1-galera-0" Jan 21 11:25:14 crc kubenswrapper[4824]: I0121 11:25:14.158798 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7cd8d62b-ab74-4a1b-95f5-7253ed6c0346-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"7cd8d62b-ab74-4a1b-95f5-7253ed6c0346\") " pod="openstack/openstack-cell1-galera-0" Jan 21 11:25:14 crc kubenswrapper[4824]: I0121 11:25:14.160646 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/7cd8d62b-ab74-4a1b-95f5-7253ed6c0346-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"7cd8d62b-ab74-4a1b-95f5-7253ed6c0346\") " pod="openstack/openstack-cell1-galera-0" Jan 21 11:25:14 crc kubenswrapper[4824]: I0121 11:25:14.160767 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7cd8d62b-ab74-4a1b-95f5-7253ed6c0346-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"7cd8d62b-ab74-4a1b-95f5-7253ed6c0346\") " pod="openstack/openstack-cell1-galera-0" Jan 21 11:25:14 crc kubenswrapper[4824]: I0121 11:25:14.175238 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"openstack-cell1-galera-0\" (UID: \"7cd8d62b-ab74-4a1b-95f5-7253ed6c0346\") " pod="openstack/openstack-cell1-galera-0" Jan 21 11:25:14 crc kubenswrapper[4824]: I0121 11:25:14.177000 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tt9bn\" (UniqueName: \"kubernetes.io/projected/7cd8d62b-ab74-4a1b-95f5-7253ed6c0346-kube-api-access-tt9bn\") pod \"openstack-cell1-galera-0\" (UID: \"7cd8d62b-ab74-4a1b-95f5-7253ed6c0346\") " pod="openstack/openstack-cell1-galera-0" Jan 21 11:25:14 crc kubenswrapper[4824]: I0121 11:25:14.262654 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Jan 21 11:25:14 crc kubenswrapper[4824]: I0121 11:25:14.266925 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Jan 21 11:25:14 crc kubenswrapper[4824]: I0121 11:25:14.270446 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Jan 21 11:25:14 crc kubenswrapper[4824]: I0121 11:25:14.270839 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-vxswb" Jan 21 11:25:14 crc kubenswrapper[4824]: I0121 11:25:14.276297 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Jan 21 11:25:14 crc kubenswrapper[4824]: I0121 11:25:14.280582 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Jan 21 11:25:14 crc kubenswrapper[4824]: I0121 11:25:14.324417 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Jan 21 11:25:14 crc kubenswrapper[4824]: I0121 11:25:14.463160 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/dba0050b-8a73-4355-a1e0-7c9a03557ead-memcached-tls-certs\") pod \"memcached-0\" (UID: \"dba0050b-8a73-4355-a1e0-7c9a03557ead\") " pod="openstack/memcached-0" Jan 21 11:25:14 crc kubenswrapper[4824]: I0121 11:25:14.463237 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7r5tc\" (UniqueName: \"kubernetes.io/projected/dba0050b-8a73-4355-a1e0-7c9a03557ead-kube-api-access-7r5tc\") pod \"memcached-0\" (UID: \"dba0050b-8a73-4355-a1e0-7c9a03557ead\") " pod="openstack/memcached-0" Jan 21 11:25:14 crc kubenswrapper[4824]: I0121 11:25:14.463258 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dba0050b-8a73-4355-a1e0-7c9a03557ead-combined-ca-bundle\") pod \"memcached-0\" (UID: \"dba0050b-8a73-4355-a1e0-7c9a03557ead\") " pod="openstack/memcached-0" Jan 21 11:25:14 crc kubenswrapper[4824]: I0121 11:25:14.463300 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/dba0050b-8a73-4355-a1e0-7c9a03557ead-kolla-config\") pod \"memcached-0\" (UID: \"dba0050b-8a73-4355-a1e0-7c9a03557ead\") " pod="openstack/memcached-0" Jan 21 11:25:14 crc kubenswrapper[4824]: I0121 11:25:14.463405 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/dba0050b-8a73-4355-a1e0-7c9a03557ead-config-data\") pod \"memcached-0\" (UID: \"dba0050b-8a73-4355-a1e0-7c9a03557ead\") " pod="openstack/memcached-0" Jan 21 11:25:14 crc kubenswrapper[4824]: I0121 11:25:14.565249 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/dba0050b-8a73-4355-a1e0-7c9a03557ead-memcached-tls-certs\") pod \"memcached-0\" (UID: \"dba0050b-8a73-4355-a1e0-7c9a03557ead\") " pod="openstack/memcached-0" Jan 21 11:25:14 crc kubenswrapper[4824]: I0121 11:25:14.565316 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7r5tc\" (UniqueName: \"kubernetes.io/projected/dba0050b-8a73-4355-a1e0-7c9a03557ead-kube-api-access-7r5tc\") pod \"memcached-0\" (UID: \"dba0050b-8a73-4355-a1e0-7c9a03557ead\") " pod="openstack/memcached-0" Jan 21 11:25:14 crc kubenswrapper[4824]: I0121 11:25:14.565336 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dba0050b-8a73-4355-a1e0-7c9a03557ead-combined-ca-bundle\") pod \"memcached-0\" (UID: \"dba0050b-8a73-4355-a1e0-7c9a03557ead\") " pod="openstack/memcached-0" Jan 21 11:25:14 crc kubenswrapper[4824]: I0121 11:25:14.565374 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/dba0050b-8a73-4355-a1e0-7c9a03557ead-kolla-config\") pod \"memcached-0\" (UID: \"dba0050b-8a73-4355-a1e0-7c9a03557ead\") " pod="openstack/memcached-0" Jan 21 11:25:14 crc kubenswrapper[4824]: I0121 11:25:14.565419 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/dba0050b-8a73-4355-a1e0-7c9a03557ead-config-data\") pod \"memcached-0\" (UID: \"dba0050b-8a73-4355-a1e0-7c9a03557ead\") " pod="openstack/memcached-0" Jan 21 11:25:14 crc kubenswrapper[4824]: I0121 11:25:14.566099 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/dba0050b-8a73-4355-a1e0-7c9a03557ead-config-data\") pod \"memcached-0\" (UID: \"dba0050b-8a73-4355-a1e0-7c9a03557ead\") " pod="openstack/memcached-0" Jan 21 11:25:14 crc kubenswrapper[4824]: I0121 11:25:14.566282 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/dba0050b-8a73-4355-a1e0-7c9a03557ead-kolla-config\") pod \"memcached-0\" (UID: \"dba0050b-8a73-4355-a1e0-7c9a03557ead\") " pod="openstack/memcached-0" Jan 21 11:25:14 crc kubenswrapper[4824]: I0121 11:25:14.570352 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/dba0050b-8a73-4355-a1e0-7c9a03557ead-memcached-tls-certs\") pod \"memcached-0\" (UID: \"dba0050b-8a73-4355-a1e0-7c9a03557ead\") " pod="openstack/memcached-0" Jan 21 11:25:14 crc kubenswrapper[4824]: I0121 11:25:14.572123 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dba0050b-8a73-4355-a1e0-7c9a03557ead-combined-ca-bundle\") pod \"memcached-0\" (UID: \"dba0050b-8a73-4355-a1e0-7c9a03557ead\") " pod="openstack/memcached-0" Jan 21 11:25:14 crc kubenswrapper[4824]: I0121 11:25:14.580654 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7r5tc\" (UniqueName: \"kubernetes.io/projected/dba0050b-8a73-4355-a1e0-7c9a03557ead-kube-api-access-7r5tc\") pod \"memcached-0\" (UID: \"dba0050b-8a73-4355-a1e0-7c9a03557ead\") " pod="openstack/memcached-0" Jan 21 11:25:14 crc kubenswrapper[4824]: I0121 11:25:14.590706 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Jan 21 11:25:16 crc kubenswrapper[4824]: I0121 11:25:16.106715 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Jan 21 11:25:16 crc kubenswrapper[4824]: I0121 11:25:16.107903 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Jan 21 11:25:16 crc kubenswrapper[4824]: I0121 11:25:16.109346 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-464bp" Jan 21 11:25:16 crc kubenswrapper[4824]: I0121 11:25:16.124321 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Jan 21 11:25:16 crc kubenswrapper[4824]: I0121 11:25:16.288516 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n6ndv\" (UniqueName: \"kubernetes.io/projected/2e3a943c-ee06-4e54-a984-34ecb1320f89-kube-api-access-n6ndv\") pod \"kube-state-metrics-0\" (UID: \"2e3a943c-ee06-4e54-a984-34ecb1320f89\") " pod="openstack/kube-state-metrics-0" Jan 21 11:25:16 crc kubenswrapper[4824]: I0121 11:25:16.390303 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n6ndv\" (UniqueName: \"kubernetes.io/projected/2e3a943c-ee06-4e54-a984-34ecb1320f89-kube-api-access-n6ndv\") pod \"kube-state-metrics-0\" (UID: \"2e3a943c-ee06-4e54-a984-34ecb1320f89\") " pod="openstack/kube-state-metrics-0" Jan 21 11:25:16 crc kubenswrapper[4824]: I0121 11:25:16.408156 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n6ndv\" (UniqueName: \"kubernetes.io/projected/2e3a943c-ee06-4e54-a984-34ecb1320f89-kube-api-access-n6ndv\") pod \"kube-state-metrics-0\" (UID: \"2e3a943c-ee06-4e54-a984-34ecb1320f89\") " pod="openstack/kube-state-metrics-0" Jan 21 11:25:16 crc kubenswrapper[4824]: I0121 11:25:16.429098 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Jan 21 11:25:19 crc kubenswrapper[4824]: I0121 11:25:19.027448 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Jan 21 11:25:19 crc kubenswrapper[4824]: W0121 11:25:19.033844 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod156fec50_b486_4e84_a7bf_b40491a863c7.slice/crio-37d36a29724a133bc6f138979451aa8f9c0eeec9cad658937121a853dcbd295e WatchSource:0}: Error finding container 37d36a29724a133bc6f138979451aa8f9c0eeec9cad658937121a853dcbd295e: Status 404 returned error can't find the container with id 37d36a29724a133bc6f138979451aa8f9c0eeec9cad658937121a853dcbd295e Jan 21 11:25:19 crc kubenswrapper[4824]: I0121 11:25:19.035950 4824 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 21 11:25:19 crc kubenswrapper[4824]: I0121 11:25:19.040221 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-zspkp"] Jan 21 11:25:19 crc kubenswrapper[4824]: W0121 11:25:19.045246 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf67dfc52_dc0e_404e_af22_75c6a3d10721.slice/crio-4c0d7f6c5bc10bc5e3cf172c7b361f178beb8f1decb24246db808a7099969563 WatchSource:0}: Error finding container 4c0d7f6c5bc10bc5e3cf172c7b361f178beb8f1decb24246db808a7099969563: Status 404 returned error can't find the container with id 4c0d7f6c5bc10bc5e3cf172c7b361f178beb8f1decb24246db808a7099969563 Jan 21 11:25:19 crc kubenswrapper[4824]: I0121 11:25:19.142211 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Jan 21 11:25:19 crc kubenswrapper[4824]: I0121 11:25:19.147392 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Jan 21 11:25:19 crc kubenswrapper[4824]: I0121 11:25:19.226467 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-nrqwx"] Jan 21 11:25:19 crc kubenswrapper[4824]: I0121 11:25:19.230618 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Jan 21 11:25:19 crc kubenswrapper[4824]: I0121 11:25:19.241180 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Jan 21 11:25:19 crc kubenswrapper[4824]: I0121 11:25:19.245725 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Jan 21 11:25:19 crc kubenswrapper[4824]: I0121 11:25:19.461368 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"2e3a943c-ee06-4e54-a984-34ecb1320f89","Type":"ContainerStarted","Data":"6cb5ad5332507e249a47db656294541a580560b8688929dae90a9899b846d8f8"} Jan 21 11:25:19 crc kubenswrapper[4824]: I0121 11:25:19.462363 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"156fec50-b486-4e84-a7bf-b40491a863c7","Type":"ContainerStarted","Data":"37d36a29724a133bc6f138979451aa8f9c0eeec9cad658937121a853dcbd295e"} Jan 21 11:25:19 crc kubenswrapper[4824]: I0121 11:25:19.463771 4824 generic.go:334] "Generic (PLEG): container finished" podID="f67dfc52-dc0e-404e-af22-75c6a3d10721" containerID="2fabf442546067371acccd858ab70f06b5ad7ea6a02d270ec355916d978e1384" exitCode=0 Jan 21 11:25:19 crc kubenswrapper[4824]: I0121 11:25:19.463810 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-zspkp" event={"ID":"f67dfc52-dc0e-404e-af22-75c6a3d10721","Type":"ContainerDied","Data":"2fabf442546067371acccd858ab70f06b5ad7ea6a02d270ec355916d978e1384"} Jan 21 11:25:19 crc kubenswrapper[4824]: I0121 11:25:19.463826 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-zspkp" event={"ID":"f67dfc52-dc0e-404e-af22-75c6a3d10721","Type":"ContainerStarted","Data":"4c0d7f6c5bc10bc5e3cf172c7b361f178beb8f1decb24246db808a7099969563"} Jan 21 11:25:19 crc kubenswrapper[4824]: I0121 11:25:19.465171 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"84e74d7b-18a0-4a3f-8680-6246ac538a6e","Type":"ContainerStarted","Data":"8ee895c662c7814313d66afa49854eb939a2dc25462867e35d0362794391ee25"} Jan 21 11:25:19 crc kubenswrapper[4824]: I0121 11:25:19.466079 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"218aca21-4c37-4862-b891-fc228baa5348","Type":"ContainerStarted","Data":"bf9686d495d7804b469a74da5f65df348ba59c6886dccf9481b92b03528d8379"} Jan 21 11:25:19 crc kubenswrapper[4824]: I0121 11:25:19.467003 4824 generic.go:334] "Generic (PLEG): container finished" podID="71b408e9-2ba6-4b0e-9b4c-22faf4ef4fd4" containerID="cd51f4db187c81d9d459bc61d2b5fff2b45f07ab86bc1d1f49ffc846a1036059" exitCode=0 Jan 21 11:25:19 crc kubenswrapper[4824]: I0121 11:25:19.467109 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-nrqwx" event={"ID":"71b408e9-2ba6-4b0e-9b4c-22faf4ef4fd4","Type":"ContainerDied","Data":"cd51f4db187c81d9d459bc61d2b5fff2b45f07ab86bc1d1f49ffc846a1036059"} Jan 21 11:25:19 crc kubenswrapper[4824]: I0121 11:25:19.467130 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-nrqwx" event={"ID":"71b408e9-2ba6-4b0e-9b4c-22faf4ef4fd4","Type":"ContainerStarted","Data":"9f481192efadf26388108ffc156d5c70e082a7630e4e6aa574c82729d2674bf6"} Jan 21 11:25:19 crc kubenswrapper[4824]: I0121 11:25:19.468311 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"dba0050b-8a73-4355-a1e0-7c9a03557ead","Type":"ContainerStarted","Data":"0a266bda78772602cb7070fdd028643409e691a2f47d4b4ff003042c9da515e1"} Jan 21 11:25:19 crc kubenswrapper[4824]: I0121 11:25:19.469342 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"7cd8d62b-ab74-4a1b-95f5-7253ed6c0346","Type":"ContainerStarted","Data":"e38b6eb1d31e1293e699d97ffa6c84831dffe4bbdbcaad6d49e66ab12b79817c"} Jan 21 11:25:19 crc kubenswrapper[4824]: I0121 11:25:19.470857 4824 generic.go:334] "Generic (PLEG): container finished" podID="b48ab67f-88cb-4921-b6a3-38b77b534bf9" containerID="415de3e8643b808347c13b2c7d4d57542a514312e1a5bd379cbc7b434241d117" exitCode=0 Jan 21 11:25:19 crc kubenswrapper[4824]: I0121 11:25:19.470902 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-lmtkp" event={"ID":"b48ab67f-88cb-4921-b6a3-38b77b534bf9","Type":"ContainerDied","Data":"415de3e8643b808347c13b2c7d4d57542a514312e1a5bd379cbc7b434241d117"} Jan 21 11:25:19 crc kubenswrapper[4824]: I0121 11:25:19.472284 4824 generic.go:334] "Generic (PLEG): container finished" podID="e8aca65b-76cd-45f9-8b20-da2e34a41fc2" containerID="aaacc227843f224a20462553c49fdbfc00aeab8ce2765ad6ee4c575c381c5973" exitCode=0 Jan 21 11:25:19 crc kubenswrapper[4824]: I0121 11:25:19.472314 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-tmr2j" event={"ID":"e8aca65b-76cd-45f9-8b20-da2e34a41fc2","Type":"ContainerDied","Data":"aaacc227843f224a20462553c49fdbfc00aeab8ce2765ad6ee4c575c381c5973"} Jan 21 11:25:19 crc kubenswrapper[4824]: I0121 11:25:19.775886 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-htwd6"] Jan 21 11:25:19 crc kubenswrapper[4824]: I0121 11:25:19.776848 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-htwd6" Jan 21 11:25:19 crc kubenswrapper[4824]: I0121 11:25:19.784748 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Jan 21 11:25:19 crc kubenswrapper[4824]: I0121 11:25:19.784951 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Jan 21 11:25:19 crc kubenswrapper[4824]: I0121 11:25:19.785106 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-ktgdl" Jan 21 11:25:19 crc kubenswrapper[4824]: I0121 11:25:19.798775 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-htwd6"] Jan 21 11:25:19 crc kubenswrapper[4824]: I0121 11:25:19.831513 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-gr9gj"] Jan 21 11:25:19 crc kubenswrapper[4824]: I0121 11:25:19.833539 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-gr9gj" Jan 21 11:25:19 crc kubenswrapper[4824]: I0121 11:25:19.845287 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-gr9gj"] Jan 21 11:25:19 crc kubenswrapper[4824]: I0121 11:25:19.915253 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-tmr2j" Jan 21 11:25:19 crc kubenswrapper[4824]: I0121 11:25:19.936183 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-lmtkp" Jan 21 11:25:19 crc kubenswrapper[4824]: I0121 11:25:19.956514 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/b27b3f77-8476-4f3e-a946-f4f26fc66cb5-var-lib\") pod \"ovn-controller-ovs-gr9gj\" (UID: \"b27b3f77-8476-4f3e-a946-f4f26fc66cb5\") " pod="openstack/ovn-controller-ovs-gr9gj" Jan 21 11:25:19 crc kubenswrapper[4824]: I0121 11:25:19.957302 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b27b3f77-8476-4f3e-a946-f4f26fc66cb5-scripts\") pod \"ovn-controller-ovs-gr9gj\" (UID: \"b27b3f77-8476-4f3e-a946-f4f26fc66cb5\") " pod="openstack/ovn-controller-ovs-gr9gj" Jan 21 11:25:19 crc kubenswrapper[4824]: I0121 11:25:19.957354 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/511f2b6d-a08d-49f8-b393-ab222219d4a7-var-run-ovn\") pod \"ovn-controller-htwd6\" (UID: \"511f2b6d-a08d-49f8-b393-ab222219d4a7\") " pod="openstack/ovn-controller-htwd6" Jan 21 11:25:19 crc kubenswrapper[4824]: I0121 11:25:19.957375 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/b27b3f77-8476-4f3e-a946-f4f26fc66cb5-var-log\") pod \"ovn-controller-ovs-gr9gj\" (UID: \"b27b3f77-8476-4f3e-a946-f4f26fc66cb5\") " pod="openstack/ovn-controller-ovs-gr9gj" Jan 21 11:25:19 crc kubenswrapper[4824]: I0121 11:25:19.957389 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/511f2b6d-a08d-49f8-b393-ab222219d4a7-combined-ca-bundle\") pod \"ovn-controller-htwd6\" (UID: \"511f2b6d-a08d-49f8-b393-ab222219d4a7\") " pod="openstack/ovn-controller-htwd6" Jan 21 11:25:19 crc kubenswrapper[4824]: I0121 11:25:19.957412 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/511f2b6d-a08d-49f8-b393-ab222219d4a7-var-run\") pod \"ovn-controller-htwd6\" (UID: \"511f2b6d-a08d-49f8-b393-ab222219d4a7\") " pod="openstack/ovn-controller-htwd6" Jan 21 11:25:19 crc kubenswrapper[4824]: I0121 11:25:19.957445 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/511f2b6d-a08d-49f8-b393-ab222219d4a7-ovn-controller-tls-certs\") pod \"ovn-controller-htwd6\" (UID: \"511f2b6d-a08d-49f8-b393-ab222219d4a7\") " pod="openstack/ovn-controller-htwd6" Jan 21 11:25:19 crc kubenswrapper[4824]: I0121 11:25:19.957495 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/511f2b6d-a08d-49f8-b393-ab222219d4a7-scripts\") pod \"ovn-controller-htwd6\" (UID: \"511f2b6d-a08d-49f8-b393-ab222219d4a7\") " pod="openstack/ovn-controller-htwd6" Jan 21 11:25:19 crc kubenswrapper[4824]: I0121 11:25:19.957530 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/b27b3f77-8476-4f3e-a946-f4f26fc66cb5-var-run\") pod \"ovn-controller-ovs-gr9gj\" (UID: \"b27b3f77-8476-4f3e-a946-f4f26fc66cb5\") " pod="openstack/ovn-controller-ovs-gr9gj" Jan 21 11:25:19 crc kubenswrapper[4824]: I0121 11:25:19.957544 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cz8l5\" (UniqueName: \"kubernetes.io/projected/511f2b6d-a08d-49f8-b393-ab222219d4a7-kube-api-access-cz8l5\") pod \"ovn-controller-htwd6\" (UID: \"511f2b6d-a08d-49f8-b393-ab222219d4a7\") " pod="openstack/ovn-controller-htwd6" Jan 21 11:25:19 crc kubenswrapper[4824]: I0121 11:25:19.957766 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6cdxj\" (UniqueName: \"kubernetes.io/projected/b27b3f77-8476-4f3e-a946-f4f26fc66cb5-kube-api-access-6cdxj\") pod \"ovn-controller-ovs-gr9gj\" (UID: \"b27b3f77-8476-4f3e-a946-f4f26fc66cb5\") " pod="openstack/ovn-controller-ovs-gr9gj" Jan 21 11:25:19 crc kubenswrapper[4824]: I0121 11:25:19.957816 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/b27b3f77-8476-4f3e-a946-f4f26fc66cb5-etc-ovs\") pod \"ovn-controller-ovs-gr9gj\" (UID: \"b27b3f77-8476-4f3e-a946-f4f26fc66cb5\") " pod="openstack/ovn-controller-ovs-gr9gj" Jan 21 11:25:19 crc kubenswrapper[4824]: I0121 11:25:19.957838 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/511f2b6d-a08d-49f8-b393-ab222219d4a7-var-log-ovn\") pod \"ovn-controller-htwd6\" (UID: \"511f2b6d-a08d-49f8-b393-ab222219d4a7\") " pod="openstack/ovn-controller-htwd6" Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.058414 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8q4rz\" (UniqueName: \"kubernetes.io/projected/b48ab67f-88cb-4921-b6a3-38b77b534bf9-kube-api-access-8q4rz\") pod \"b48ab67f-88cb-4921-b6a3-38b77b534bf9\" (UID: \"b48ab67f-88cb-4921-b6a3-38b77b534bf9\") " Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.058577 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e8aca65b-76cd-45f9-8b20-da2e34a41fc2-dns-svc\") pod \"e8aca65b-76cd-45f9-8b20-da2e34a41fc2\" (UID: \"e8aca65b-76cd-45f9-8b20-da2e34a41fc2\") " Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.058676 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e8aca65b-76cd-45f9-8b20-da2e34a41fc2-config\") pod \"e8aca65b-76cd-45f9-8b20-da2e34a41fc2\" (UID: \"e8aca65b-76cd-45f9-8b20-da2e34a41fc2\") " Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.058736 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b48ab67f-88cb-4921-b6a3-38b77b534bf9-config\") pod \"b48ab67f-88cb-4921-b6a3-38b77b534bf9\" (UID: \"b48ab67f-88cb-4921-b6a3-38b77b534bf9\") " Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.058766 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jt49s\" (UniqueName: \"kubernetes.io/projected/e8aca65b-76cd-45f9-8b20-da2e34a41fc2-kube-api-access-jt49s\") pod \"e8aca65b-76cd-45f9-8b20-da2e34a41fc2\" (UID: \"e8aca65b-76cd-45f9-8b20-da2e34a41fc2\") " Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.058946 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/b27b3f77-8476-4f3e-a946-f4f26fc66cb5-etc-ovs\") pod \"ovn-controller-ovs-gr9gj\" (UID: \"b27b3f77-8476-4f3e-a946-f4f26fc66cb5\") " pod="openstack/ovn-controller-ovs-gr9gj" Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.058980 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/511f2b6d-a08d-49f8-b393-ab222219d4a7-var-log-ovn\") pod \"ovn-controller-htwd6\" (UID: \"511f2b6d-a08d-49f8-b393-ab222219d4a7\") " pod="openstack/ovn-controller-htwd6" Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.059045 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/b27b3f77-8476-4f3e-a946-f4f26fc66cb5-var-lib\") pod \"ovn-controller-ovs-gr9gj\" (UID: \"b27b3f77-8476-4f3e-a946-f4f26fc66cb5\") " pod="openstack/ovn-controller-ovs-gr9gj" Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.059065 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b27b3f77-8476-4f3e-a946-f4f26fc66cb5-scripts\") pod \"ovn-controller-ovs-gr9gj\" (UID: \"b27b3f77-8476-4f3e-a946-f4f26fc66cb5\") " pod="openstack/ovn-controller-ovs-gr9gj" Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.059087 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/b27b3f77-8476-4f3e-a946-f4f26fc66cb5-var-log\") pod \"ovn-controller-ovs-gr9gj\" (UID: \"b27b3f77-8476-4f3e-a946-f4f26fc66cb5\") " pod="openstack/ovn-controller-ovs-gr9gj" Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.059099 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/511f2b6d-a08d-49f8-b393-ab222219d4a7-var-run-ovn\") pod \"ovn-controller-htwd6\" (UID: \"511f2b6d-a08d-49f8-b393-ab222219d4a7\") " pod="openstack/ovn-controller-htwd6" Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.059116 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/511f2b6d-a08d-49f8-b393-ab222219d4a7-combined-ca-bundle\") pod \"ovn-controller-htwd6\" (UID: \"511f2b6d-a08d-49f8-b393-ab222219d4a7\") " pod="openstack/ovn-controller-htwd6" Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.059130 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/511f2b6d-a08d-49f8-b393-ab222219d4a7-var-run\") pod \"ovn-controller-htwd6\" (UID: \"511f2b6d-a08d-49f8-b393-ab222219d4a7\") " pod="openstack/ovn-controller-htwd6" Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.059432 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/511f2b6d-a08d-49f8-b393-ab222219d4a7-ovn-controller-tls-certs\") pod \"ovn-controller-htwd6\" (UID: \"511f2b6d-a08d-49f8-b393-ab222219d4a7\") " pod="openstack/ovn-controller-htwd6" Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.059942 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/511f2b6d-a08d-49f8-b393-ab222219d4a7-var-log-ovn\") pod \"ovn-controller-htwd6\" (UID: \"511f2b6d-a08d-49f8-b393-ab222219d4a7\") " pod="openstack/ovn-controller-htwd6" Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.060125 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/b27b3f77-8476-4f3e-a946-f4f26fc66cb5-var-log\") pod \"ovn-controller-ovs-gr9gj\" (UID: \"b27b3f77-8476-4f3e-a946-f4f26fc66cb5\") " pod="openstack/ovn-controller-ovs-gr9gj" Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.061448 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/b27b3f77-8476-4f3e-a946-f4f26fc66cb5-etc-ovs\") pod \"ovn-controller-ovs-gr9gj\" (UID: \"b27b3f77-8476-4f3e-a946-f4f26fc66cb5\") " pod="openstack/ovn-controller-ovs-gr9gj" Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.061624 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/511f2b6d-a08d-49f8-b393-ab222219d4a7-scripts\") pod \"ovn-controller-htwd6\" (UID: \"511f2b6d-a08d-49f8-b393-ab222219d4a7\") " pod="openstack/ovn-controller-htwd6" Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.061671 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/b27b3f77-8476-4f3e-a946-f4f26fc66cb5-var-run\") pod \"ovn-controller-ovs-gr9gj\" (UID: \"b27b3f77-8476-4f3e-a946-f4f26fc66cb5\") " pod="openstack/ovn-controller-ovs-gr9gj" Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.061702 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cz8l5\" (UniqueName: \"kubernetes.io/projected/511f2b6d-a08d-49f8-b393-ab222219d4a7-kube-api-access-cz8l5\") pod \"ovn-controller-htwd6\" (UID: \"511f2b6d-a08d-49f8-b393-ab222219d4a7\") " pod="openstack/ovn-controller-htwd6" Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.061719 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/b27b3f77-8476-4f3e-a946-f4f26fc66cb5-var-lib\") pod \"ovn-controller-ovs-gr9gj\" (UID: \"b27b3f77-8476-4f3e-a946-f4f26fc66cb5\") " pod="openstack/ovn-controller-ovs-gr9gj" Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.062162 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/511f2b6d-a08d-49f8-b393-ab222219d4a7-var-run\") pod \"ovn-controller-htwd6\" (UID: \"511f2b6d-a08d-49f8-b393-ab222219d4a7\") " pod="openstack/ovn-controller-htwd6" Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.062232 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/b27b3f77-8476-4f3e-a946-f4f26fc66cb5-var-run\") pod \"ovn-controller-ovs-gr9gj\" (UID: \"b27b3f77-8476-4f3e-a946-f4f26fc66cb5\") " pod="openstack/ovn-controller-ovs-gr9gj" Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.063050 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6cdxj\" (UniqueName: \"kubernetes.io/projected/b27b3f77-8476-4f3e-a946-f4f26fc66cb5-kube-api-access-6cdxj\") pod \"ovn-controller-ovs-gr9gj\" (UID: \"b27b3f77-8476-4f3e-a946-f4f26fc66cb5\") " pod="openstack/ovn-controller-ovs-gr9gj" Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.063168 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e8aca65b-76cd-45f9-8b20-da2e34a41fc2-kube-api-access-jt49s" (OuterVolumeSpecName: "kube-api-access-jt49s") pod "e8aca65b-76cd-45f9-8b20-da2e34a41fc2" (UID: "e8aca65b-76cd-45f9-8b20-da2e34a41fc2"). InnerVolumeSpecName "kube-api-access-jt49s". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.063208 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/511f2b6d-a08d-49f8-b393-ab222219d4a7-var-run-ovn\") pod \"ovn-controller-htwd6\" (UID: \"511f2b6d-a08d-49f8-b393-ab222219d4a7\") " pod="openstack/ovn-controller-htwd6" Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.063349 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b48ab67f-88cb-4921-b6a3-38b77b534bf9-kube-api-access-8q4rz" (OuterVolumeSpecName: "kube-api-access-8q4rz") pod "b48ab67f-88cb-4921-b6a3-38b77b534bf9" (UID: "b48ab67f-88cb-4921-b6a3-38b77b534bf9"). InnerVolumeSpecName "kube-api-access-8q4rz". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.067113 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b27b3f77-8476-4f3e-a946-f4f26fc66cb5-scripts\") pod \"ovn-controller-ovs-gr9gj\" (UID: \"b27b3f77-8476-4f3e-a946-f4f26fc66cb5\") " pod="openstack/ovn-controller-ovs-gr9gj" Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.067269 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/511f2b6d-a08d-49f8-b393-ab222219d4a7-combined-ca-bundle\") pod \"ovn-controller-htwd6\" (UID: \"511f2b6d-a08d-49f8-b393-ab222219d4a7\") " pod="openstack/ovn-controller-htwd6" Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.067402 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/511f2b6d-a08d-49f8-b393-ab222219d4a7-scripts\") pod \"ovn-controller-htwd6\" (UID: \"511f2b6d-a08d-49f8-b393-ab222219d4a7\") " pod="openstack/ovn-controller-htwd6" Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.074740 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e8aca65b-76cd-45f9-8b20-da2e34a41fc2-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "e8aca65b-76cd-45f9-8b20-da2e34a41fc2" (UID: "e8aca65b-76cd-45f9-8b20-da2e34a41fc2"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.074913 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/511f2b6d-a08d-49f8-b393-ab222219d4a7-ovn-controller-tls-certs\") pod \"ovn-controller-htwd6\" (UID: \"511f2b6d-a08d-49f8-b393-ab222219d4a7\") " pod="openstack/ovn-controller-htwd6" Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.075586 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e8aca65b-76cd-45f9-8b20-da2e34a41fc2-config" (OuterVolumeSpecName: "config") pod "e8aca65b-76cd-45f9-8b20-da2e34a41fc2" (UID: "e8aca65b-76cd-45f9-8b20-da2e34a41fc2"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.076992 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cz8l5\" (UniqueName: \"kubernetes.io/projected/511f2b6d-a08d-49f8-b393-ab222219d4a7-kube-api-access-cz8l5\") pod \"ovn-controller-htwd6\" (UID: \"511f2b6d-a08d-49f8-b393-ab222219d4a7\") " pod="openstack/ovn-controller-htwd6" Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.077558 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6cdxj\" (UniqueName: \"kubernetes.io/projected/b27b3f77-8476-4f3e-a946-f4f26fc66cb5-kube-api-access-6cdxj\") pod \"ovn-controller-ovs-gr9gj\" (UID: \"b27b3f77-8476-4f3e-a946-f4f26fc66cb5\") " pod="openstack/ovn-controller-ovs-gr9gj" Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.083243 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b48ab67f-88cb-4921-b6a3-38b77b534bf9-config" (OuterVolumeSpecName: "config") pod "b48ab67f-88cb-4921-b6a3-38b77b534bf9" (UID: "b48ab67f-88cb-4921-b6a3-38b77b534bf9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.106911 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-htwd6" Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.165464 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8q4rz\" (UniqueName: \"kubernetes.io/projected/b48ab67f-88cb-4921-b6a3-38b77b534bf9-kube-api-access-8q4rz\") on node \"crc\" DevicePath \"\"" Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.165490 4824 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e8aca65b-76cd-45f9-8b20-da2e34a41fc2-dns-svc\") on node \"crc\" DevicePath \"\"" Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.165499 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e8aca65b-76cd-45f9-8b20-da2e34a41fc2-config\") on node \"crc\" DevicePath \"\"" Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.165507 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b48ab67f-88cb-4921-b6a3-38b77b534bf9-config\") on node \"crc\" DevicePath \"\"" Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.165515 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jt49s\" (UniqueName: \"kubernetes.io/projected/e8aca65b-76cd-45f9-8b20-da2e34a41fc2-kube-api-access-jt49s\") on node \"crc\" DevicePath \"\"" Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.211781 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-gr9gj" Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.319021 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Jan 21 11:25:20 crc kubenswrapper[4824]: E0121 11:25:20.319284 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b48ab67f-88cb-4921-b6a3-38b77b534bf9" containerName="init" Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.319298 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="b48ab67f-88cb-4921-b6a3-38b77b534bf9" containerName="init" Jan 21 11:25:20 crc kubenswrapper[4824]: E0121 11:25:20.319316 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8aca65b-76cd-45f9-8b20-da2e34a41fc2" containerName="init" Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.319321 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8aca65b-76cd-45f9-8b20-da2e34a41fc2" containerName="init" Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.319452 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="e8aca65b-76cd-45f9-8b20-da2e34a41fc2" containerName="init" Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.319468 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="b48ab67f-88cb-4921-b6a3-38b77b534bf9" containerName="init" Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.320122 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.322531 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.322826 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.326133 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.326259 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.326520 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-l8pwl" Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.331371 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.367602 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5c0b5876-da04-49e6-8946-4fc6060a64ec-config\") pod \"ovsdbserver-nb-0\" (UID: \"5c0b5876-da04-49e6-8946-4fc6060a64ec\") " pod="openstack/ovsdbserver-nb-0" Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.367636 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/5c0b5876-da04-49e6-8946-4fc6060a64ec-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"5c0b5876-da04-49e6-8946-4fc6060a64ec\") " pod="openstack/ovsdbserver-nb-0" Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.367678 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5c0b5876-da04-49e6-8946-4fc6060a64ec-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"5c0b5876-da04-49e6-8946-4fc6060a64ec\") " pod="openstack/ovsdbserver-nb-0" Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.367789 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/5c0b5876-da04-49e6-8946-4fc6060a64ec-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"5c0b5876-da04-49e6-8946-4fc6060a64ec\") " pod="openstack/ovsdbserver-nb-0" Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.367860 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/5c0b5876-da04-49e6-8946-4fc6060a64ec-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"5c0b5876-da04-49e6-8946-4fc6060a64ec\") " pod="openstack/ovsdbserver-nb-0" Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.367884 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x2nbk\" (UniqueName: \"kubernetes.io/projected/5c0b5876-da04-49e6-8946-4fc6060a64ec-kube-api-access-x2nbk\") pod \"ovsdbserver-nb-0\" (UID: \"5c0b5876-da04-49e6-8946-4fc6060a64ec\") " pod="openstack/ovsdbserver-nb-0" Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.367986 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"ovsdbserver-nb-0\" (UID: \"5c0b5876-da04-49e6-8946-4fc6060a64ec\") " pod="openstack/ovsdbserver-nb-0" Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.368067 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c0b5876-da04-49e6-8946-4fc6060a64ec-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"5c0b5876-da04-49e6-8946-4fc6060a64ec\") " pod="openstack/ovsdbserver-nb-0" Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.471027 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c0b5876-da04-49e6-8946-4fc6060a64ec-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"5c0b5876-da04-49e6-8946-4fc6060a64ec\") " pod="openstack/ovsdbserver-nb-0" Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.471096 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5c0b5876-da04-49e6-8946-4fc6060a64ec-config\") pod \"ovsdbserver-nb-0\" (UID: \"5c0b5876-da04-49e6-8946-4fc6060a64ec\") " pod="openstack/ovsdbserver-nb-0" Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.471123 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/5c0b5876-da04-49e6-8946-4fc6060a64ec-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"5c0b5876-da04-49e6-8946-4fc6060a64ec\") " pod="openstack/ovsdbserver-nb-0" Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.471174 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5c0b5876-da04-49e6-8946-4fc6060a64ec-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"5c0b5876-da04-49e6-8946-4fc6060a64ec\") " pod="openstack/ovsdbserver-nb-0" Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.471208 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/5c0b5876-da04-49e6-8946-4fc6060a64ec-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"5c0b5876-da04-49e6-8946-4fc6060a64ec\") " pod="openstack/ovsdbserver-nb-0" Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.471258 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/5c0b5876-da04-49e6-8946-4fc6060a64ec-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"5c0b5876-da04-49e6-8946-4fc6060a64ec\") " pod="openstack/ovsdbserver-nb-0" Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.471280 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x2nbk\" (UniqueName: \"kubernetes.io/projected/5c0b5876-da04-49e6-8946-4fc6060a64ec-kube-api-access-x2nbk\") pod \"ovsdbserver-nb-0\" (UID: \"5c0b5876-da04-49e6-8946-4fc6060a64ec\") " pod="openstack/ovsdbserver-nb-0" Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.471327 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"ovsdbserver-nb-0\" (UID: \"5c0b5876-da04-49e6-8946-4fc6060a64ec\") " pod="openstack/ovsdbserver-nb-0" Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.471599 4824 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"ovsdbserver-nb-0\" (UID: \"5c0b5876-da04-49e6-8946-4fc6060a64ec\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/ovsdbserver-nb-0" Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.471867 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/5c0b5876-da04-49e6-8946-4fc6060a64ec-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"5c0b5876-da04-49e6-8946-4fc6060a64ec\") " pod="openstack/ovsdbserver-nb-0" Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.471982 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5c0b5876-da04-49e6-8946-4fc6060a64ec-config\") pod \"ovsdbserver-nb-0\" (UID: \"5c0b5876-da04-49e6-8946-4fc6060a64ec\") " pod="openstack/ovsdbserver-nb-0" Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.472242 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5c0b5876-da04-49e6-8946-4fc6060a64ec-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"5c0b5876-da04-49e6-8946-4fc6060a64ec\") " pod="openstack/ovsdbserver-nb-0" Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.474547 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/5c0b5876-da04-49e6-8946-4fc6060a64ec-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"5c0b5876-da04-49e6-8946-4fc6060a64ec\") " pod="openstack/ovsdbserver-nb-0" Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.474722 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/5c0b5876-da04-49e6-8946-4fc6060a64ec-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"5c0b5876-da04-49e6-8946-4fc6060a64ec\") " pod="openstack/ovsdbserver-nb-0" Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.478521 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c0b5876-da04-49e6-8946-4fc6060a64ec-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"5c0b5876-da04-49e6-8946-4fc6060a64ec\") " pod="openstack/ovsdbserver-nb-0" Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.483830 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-nrqwx" event={"ID":"71b408e9-2ba6-4b0e-9b4c-22faf4ef4fd4","Type":"ContainerStarted","Data":"bdaacbc24f5cf64fe4e60d7403f3ba4bc3f083322e15a6cc88509aa65a02a4b8"} Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.483884 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-57d769cc4f-nrqwx" Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.485157 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x2nbk\" (UniqueName: \"kubernetes.io/projected/5c0b5876-da04-49e6-8946-4fc6060a64ec-kube-api-access-x2nbk\") pod \"ovsdbserver-nb-0\" (UID: \"5c0b5876-da04-49e6-8946-4fc6060a64ec\") " pod="openstack/ovsdbserver-nb-0" Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.489639 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-zspkp" event={"ID":"f67dfc52-dc0e-404e-af22-75c6a3d10721","Type":"ContainerStarted","Data":"0ad321acd47c8d7e7d82f491f6ee8db60e60e3e1cab256cda20de9be81202aa0"} Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.489864 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-666b6646f7-zspkp" Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.491122 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"ovsdbserver-nb-0\" (UID: \"5c0b5876-da04-49e6-8946-4fc6060a64ec\") " pod="openstack/ovsdbserver-nb-0" Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.491564 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-lmtkp" event={"ID":"b48ab67f-88cb-4921-b6a3-38b77b534bf9","Type":"ContainerDied","Data":"c935cf67bf3bb60e71d9e8fa32ea33c10f87bb841cf838b50e9a195acf0ce5f3"} Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.491600 4824 scope.go:117] "RemoveContainer" containerID="415de3e8643b808347c13b2c7d4d57542a514312e1a5bd379cbc7b434241d117" Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.491775 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-lmtkp" Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.500893 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-tmr2j" event={"ID":"e8aca65b-76cd-45f9-8b20-da2e34a41fc2","Type":"ContainerDied","Data":"8c626b420278544fc13bdb44a0fc27116c48e8b3b2e62a1f886c8aa8d6e59139"} Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.500969 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-tmr2j" Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.510400 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-57d769cc4f-nrqwx" podStartSLOduration=10.510386 podStartE2EDuration="10.510386s" podCreationTimestamp="2026-01-21 11:25:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:25:20.498636132 +0000 UTC m=+862.791665423" watchObservedRunningTime="2026-01-21 11:25:20.510386 +0000 UTC m=+862.803415291" Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.522703 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-666b6646f7-zspkp" podStartSLOduration=11.522682457 podStartE2EDuration="11.522682457s" podCreationTimestamp="2026-01-21 11:25:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:25:20.517763245 +0000 UTC m=+862.810792537" watchObservedRunningTime="2026-01-21 11:25:20.522682457 +0000 UTC m=+862.815711749" Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.545770 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-lmtkp"] Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.561132 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-lmtkp"] Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.607985 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-tmr2j"] Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.613792 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-tmr2j"] Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.647199 4824 scope.go:117] "RemoveContainer" containerID="aaacc227843f224a20462553c49fdbfc00aeab8ce2765ad6ee4c575c381c5973" Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.647597 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Jan 21 11:25:20 crc kubenswrapper[4824]: I0121 11:25:20.966787 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-htwd6"] Jan 21 11:25:20 crc kubenswrapper[4824]: W0121 11:25:20.976582 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod511f2b6d_a08d_49f8_b393_ab222219d4a7.slice/crio-6ed1172c8819aecfe47f2a831be778c84a6bcec641f7cd2728eaf87d11c3cd23 WatchSource:0}: Error finding container 6ed1172c8819aecfe47f2a831be778c84a6bcec641f7cd2728eaf87d11c3cd23: Status 404 returned error can't find the container with id 6ed1172c8819aecfe47f2a831be778c84a6bcec641f7cd2728eaf87d11c3cd23 Jan 21 11:25:21 crc kubenswrapper[4824]: I0121 11:25:21.128600 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-gr9gj"] Jan 21 11:25:21 crc kubenswrapper[4824]: I0121 11:25:21.245158 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-vrpmw"] Jan 21 11:25:21 crc kubenswrapper[4824]: I0121 11:25:21.246373 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-vrpmw" Jan 21 11:25:21 crc kubenswrapper[4824]: I0121 11:25:21.248176 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-vrpmw"] Jan 21 11:25:21 crc kubenswrapper[4824]: I0121 11:25:21.254731 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Jan 21 11:25:21 crc kubenswrapper[4824]: I0121 11:25:21.282036 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/8652f3f6-1ad7-49d9-9048-827be17a3dd0-ovs-rundir\") pod \"ovn-controller-metrics-vrpmw\" (UID: \"8652f3f6-1ad7-49d9-9048-827be17a3dd0\") " pod="openstack/ovn-controller-metrics-vrpmw" Jan 21 11:25:21 crc kubenswrapper[4824]: I0121 11:25:21.282078 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8652f3f6-1ad7-49d9-9048-827be17a3dd0-combined-ca-bundle\") pod \"ovn-controller-metrics-vrpmw\" (UID: \"8652f3f6-1ad7-49d9-9048-827be17a3dd0\") " pod="openstack/ovn-controller-metrics-vrpmw" Jan 21 11:25:21 crc kubenswrapper[4824]: I0121 11:25:21.282102 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/8652f3f6-1ad7-49d9-9048-827be17a3dd0-ovn-rundir\") pod \"ovn-controller-metrics-vrpmw\" (UID: \"8652f3f6-1ad7-49d9-9048-827be17a3dd0\") " pod="openstack/ovn-controller-metrics-vrpmw" Jan 21 11:25:21 crc kubenswrapper[4824]: I0121 11:25:21.282134 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8652f3f6-1ad7-49d9-9048-827be17a3dd0-config\") pod \"ovn-controller-metrics-vrpmw\" (UID: \"8652f3f6-1ad7-49d9-9048-827be17a3dd0\") " pod="openstack/ovn-controller-metrics-vrpmw" Jan 21 11:25:21 crc kubenswrapper[4824]: I0121 11:25:21.282182 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sqtdg\" (UniqueName: \"kubernetes.io/projected/8652f3f6-1ad7-49d9-9048-827be17a3dd0-kube-api-access-sqtdg\") pod \"ovn-controller-metrics-vrpmw\" (UID: \"8652f3f6-1ad7-49d9-9048-827be17a3dd0\") " pod="openstack/ovn-controller-metrics-vrpmw" Jan 21 11:25:21 crc kubenswrapper[4824]: I0121 11:25:21.282202 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/8652f3f6-1ad7-49d9-9048-827be17a3dd0-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-vrpmw\" (UID: \"8652f3f6-1ad7-49d9-9048-827be17a3dd0\") " pod="openstack/ovn-controller-metrics-vrpmw" Jan 21 11:25:21 crc kubenswrapper[4824]: I0121 11:25:21.330983 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Jan 21 11:25:21 crc kubenswrapper[4824]: I0121 11:25:21.383560 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sqtdg\" (UniqueName: \"kubernetes.io/projected/8652f3f6-1ad7-49d9-9048-827be17a3dd0-kube-api-access-sqtdg\") pod \"ovn-controller-metrics-vrpmw\" (UID: \"8652f3f6-1ad7-49d9-9048-827be17a3dd0\") " pod="openstack/ovn-controller-metrics-vrpmw" Jan 21 11:25:21 crc kubenswrapper[4824]: I0121 11:25:21.383605 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/8652f3f6-1ad7-49d9-9048-827be17a3dd0-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-vrpmw\" (UID: \"8652f3f6-1ad7-49d9-9048-827be17a3dd0\") " pod="openstack/ovn-controller-metrics-vrpmw" Jan 21 11:25:21 crc kubenswrapper[4824]: I0121 11:25:21.383662 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/8652f3f6-1ad7-49d9-9048-827be17a3dd0-ovs-rundir\") pod \"ovn-controller-metrics-vrpmw\" (UID: \"8652f3f6-1ad7-49d9-9048-827be17a3dd0\") " pod="openstack/ovn-controller-metrics-vrpmw" Jan 21 11:25:21 crc kubenswrapper[4824]: I0121 11:25:21.383701 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8652f3f6-1ad7-49d9-9048-827be17a3dd0-combined-ca-bundle\") pod \"ovn-controller-metrics-vrpmw\" (UID: \"8652f3f6-1ad7-49d9-9048-827be17a3dd0\") " pod="openstack/ovn-controller-metrics-vrpmw" Jan 21 11:25:21 crc kubenswrapper[4824]: I0121 11:25:21.383720 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/8652f3f6-1ad7-49d9-9048-827be17a3dd0-ovn-rundir\") pod \"ovn-controller-metrics-vrpmw\" (UID: \"8652f3f6-1ad7-49d9-9048-827be17a3dd0\") " pod="openstack/ovn-controller-metrics-vrpmw" Jan 21 11:25:21 crc kubenswrapper[4824]: I0121 11:25:21.383748 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8652f3f6-1ad7-49d9-9048-827be17a3dd0-config\") pod \"ovn-controller-metrics-vrpmw\" (UID: \"8652f3f6-1ad7-49d9-9048-827be17a3dd0\") " pod="openstack/ovn-controller-metrics-vrpmw" Jan 21 11:25:21 crc kubenswrapper[4824]: I0121 11:25:21.384182 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/8652f3f6-1ad7-49d9-9048-827be17a3dd0-ovn-rundir\") pod \"ovn-controller-metrics-vrpmw\" (UID: \"8652f3f6-1ad7-49d9-9048-827be17a3dd0\") " pod="openstack/ovn-controller-metrics-vrpmw" Jan 21 11:25:21 crc kubenswrapper[4824]: I0121 11:25:21.384392 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/8652f3f6-1ad7-49d9-9048-827be17a3dd0-ovs-rundir\") pod \"ovn-controller-metrics-vrpmw\" (UID: \"8652f3f6-1ad7-49d9-9048-827be17a3dd0\") " pod="openstack/ovn-controller-metrics-vrpmw" Jan 21 11:25:21 crc kubenswrapper[4824]: I0121 11:25:21.384852 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8652f3f6-1ad7-49d9-9048-827be17a3dd0-config\") pod \"ovn-controller-metrics-vrpmw\" (UID: \"8652f3f6-1ad7-49d9-9048-827be17a3dd0\") " pod="openstack/ovn-controller-metrics-vrpmw" Jan 21 11:25:21 crc kubenswrapper[4824]: I0121 11:25:21.388591 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8652f3f6-1ad7-49d9-9048-827be17a3dd0-combined-ca-bundle\") pod \"ovn-controller-metrics-vrpmw\" (UID: \"8652f3f6-1ad7-49d9-9048-827be17a3dd0\") " pod="openstack/ovn-controller-metrics-vrpmw" Jan 21 11:25:21 crc kubenswrapper[4824]: I0121 11:25:21.388610 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/8652f3f6-1ad7-49d9-9048-827be17a3dd0-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-vrpmw\" (UID: \"8652f3f6-1ad7-49d9-9048-827be17a3dd0\") " pod="openstack/ovn-controller-metrics-vrpmw" Jan 21 11:25:21 crc kubenswrapper[4824]: I0121 11:25:21.397918 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sqtdg\" (UniqueName: \"kubernetes.io/projected/8652f3f6-1ad7-49d9-9048-827be17a3dd0-kube-api-access-sqtdg\") pod \"ovn-controller-metrics-vrpmw\" (UID: \"8652f3f6-1ad7-49d9-9048-827be17a3dd0\") " pod="openstack/ovn-controller-metrics-vrpmw" Jan 21 11:25:21 crc kubenswrapper[4824]: I0121 11:25:21.510482 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-htwd6" event={"ID":"511f2b6d-a08d-49f8-b393-ab222219d4a7","Type":"ContainerStarted","Data":"6ed1172c8819aecfe47f2a831be778c84a6bcec641f7cd2728eaf87d11c3cd23"} Jan 21 11:25:21 crc kubenswrapper[4824]: W0121 11:25:21.549042 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb27b3f77_8476_4f3e_a946_f4f26fc66cb5.slice/crio-b01683b60bc1b3ef97826d5ab6262af5f63c4d702d8d2950e38ddae2e977f4b2 WatchSource:0}: Error finding container b01683b60bc1b3ef97826d5ab6262af5f63c4d702d8d2950e38ddae2e977f4b2: Status 404 returned error can't find the container with id b01683b60bc1b3ef97826d5ab6262af5f63c4d702d8d2950e38ddae2e977f4b2 Jan 21 11:25:21 crc kubenswrapper[4824]: I0121 11:25:21.572572 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-vrpmw" Jan 21 11:25:22 crc kubenswrapper[4824]: I0121 11:25:22.065338 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b48ab67f-88cb-4921-b6a3-38b77b534bf9" path="/var/lib/kubelet/pods/b48ab67f-88cb-4921-b6a3-38b77b534bf9/volumes" Jan 21 11:25:22 crc kubenswrapper[4824]: I0121 11:25:22.066621 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e8aca65b-76cd-45f9-8b20-da2e34a41fc2" path="/var/lib/kubelet/pods/e8aca65b-76cd-45f9-8b20-da2e34a41fc2/volumes" Jan 21 11:25:22 crc kubenswrapper[4824]: I0121 11:25:22.334039 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-zspkp"] Jan 21 11:25:22 crc kubenswrapper[4824]: I0121 11:25:22.351143 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-4xld6"] Jan 21 11:25:22 crc kubenswrapper[4824]: I0121 11:25:22.352148 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-4xld6" Jan 21 11:25:22 crc kubenswrapper[4824]: I0121 11:25:22.360381 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Jan 21 11:25:22 crc kubenswrapper[4824]: I0121 11:25:22.371383 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-4xld6"] Jan 21 11:25:22 crc kubenswrapper[4824]: I0121 11:25:22.401539 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/56a2624d-7ab8-43ea-967e-e4112c693148-dns-svc\") pod \"dnsmasq-dns-7fd796d7df-4xld6\" (UID: \"56a2624d-7ab8-43ea-967e-e4112c693148\") " pod="openstack/dnsmasq-dns-7fd796d7df-4xld6" Jan 21 11:25:22 crc kubenswrapper[4824]: I0121 11:25:22.401574 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/56a2624d-7ab8-43ea-967e-e4112c693148-config\") pod \"dnsmasq-dns-7fd796d7df-4xld6\" (UID: \"56a2624d-7ab8-43ea-967e-e4112c693148\") " pod="openstack/dnsmasq-dns-7fd796d7df-4xld6" Jan 21 11:25:22 crc kubenswrapper[4824]: I0121 11:25:22.401862 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/56a2624d-7ab8-43ea-967e-e4112c693148-ovsdbserver-nb\") pod \"dnsmasq-dns-7fd796d7df-4xld6\" (UID: \"56a2624d-7ab8-43ea-967e-e4112c693148\") " pod="openstack/dnsmasq-dns-7fd796d7df-4xld6" Jan 21 11:25:22 crc kubenswrapper[4824]: I0121 11:25:22.401969 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rlmf6\" (UniqueName: \"kubernetes.io/projected/56a2624d-7ab8-43ea-967e-e4112c693148-kube-api-access-rlmf6\") pod \"dnsmasq-dns-7fd796d7df-4xld6\" (UID: \"56a2624d-7ab8-43ea-967e-e4112c693148\") " pod="openstack/dnsmasq-dns-7fd796d7df-4xld6" Jan 21 11:25:22 crc kubenswrapper[4824]: I0121 11:25:22.504280 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/56a2624d-7ab8-43ea-967e-e4112c693148-dns-svc\") pod \"dnsmasq-dns-7fd796d7df-4xld6\" (UID: \"56a2624d-7ab8-43ea-967e-e4112c693148\") " pod="openstack/dnsmasq-dns-7fd796d7df-4xld6" Jan 21 11:25:22 crc kubenswrapper[4824]: I0121 11:25:22.504339 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/56a2624d-7ab8-43ea-967e-e4112c693148-dns-svc\") pod \"dnsmasq-dns-7fd796d7df-4xld6\" (UID: \"56a2624d-7ab8-43ea-967e-e4112c693148\") " pod="openstack/dnsmasq-dns-7fd796d7df-4xld6" Jan 21 11:25:22 crc kubenswrapper[4824]: I0121 11:25:22.504374 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/56a2624d-7ab8-43ea-967e-e4112c693148-config\") pod \"dnsmasq-dns-7fd796d7df-4xld6\" (UID: \"56a2624d-7ab8-43ea-967e-e4112c693148\") " pod="openstack/dnsmasq-dns-7fd796d7df-4xld6" Jan 21 11:25:22 crc kubenswrapper[4824]: I0121 11:25:22.504968 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/56a2624d-7ab8-43ea-967e-e4112c693148-config\") pod \"dnsmasq-dns-7fd796d7df-4xld6\" (UID: \"56a2624d-7ab8-43ea-967e-e4112c693148\") " pod="openstack/dnsmasq-dns-7fd796d7df-4xld6" Jan 21 11:25:22 crc kubenswrapper[4824]: I0121 11:25:22.505130 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/56a2624d-7ab8-43ea-967e-e4112c693148-ovsdbserver-nb\") pod \"dnsmasq-dns-7fd796d7df-4xld6\" (UID: \"56a2624d-7ab8-43ea-967e-e4112c693148\") " pod="openstack/dnsmasq-dns-7fd796d7df-4xld6" Jan 21 11:25:22 crc kubenswrapper[4824]: I0121 11:25:22.505663 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/56a2624d-7ab8-43ea-967e-e4112c693148-ovsdbserver-nb\") pod \"dnsmasq-dns-7fd796d7df-4xld6\" (UID: \"56a2624d-7ab8-43ea-967e-e4112c693148\") " pod="openstack/dnsmasq-dns-7fd796d7df-4xld6" Jan 21 11:25:22 crc kubenswrapper[4824]: I0121 11:25:22.505790 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rlmf6\" (UniqueName: \"kubernetes.io/projected/56a2624d-7ab8-43ea-967e-e4112c693148-kube-api-access-rlmf6\") pod \"dnsmasq-dns-7fd796d7df-4xld6\" (UID: \"56a2624d-7ab8-43ea-967e-e4112c693148\") " pod="openstack/dnsmasq-dns-7fd796d7df-4xld6" Jan 21 11:25:22 crc kubenswrapper[4824]: I0121 11:25:22.518571 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-gr9gj" event={"ID":"b27b3f77-8476-4f3e-a946-f4f26fc66cb5","Type":"ContainerStarted","Data":"b01683b60bc1b3ef97826d5ab6262af5f63c4d702d8d2950e38ddae2e977f4b2"} Jan 21 11:25:22 crc kubenswrapper[4824]: I0121 11:25:22.518721 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-666b6646f7-zspkp" podUID="f67dfc52-dc0e-404e-af22-75c6a3d10721" containerName="dnsmasq-dns" containerID="cri-o://0ad321acd47c8d7e7d82f491f6ee8db60e60e3e1cab256cda20de9be81202aa0" gracePeriod=10 Jan 21 11:25:22 crc kubenswrapper[4824]: I0121 11:25:22.518978 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rlmf6\" (UniqueName: \"kubernetes.io/projected/56a2624d-7ab8-43ea-967e-e4112c693148-kube-api-access-rlmf6\") pod \"dnsmasq-dns-7fd796d7df-4xld6\" (UID: \"56a2624d-7ab8-43ea-967e-e4112c693148\") " pod="openstack/dnsmasq-dns-7fd796d7df-4xld6" Jan 21 11:25:22 crc kubenswrapper[4824]: I0121 11:25:22.670352 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-4xld6" Jan 21 11:25:22 crc kubenswrapper[4824]: W0121 11:25:22.721794 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5c0b5876_da04_49e6_8946_4fc6060a64ec.slice/crio-a9120e7f79937c25195d3b830feb6ca41ab891b48425d59e1a0ca2c0018a23cd WatchSource:0}: Error finding container a9120e7f79937c25195d3b830feb6ca41ab891b48425d59e1a0ca2c0018a23cd: Status 404 returned error can't find the container with id a9120e7f79937c25195d3b830feb6ca41ab891b48425d59e1a0ca2c0018a23cd Jan 21 11:25:22 crc kubenswrapper[4824]: I0121 11:25:22.842734 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Jan 21 11:25:22 crc kubenswrapper[4824]: I0121 11:25:22.844084 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Jan 21 11:25:22 crc kubenswrapper[4824]: I0121 11:25:22.845724 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Jan 21 11:25:22 crc kubenswrapper[4824]: I0121 11:25:22.845932 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Jan 21 11:25:22 crc kubenswrapper[4824]: I0121 11:25:22.846177 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-rn62s" Jan 21 11:25:22 crc kubenswrapper[4824]: I0121 11:25:22.847204 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Jan 21 11:25:22 crc kubenswrapper[4824]: I0121 11:25:22.855280 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Jan 21 11:25:22 crc kubenswrapper[4824]: I0121 11:25:22.911558 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e86f23d4-099d-4e8b-b3a4-d2fc7152c1ef-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"e86f23d4-099d-4e8b-b3a4-d2fc7152c1ef\") " pod="openstack/ovsdbserver-sb-0" Jan 21 11:25:22 crc kubenswrapper[4824]: I0121 11:25:22.911636 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"ovsdbserver-sb-0\" (UID: \"e86f23d4-099d-4e8b-b3a4-d2fc7152c1ef\") " pod="openstack/ovsdbserver-sb-0" Jan 21 11:25:22 crc kubenswrapper[4824]: I0121 11:25:22.911658 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e86f23d4-099d-4e8b-b3a4-d2fc7152c1ef-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"e86f23d4-099d-4e8b-b3a4-d2fc7152c1ef\") " pod="openstack/ovsdbserver-sb-0" Jan 21 11:25:22 crc kubenswrapper[4824]: I0121 11:25:22.911730 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e86f23d4-099d-4e8b-b3a4-d2fc7152c1ef-config\") pod \"ovsdbserver-sb-0\" (UID: \"e86f23d4-099d-4e8b-b3a4-d2fc7152c1ef\") " pod="openstack/ovsdbserver-sb-0" Jan 21 11:25:22 crc kubenswrapper[4824]: I0121 11:25:22.911788 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/e86f23d4-099d-4e8b-b3a4-d2fc7152c1ef-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"e86f23d4-099d-4e8b-b3a4-d2fc7152c1ef\") " pod="openstack/ovsdbserver-sb-0" Jan 21 11:25:22 crc kubenswrapper[4824]: I0121 11:25:22.911823 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/e86f23d4-099d-4e8b-b3a4-d2fc7152c1ef-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"e86f23d4-099d-4e8b-b3a4-d2fc7152c1ef\") " pod="openstack/ovsdbserver-sb-0" Jan 21 11:25:22 crc kubenswrapper[4824]: I0121 11:25:22.911869 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/e86f23d4-099d-4e8b-b3a4-d2fc7152c1ef-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"e86f23d4-099d-4e8b-b3a4-d2fc7152c1ef\") " pod="openstack/ovsdbserver-sb-0" Jan 21 11:25:22 crc kubenswrapper[4824]: I0121 11:25:22.911906 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lxnpp\" (UniqueName: \"kubernetes.io/projected/e86f23d4-099d-4e8b-b3a4-d2fc7152c1ef-kube-api-access-lxnpp\") pod \"ovsdbserver-sb-0\" (UID: \"e86f23d4-099d-4e8b-b3a4-d2fc7152c1ef\") " pod="openstack/ovsdbserver-sb-0" Jan 21 11:25:23 crc kubenswrapper[4824]: I0121 11:25:23.014610 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e86f23d4-099d-4e8b-b3a4-d2fc7152c1ef-config\") pod \"ovsdbserver-sb-0\" (UID: \"e86f23d4-099d-4e8b-b3a4-d2fc7152c1ef\") " pod="openstack/ovsdbserver-sb-0" Jan 21 11:25:23 crc kubenswrapper[4824]: I0121 11:25:23.014670 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/e86f23d4-099d-4e8b-b3a4-d2fc7152c1ef-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"e86f23d4-099d-4e8b-b3a4-d2fc7152c1ef\") " pod="openstack/ovsdbserver-sb-0" Jan 21 11:25:23 crc kubenswrapper[4824]: I0121 11:25:23.014735 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/e86f23d4-099d-4e8b-b3a4-d2fc7152c1ef-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"e86f23d4-099d-4e8b-b3a4-d2fc7152c1ef\") " pod="openstack/ovsdbserver-sb-0" Jan 21 11:25:23 crc kubenswrapper[4824]: I0121 11:25:23.014765 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/e86f23d4-099d-4e8b-b3a4-d2fc7152c1ef-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"e86f23d4-099d-4e8b-b3a4-d2fc7152c1ef\") " pod="openstack/ovsdbserver-sb-0" Jan 21 11:25:23 crc kubenswrapper[4824]: I0121 11:25:23.014816 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lxnpp\" (UniqueName: \"kubernetes.io/projected/e86f23d4-099d-4e8b-b3a4-d2fc7152c1ef-kube-api-access-lxnpp\") pod \"ovsdbserver-sb-0\" (UID: \"e86f23d4-099d-4e8b-b3a4-d2fc7152c1ef\") " pod="openstack/ovsdbserver-sb-0" Jan 21 11:25:23 crc kubenswrapper[4824]: I0121 11:25:23.014919 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e86f23d4-099d-4e8b-b3a4-d2fc7152c1ef-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"e86f23d4-099d-4e8b-b3a4-d2fc7152c1ef\") " pod="openstack/ovsdbserver-sb-0" Jan 21 11:25:23 crc kubenswrapper[4824]: I0121 11:25:23.014978 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"ovsdbserver-sb-0\" (UID: \"e86f23d4-099d-4e8b-b3a4-d2fc7152c1ef\") " pod="openstack/ovsdbserver-sb-0" Jan 21 11:25:23 crc kubenswrapper[4824]: I0121 11:25:23.014992 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e86f23d4-099d-4e8b-b3a4-d2fc7152c1ef-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"e86f23d4-099d-4e8b-b3a4-d2fc7152c1ef\") " pod="openstack/ovsdbserver-sb-0" Jan 21 11:25:23 crc kubenswrapper[4824]: I0121 11:25:23.015588 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/e86f23d4-099d-4e8b-b3a4-d2fc7152c1ef-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"e86f23d4-099d-4e8b-b3a4-d2fc7152c1ef\") " pod="openstack/ovsdbserver-sb-0" Jan 21 11:25:23 crc kubenswrapper[4824]: I0121 11:25:23.015663 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e86f23d4-099d-4e8b-b3a4-d2fc7152c1ef-config\") pod \"ovsdbserver-sb-0\" (UID: \"e86f23d4-099d-4e8b-b3a4-d2fc7152c1ef\") " pod="openstack/ovsdbserver-sb-0" Jan 21 11:25:23 crc kubenswrapper[4824]: I0121 11:25:23.015886 4824 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"ovsdbserver-sb-0\" (UID: \"e86f23d4-099d-4e8b-b3a4-d2fc7152c1ef\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/ovsdbserver-sb-0" Jan 21 11:25:23 crc kubenswrapper[4824]: I0121 11:25:23.017309 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e86f23d4-099d-4e8b-b3a4-d2fc7152c1ef-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"e86f23d4-099d-4e8b-b3a4-d2fc7152c1ef\") " pod="openstack/ovsdbserver-sb-0" Jan 21 11:25:23 crc kubenswrapper[4824]: I0121 11:25:23.019371 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e86f23d4-099d-4e8b-b3a4-d2fc7152c1ef-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"e86f23d4-099d-4e8b-b3a4-d2fc7152c1ef\") " pod="openstack/ovsdbserver-sb-0" Jan 21 11:25:23 crc kubenswrapper[4824]: I0121 11:25:23.019609 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/e86f23d4-099d-4e8b-b3a4-d2fc7152c1ef-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"e86f23d4-099d-4e8b-b3a4-d2fc7152c1ef\") " pod="openstack/ovsdbserver-sb-0" Jan 21 11:25:23 crc kubenswrapper[4824]: I0121 11:25:23.023057 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/e86f23d4-099d-4e8b-b3a4-d2fc7152c1ef-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"e86f23d4-099d-4e8b-b3a4-d2fc7152c1ef\") " pod="openstack/ovsdbserver-sb-0" Jan 21 11:25:23 crc kubenswrapper[4824]: I0121 11:25:23.029678 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lxnpp\" (UniqueName: \"kubernetes.io/projected/e86f23d4-099d-4e8b-b3a4-d2fc7152c1ef-kube-api-access-lxnpp\") pod \"ovsdbserver-sb-0\" (UID: \"e86f23d4-099d-4e8b-b3a4-d2fc7152c1ef\") " pod="openstack/ovsdbserver-sb-0" Jan 21 11:25:23 crc kubenswrapper[4824]: I0121 11:25:23.047167 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"ovsdbserver-sb-0\" (UID: \"e86f23d4-099d-4e8b-b3a4-d2fc7152c1ef\") " pod="openstack/ovsdbserver-sb-0" Jan 21 11:25:23 crc kubenswrapper[4824]: I0121 11:25:23.168132 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Jan 21 11:25:23 crc kubenswrapper[4824]: I0121 11:25:23.525218 4824 generic.go:334] "Generic (PLEG): container finished" podID="f67dfc52-dc0e-404e-af22-75c6a3d10721" containerID="0ad321acd47c8d7e7d82f491f6ee8db60e60e3e1cab256cda20de9be81202aa0" exitCode=0 Jan 21 11:25:23 crc kubenswrapper[4824]: I0121 11:25:23.525290 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-zspkp" event={"ID":"f67dfc52-dc0e-404e-af22-75c6a3d10721","Type":"ContainerDied","Data":"0ad321acd47c8d7e7d82f491f6ee8db60e60e3e1cab256cda20de9be81202aa0"} Jan 21 11:25:23 crc kubenswrapper[4824]: I0121 11:25:23.526128 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"5c0b5876-da04-49e6-8946-4fc6060a64ec","Type":"ContainerStarted","Data":"a9120e7f79937c25195d3b830feb6ca41ab891b48425d59e1a0ca2c0018a23cd"} Jan 21 11:25:24 crc kubenswrapper[4824]: I0121 11:25:24.952416 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-zspkp" Jan 21 11:25:25 crc kubenswrapper[4824]: I0121 11:25:25.044268 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f67dfc52-dc0e-404e-af22-75c6a3d10721-dns-svc\") pod \"f67dfc52-dc0e-404e-af22-75c6a3d10721\" (UID: \"f67dfc52-dc0e-404e-af22-75c6a3d10721\") " Jan 21 11:25:25 crc kubenswrapper[4824]: I0121 11:25:25.044320 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f67dfc52-dc0e-404e-af22-75c6a3d10721-config\") pod \"f67dfc52-dc0e-404e-af22-75c6a3d10721\" (UID: \"f67dfc52-dc0e-404e-af22-75c6a3d10721\") " Jan 21 11:25:25 crc kubenswrapper[4824]: I0121 11:25:25.077781 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-vrpmw"] Jan 21 11:25:25 crc kubenswrapper[4824]: I0121 11:25:25.091583 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-4xld6"] Jan 21 11:25:25 crc kubenswrapper[4824]: W0121 11:25:25.093812 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8652f3f6_1ad7_49d9_9048_827be17a3dd0.slice/crio-352bf16a65066bf3212e9a4c7cf5a4cc64f59beec5baff4e34e6c30d19cd91e0 WatchSource:0}: Error finding container 352bf16a65066bf3212e9a4c7cf5a4cc64f59beec5baff4e34e6c30d19cd91e0: Status 404 returned error can't find the container with id 352bf16a65066bf3212e9a4c7cf5a4cc64f59beec5baff4e34e6c30d19cd91e0 Jan 21 11:25:25 crc kubenswrapper[4824]: W0121 11:25:25.100056 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod56a2624d_7ab8_43ea_967e_e4112c693148.slice/crio-bbfbc490e1ea828e4c1b756158c30f088c089cc696efd16dfed5b06bd0756946 WatchSource:0}: Error finding container bbfbc490e1ea828e4c1b756158c30f088c089cc696efd16dfed5b06bd0756946: Status 404 returned error can't find the container with id bbfbc490e1ea828e4c1b756158c30f088c089cc696efd16dfed5b06bd0756946 Jan 21 11:25:25 crc kubenswrapper[4824]: I0121 11:25:25.120439 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Jan 21 11:25:25 crc kubenswrapper[4824]: I0121 11:25:25.140927 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f67dfc52-dc0e-404e-af22-75c6a3d10721-config" (OuterVolumeSpecName: "config") pod "f67dfc52-dc0e-404e-af22-75c6a3d10721" (UID: "f67dfc52-dc0e-404e-af22-75c6a3d10721"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:25:25 crc kubenswrapper[4824]: I0121 11:25:25.145999 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-msx57\" (UniqueName: \"kubernetes.io/projected/f67dfc52-dc0e-404e-af22-75c6a3d10721-kube-api-access-msx57\") pod \"f67dfc52-dc0e-404e-af22-75c6a3d10721\" (UID: \"f67dfc52-dc0e-404e-af22-75c6a3d10721\") " Jan 21 11:25:25 crc kubenswrapper[4824]: I0121 11:25:25.146508 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f67dfc52-dc0e-404e-af22-75c6a3d10721-config\") on node \"crc\" DevicePath \"\"" Jan 21 11:25:25 crc kubenswrapper[4824]: I0121 11:25:25.149627 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f67dfc52-dc0e-404e-af22-75c6a3d10721-kube-api-access-msx57" (OuterVolumeSpecName: "kube-api-access-msx57") pod "f67dfc52-dc0e-404e-af22-75c6a3d10721" (UID: "f67dfc52-dc0e-404e-af22-75c6a3d10721"). InnerVolumeSpecName "kube-api-access-msx57". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:25:25 crc kubenswrapper[4824]: I0121 11:25:25.174614 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f67dfc52-dc0e-404e-af22-75c6a3d10721-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "f67dfc52-dc0e-404e-af22-75c6a3d10721" (UID: "f67dfc52-dc0e-404e-af22-75c6a3d10721"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:25:25 crc kubenswrapper[4824]: I0121 11:25:25.247406 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-msx57\" (UniqueName: \"kubernetes.io/projected/f67dfc52-dc0e-404e-af22-75c6a3d10721-kube-api-access-msx57\") on node \"crc\" DevicePath \"\"" Jan 21 11:25:25 crc kubenswrapper[4824]: I0121 11:25:25.247433 4824 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f67dfc52-dc0e-404e-af22-75c6a3d10721-dns-svc\") on node \"crc\" DevicePath \"\"" Jan 21 11:25:25 crc kubenswrapper[4824]: I0121 11:25:25.497121 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-57d769cc4f-nrqwx" Jan 21 11:25:25 crc kubenswrapper[4824]: I0121 11:25:25.544717 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"dba0050b-8a73-4355-a1e0-7c9a03557ead","Type":"ContainerStarted","Data":"15764a7d669d69b32030fea1055b01aba131468d39177b7c5dbb1e56a61e22a7"} Jan 21 11:25:25 crc kubenswrapper[4824]: I0121 11:25:25.545886 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Jan 21 11:25:25 crc kubenswrapper[4824]: I0121 11:25:25.547864 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"2e3a943c-ee06-4e54-a984-34ecb1320f89","Type":"ContainerStarted","Data":"6e0a2a51cb10bfe8bc65ec0123c4fa76489cf9873a673344ebc1fbf8065155d5"} Jan 21 11:25:25 crc kubenswrapper[4824]: I0121 11:25:25.548355 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Jan 21 11:25:25 crc kubenswrapper[4824]: I0121 11:25:25.549293 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"e86f23d4-099d-4e8b-b3a4-d2fc7152c1ef","Type":"ContainerStarted","Data":"4ffdf56ad46938e9ac8326bf83cd4631d601d678ad6cb95715b285cb0cfb9687"} Jan 21 11:25:25 crc kubenswrapper[4824]: I0121 11:25:25.550798 4824 generic.go:334] "Generic (PLEG): container finished" podID="56a2624d-7ab8-43ea-967e-e4112c693148" containerID="7a62ec9112a17b21db69ae3a05c4ff4ba7010a9cc978e58063d41683f7c63e93" exitCode=0 Jan 21 11:25:25 crc kubenswrapper[4824]: I0121 11:25:25.550843 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-4xld6" event={"ID":"56a2624d-7ab8-43ea-967e-e4112c693148","Type":"ContainerDied","Data":"7a62ec9112a17b21db69ae3a05c4ff4ba7010a9cc978e58063d41683f7c63e93"} Jan 21 11:25:25 crc kubenswrapper[4824]: I0121 11:25:25.550859 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-4xld6" event={"ID":"56a2624d-7ab8-43ea-967e-e4112c693148","Type":"ContainerStarted","Data":"bbfbc490e1ea828e4c1b756158c30f088c089cc696efd16dfed5b06bd0756946"} Jan 21 11:25:25 crc kubenswrapper[4824]: I0121 11:25:25.557848 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"7cd8d62b-ab74-4a1b-95f5-7253ed6c0346","Type":"ContainerStarted","Data":"17209275dbf42e0eb684762f5bbf9fbe138f725455b92becb5b33da1dce19d39"} Jan 21 11:25:25 crc kubenswrapper[4824]: I0121 11:25:25.561933 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-zspkp" event={"ID":"f67dfc52-dc0e-404e-af22-75c6a3d10721","Type":"ContainerDied","Data":"4c0d7f6c5bc10bc5e3cf172c7b361f178beb8f1decb24246db808a7099969563"} Jan 21 11:25:25 crc kubenswrapper[4824]: I0121 11:25:25.562006 4824 scope.go:117] "RemoveContainer" containerID="0ad321acd47c8d7e7d82f491f6ee8db60e60e3e1cab256cda20de9be81202aa0" Jan 21 11:25:25 crc kubenswrapper[4824]: I0121 11:25:25.562084 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-zspkp" Jan 21 11:25:25 crc kubenswrapper[4824]: I0121 11:25:25.563104 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=6.286770926 podStartE2EDuration="11.563090052s" podCreationTimestamp="2026-01-21 11:25:14 +0000 UTC" firstStartedPulling="2026-01-21 11:25:19.272123382 +0000 UTC m=+861.565152674" lastFinishedPulling="2026-01-21 11:25:24.548442509 +0000 UTC m=+866.841471800" observedRunningTime="2026-01-21 11:25:25.559134817 +0000 UTC m=+867.852164108" watchObservedRunningTime="2026-01-21 11:25:25.563090052 +0000 UTC m=+867.856119344" Jan 21 11:25:25 crc kubenswrapper[4824]: I0121 11:25:25.566411 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-vrpmw" event={"ID":"8652f3f6-1ad7-49d9-9048-827be17a3dd0","Type":"ContainerStarted","Data":"352bf16a65066bf3212e9a4c7cf5a4cc64f59beec5baff4e34e6c30d19cd91e0"} Jan 21 11:25:25 crc kubenswrapper[4824]: I0121 11:25:25.575099 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=4.11025354 podStartE2EDuration="9.575084411s" podCreationTimestamp="2026-01-21 11:25:16 +0000 UTC" firstStartedPulling="2026-01-21 11:25:19.157110037 +0000 UTC m=+861.450139330" lastFinishedPulling="2026-01-21 11:25:24.621940909 +0000 UTC m=+866.914970201" observedRunningTime="2026-01-21 11:25:25.569416167 +0000 UTC m=+867.862445459" watchObservedRunningTime="2026-01-21 11:25:25.575084411 +0000 UTC m=+867.868113703" Jan 21 11:25:25 crc kubenswrapper[4824]: I0121 11:25:25.575773 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"218aca21-4c37-4862-b891-fc228baa5348","Type":"ContainerStarted","Data":"0cdcef9637307812b896082a0b7c8e830ed364eb95d323dde4a524f05b45fc8d"} Jan 21 11:25:25 crc kubenswrapper[4824]: I0121 11:25:25.695398 4824 scope.go:117] "RemoveContainer" containerID="2fabf442546067371acccd858ab70f06b5ad7ea6a02d270ec355916d978e1384" Jan 21 11:25:25 crc kubenswrapper[4824]: I0121 11:25:25.712729 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-zspkp"] Jan 21 11:25:25 crc kubenswrapper[4824]: I0121 11:25:25.717139 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-zspkp"] Jan 21 11:25:26 crc kubenswrapper[4824]: I0121 11:25:26.061518 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f67dfc52-dc0e-404e-af22-75c6a3d10721" path="/var/lib/kubelet/pods/f67dfc52-dc0e-404e-af22-75c6a3d10721/volumes" Jan 21 11:25:26 crc kubenswrapper[4824]: I0121 11:25:26.582396 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"156fec50-b486-4e84-a7bf-b40491a863c7","Type":"ContainerStarted","Data":"57e1badc440cdd38a60828271554d42b57947d40b89d5290cb4d30f5fcb54487"} Jan 21 11:25:26 crc kubenswrapper[4824]: I0121 11:25:26.585666 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-4xld6" event={"ID":"56a2624d-7ab8-43ea-967e-e4112c693148","Type":"ContainerStarted","Data":"9d5f40499372167e6cfebad13e72e3e53477602cccd684ee9c4ce932d1f266aa"} Jan 21 11:25:26 crc kubenswrapper[4824]: I0121 11:25:26.586453 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7fd796d7df-4xld6" Jan 21 11:25:26 crc kubenswrapper[4824]: I0121 11:25:26.588682 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"84e74d7b-18a0-4a3f-8680-6246ac538a6e","Type":"ContainerStarted","Data":"12b90320ffa88a78c84d62cf8b65c324fb6c4b884b2714e0a27b33b29cca2e17"} Jan 21 11:25:26 crc kubenswrapper[4824]: I0121 11:25:26.614747 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7fd796d7df-4xld6" podStartSLOduration=4.614732308 podStartE2EDuration="4.614732308s" podCreationTimestamp="2026-01-21 11:25:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:25:26.613557274 +0000 UTC m=+868.906586565" watchObservedRunningTime="2026-01-21 11:25:26.614732308 +0000 UTC m=+868.907761600" Jan 21 11:25:28 crc kubenswrapper[4824]: I0121 11:25:28.600180 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"5c0b5876-da04-49e6-8946-4fc6060a64ec","Type":"ContainerStarted","Data":"c5bc540e824f8315de4aaa98521390c3084ae1ca6f114bf24ca648548b63b037"} Jan 21 11:25:28 crc kubenswrapper[4824]: I0121 11:25:28.600221 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"5c0b5876-da04-49e6-8946-4fc6060a64ec","Type":"ContainerStarted","Data":"5a4b275ec2f8e3f778381b1131554cd298f52f4fc6b173ea49a8770ea4a60915"} Jan 21 11:25:28 crc kubenswrapper[4824]: I0121 11:25:28.601829 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"e86f23d4-099d-4e8b-b3a4-d2fc7152c1ef","Type":"ContainerStarted","Data":"db803036a9410506d79d5627d9543774fc019eb3ab50fab160d3258604367f1d"} Jan 21 11:25:28 crc kubenswrapper[4824]: I0121 11:25:28.601869 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"e86f23d4-099d-4e8b-b3a4-d2fc7152c1ef","Type":"ContainerStarted","Data":"719b4c4c84cd5cc9db2a9703e4611c3374447f18dbbb7225f53d8bbfa4a32c07"} Jan 21 11:25:28 crc kubenswrapper[4824]: I0121 11:25:28.603196 4824 generic.go:334] "Generic (PLEG): container finished" podID="7cd8d62b-ab74-4a1b-95f5-7253ed6c0346" containerID="17209275dbf42e0eb684762f5bbf9fbe138f725455b92becb5b33da1dce19d39" exitCode=0 Jan 21 11:25:28 crc kubenswrapper[4824]: I0121 11:25:28.603253 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"7cd8d62b-ab74-4a1b-95f5-7253ed6c0346","Type":"ContainerDied","Data":"17209275dbf42e0eb684762f5bbf9fbe138f725455b92becb5b33da1dce19d39"} Jan 21 11:25:28 crc kubenswrapper[4824]: I0121 11:25:28.604564 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-vrpmw" event={"ID":"8652f3f6-1ad7-49d9-9048-827be17a3dd0","Type":"ContainerStarted","Data":"aae178d791168024804772a47be73fd13944340220d55c19a31acecf8d0e4018"} Jan 21 11:25:28 crc kubenswrapper[4824]: I0121 11:25:28.605910 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-gr9gj" event={"ID":"b27b3f77-8476-4f3e-a946-f4f26fc66cb5","Type":"ContainerStarted","Data":"f2c83cbd5821aafcf83f1956f57760cd86d073accfd9a9bee311534424f7f71f"} Jan 21 11:25:28 crc kubenswrapper[4824]: I0121 11:25:28.606970 4824 generic.go:334] "Generic (PLEG): container finished" podID="218aca21-4c37-4862-b891-fc228baa5348" containerID="0cdcef9637307812b896082a0b7c8e830ed364eb95d323dde4a524f05b45fc8d" exitCode=0 Jan 21 11:25:28 crc kubenswrapper[4824]: I0121 11:25:28.607024 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"218aca21-4c37-4862-b891-fc228baa5348","Type":"ContainerDied","Data":"0cdcef9637307812b896082a0b7c8e830ed364eb95d323dde4a524f05b45fc8d"} Jan 21 11:25:28 crc kubenswrapper[4824]: I0121 11:25:28.610422 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-htwd6" event={"ID":"511f2b6d-a08d-49f8-b393-ab222219d4a7","Type":"ContainerStarted","Data":"fbe475794bcef10bb1b0130b5c614e06187deb33b9d1cd49cfa882f3d3c6711d"} Jan 21 11:25:28 crc kubenswrapper[4824]: I0121 11:25:28.610453 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-htwd6" Jan 21 11:25:28 crc kubenswrapper[4824]: I0121 11:25:28.617509 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=4.078714048 podStartE2EDuration="9.617492879s" podCreationTimestamp="2026-01-21 11:25:19 +0000 UTC" firstStartedPulling="2026-01-21 11:25:22.723951597 +0000 UTC m=+865.016980889" lastFinishedPulling="2026-01-21 11:25:28.262730428 +0000 UTC m=+870.555759720" observedRunningTime="2026-01-21 11:25:28.615473622 +0000 UTC m=+870.908502914" watchObservedRunningTime="2026-01-21 11:25:28.617492879 +0000 UTC m=+870.910522170" Jan 21 11:25:28 crc kubenswrapper[4824]: I0121 11:25:28.672394 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-vrpmw" podStartSLOduration=4.489699936 podStartE2EDuration="7.672376763s" podCreationTimestamp="2026-01-21 11:25:21 +0000 UTC" firstStartedPulling="2026-01-21 11:25:25.097757642 +0000 UTC m=+867.390786934" lastFinishedPulling="2026-01-21 11:25:28.280434469 +0000 UTC m=+870.573463761" observedRunningTime="2026-01-21 11:25:28.670622175 +0000 UTC m=+870.963651477" watchObservedRunningTime="2026-01-21 11:25:28.672376763 +0000 UTC m=+870.965406055" Jan 21 11:25:28 crc kubenswrapper[4824]: I0121 11:25:28.694179 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=4.554566417 podStartE2EDuration="7.694162969s" podCreationTimestamp="2026-01-21 11:25:21 +0000 UTC" firstStartedPulling="2026-01-21 11:25:25.128670532 +0000 UTC m=+867.421699824" lastFinishedPulling="2026-01-21 11:25:28.268267085 +0000 UTC m=+870.561296376" observedRunningTime="2026-01-21 11:25:28.688630249 +0000 UTC m=+870.981659541" watchObservedRunningTime="2026-01-21 11:25:28.694162969 +0000 UTC m=+870.987192261" Jan 21 11:25:28 crc kubenswrapper[4824]: I0121 11:25:28.711561 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-htwd6" podStartSLOduration=2.43187333 podStartE2EDuration="9.711533842s" podCreationTimestamp="2026-01-21 11:25:19 +0000 UTC" firstStartedPulling="2026-01-21 11:25:20.980756004 +0000 UTC m=+863.273785297" lastFinishedPulling="2026-01-21 11:25:28.260416527 +0000 UTC m=+870.553445809" observedRunningTime="2026-01-21 11:25:28.706154732 +0000 UTC m=+870.999184025" watchObservedRunningTime="2026-01-21 11:25:28.711533842 +0000 UTC m=+871.004563134" Jan 21 11:25:28 crc kubenswrapper[4824]: I0121 11:25:28.954804 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-4xld6"] Jan 21 11:25:28 crc kubenswrapper[4824]: I0121 11:25:28.972750 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-vfbzh"] Jan 21 11:25:28 crc kubenswrapper[4824]: E0121 11:25:28.973058 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f67dfc52-dc0e-404e-af22-75c6a3d10721" containerName="init" Jan 21 11:25:28 crc kubenswrapper[4824]: I0121 11:25:28.973075 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="f67dfc52-dc0e-404e-af22-75c6a3d10721" containerName="init" Jan 21 11:25:28 crc kubenswrapper[4824]: E0121 11:25:28.973088 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f67dfc52-dc0e-404e-af22-75c6a3d10721" containerName="dnsmasq-dns" Jan 21 11:25:28 crc kubenswrapper[4824]: I0121 11:25:28.973094 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="f67dfc52-dc0e-404e-af22-75c6a3d10721" containerName="dnsmasq-dns" Jan 21 11:25:28 crc kubenswrapper[4824]: I0121 11:25:28.973243 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="f67dfc52-dc0e-404e-af22-75c6a3d10721" containerName="dnsmasq-dns" Jan 21 11:25:28 crc kubenswrapper[4824]: I0121 11:25:28.973923 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-vfbzh" Jan 21 11:25:28 crc kubenswrapper[4824]: I0121 11:25:28.976448 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Jan 21 11:25:28 crc kubenswrapper[4824]: I0121 11:25:28.990986 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-vfbzh"] Jan 21 11:25:29 crc kubenswrapper[4824]: I0121 11:25:29.110405 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/292b341f-0f9e-411e-9f85-47bde4dcb2c4-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-vfbzh\" (UID: \"292b341f-0f9e-411e-9f85-47bde4dcb2c4\") " pod="openstack/dnsmasq-dns-86db49b7ff-vfbzh" Jan 21 11:25:29 crc kubenswrapper[4824]: I0121 11:25:29.110481 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/292b341f-0f9e-411e-9f85-47bde4dcb2c4-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-vfbzh\" (UID: \"292b341f-0f9e-411e-9f85-47bde4dcb2c4\") " pod="openstack/dnsmasq-dns-86db49b7ff-vfbzh" Jan 21 11:25:29 crc kubenswrapper[4824]: I0121 11:25:29.110527 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hg5nl\" (UniqueName: \"kubernetes.io/projected/292b341f-0f9e-411e-9f85-47bde4dcb2c4-kube-api-access-hg5nl\") pod \"dnsmasq-dns-86db49b7ff-vfbzh\" (UID: \"292b341f-0f9e-411e-9f85-47bde4dcb2c4\") " pod="openstack/dnsmasq-dns-86db49b7ff-vfbzh" Jan 21 11:25:29 crc kubenswrapper[4824]: I0121 11:25:29.110635 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/292b341f-0f9e-411e-9f85-47bde4dcb2c4-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-vfbzh\" (UID: \"292b341f-0f9e-411e-9f85-47bde4dcb2c4\") " pod="openstack/dnsmasq-dns-86db49b7ff-vfbzh" Jan 21 11:25:29 crc kubenswrapper[4824]: I0121 11:25:29.110708 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/292b341f-0f9e-411e-9f85-47bde4dcb2c4-config\") pod \"dnsmasq-dns-86db49b7ff-vfbzh\" (UID: \"292b341f-0f9e-411e-9f85-47bde4dcb2c4\") " pod="openstack/dnsmasq-dns-86db49b7ff-vfbzh" Jan 21 11:25:29 crc kubenswrapper[4824]: I0121 11:25:29.168923 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Jan 21 11:25:29 crc kubenswrapper[4824]: I0121 11:25:29.212020 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/292b341f-0f9e-411e-9f85-47bde4dcb2c4-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-vfbzh\" (UID: \"292b341f-0f9e-411e-9f85-47bde4dcb2c4\") " pod="openstack/dnsmasq-dns-86db49b7ff-vfbzh" Jan 21 11:25:29 crc kubenswrapper[4824]: I0121 11:25:29.212089 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/292b341f-0f9e-411e-9f85-47bde4dcb2c4-config\") pod \"dnsmasq-dns-86db49b7ff-vfbzh\" (UID: \"292b341f-0f9e-411e-9f85-47bde4dcb2c4\") " pod="openstack/dnsmasq-dns-86db49b7ff-vfbzh" Jan 21 11:25:29 crc kubenswrapper[4824]: I0121 11:25:29.212153 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/292b341f-0f9e-411e-9f85-47bde4dcb2c4-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-vfbzh\" (UID: \"292b341f-0f9e-411e-9f85-47bde4dcb2c4\") " pod="openstack/dnsmasq-dns-86db49b7ff-vfbzh" Jan 21 11:25:29 crc kubenswrapper[4824]: I0121 11:25:29.212182 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/292b341f-0f9e-411e-9f85-47bde4dcb2c4-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-vfbzh\" (UID: \"292b341f-0f9e-411e-9f85-47bde4dcb2c4\") " pod="openstack/dnsmasq-dns-86db49b7ff-vfbzh" Jan 21 11:25:29 crc kubenswrapper[4824]: I0121 11:25:29.212220 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hg5nl\" (UniqueName: \"kubernetes.io/projected/292b341f-0f9e-411e-9f85-47bde4dcb2c4-kube-api-access-hg5nl\") pod \"dnsmasq-dns-86db49b7ff-vfbzh\" (UID: \"292b341f-0f9e-411e-9f85-47bde4dcb2c4\") " pod="openstack/dnsmasq-dns-86db49b7ff-vfbzh" Jan 21 11:25:29 crc kubenswrapper[4824]: I0121 11:25:29.212868 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/292b341f-0f9e-411e-9f85-47bde4dcb2c4-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-vfbzh\" (UID: \"292b341f-0f9e-411e-9f85-47bde4dcb2c4\") " pod="openstack/dnsmasq-dns-86db49b7ff-vfbzh" Jan 21 11:25:29 crc kubenswrapper[4824]: I0121 11:25:29.213172 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/292b341f-0f9e-411e-9f85-47bde4dcb2c4-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-vfbzh\" (UID: \"292b341f-0f9e-411e-9f85-47bde4dcb2c4\") " pod="openstack/dnsmasq-dns-86db49b7ff-vfbzh" Jan 21 11:25:29 crc kubenswrapper[4824]: I0121 11:25:29.213310 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/292b341f-0f9e-411e-9f85-47bde4dcb2c4-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-vfbzh\" (UID: \"292b341f-0f9e-411e-9f85-47bde4dcb2c4\") " pod="openstack/dnsmasq-dns-86db49b7ff-vfbzh" Jan 21 11:25:29 crc kubenswrapper[4824]: I0121 11:25:29.213353 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/292b341f-0f9e-411e-9f85-47bde4dcb2c4-config\") pod \"dnsmasq-dns-86db49b7ff-vfbzh\" (UID: \"292b341f-0f9e-411e-9f85-47bde4dcb2c4\") " pod="openstack/dnsmasq-dns-86db49b7ff-vfbzh" Jan 21 11:25:29 crc kubenswrapper[4824]: I0121 11:25:29.226236 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hg5nl\" (UniqueName: \"kubernetes.io/projected/292b341f-0f9e-411e-9f85-47bde4dcb2c4-kube-api-access-hg5nl\") pod \"dnsmasq-dns-86db49b7ff-vfbzh\" (UID: \"292b341f-0f9e-411e-9f85-47bde4dcb2c4\") " pod="openstack/dnsmasq-dns-86db49b7ff-vfbzh" Jan 21 11:25:29 crc kubenswrapper[4824]: I0121 11:25:29.286492 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-vfbzh" Jan 21 11:25:29 crc kubenswrapper[4824]: I0121 11:25:29.594527 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Jan 21 11:25:29 crc kubenswrapper[4824]: I0121 11:25:29.616971 4824 generic.go:334] "Generic (PLEG): container finished" podID="b27b3f77-8476-4f3e-a946-f4f26fc66cb5" containerID="f2c83cbd5821aafcf83f1956f57760cd86d073accfd9a9bee311534424f7f71f" exitCode=0 Jan 21 11:25:29 crc kubenswrapper[4824]: I0121 11:25:29.617001 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-gr9gj" event={"ID":"b27b3f77-8476-4f3e-a946-f4f26fc66cb5","Type":"ContainerDied","Data":"f2c83cbd5821aafcf83f1956f57760cd86d073accfd9a9bee311534424f7f71f"} Jan 21 11:25:29 crc kubenswrapper[4824]: I0121 11:25:29.619603 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"218aca21-4c37-4862-b891-fc228baa5348","Type":"ContainerStarted","Data":"b75f9993c8911f138ee27d672fe969e86c04facee2fd173c5eb3fe3d2c104eb2"} Jan 21 11:25:29 crc kubenswrapper[4824]: I0121 11:25:29.621883 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"7cd8d62b-ab74-4a1b-95f5-7253ed6c0346","Type":"ContainerStarted","Data":"91180231ce54266600bab28049663d812ad89e282ad5a82636c446ff381b8e4c"} Jan 21 11:25:29 crc kubenswrapper[4824]: I0121 11:25:29.622335 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7fd796d7df-4xld6" podUID="56a2624d-7ab8-43ea-967e-e4112c693148" containerName="dnsmasq-dns" containerID="cri-o://9d5f40499372167e6cfebad13e72e3e53477602cccd684ee9c4ce932d1f266aa" gracePeriod=10 Jan 21 11:25:29 crc kubenswrapper[4824]: I0121 11:25:29.647780 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Jan 21 11:25:29 crc kubenswrapper[4824]: I0121 11:25:29.657802 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-vfbzh"] Jan 21 11:25:29 crc kubenswrapper[4824]: I0121 11:25:29.668527 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=12.364585947 podStartE2EDuration="17.668513293s" podCreationTimestamp="2026-01-21 11:25:12 +0000 UTC" firstStartedPulling="2026-01-21 11:25:19.272909113 +0000 UTC m=+861.565938405" lastFinishedPulling="2026-01-21 11:25:24.576836459 +0000 UTC m=+866.869865751" observedRunningTime="2026-01-21 11:25:29.664151141 +0000 UTC m=+871.957180433" watchObservedRunningTime="2026-01-21 11:25:29.668513293 +0000 UTC m=+871.961542586" Jan 21 11:25:29 crc kubenswrapper[4824]: I0121 11:25:29.697529 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=13.346555629 podStartE2EDuration="18.697515361s" podCreationTimestamp="2026-01-21 11:25:11 +0000 UTC" firstStartedPulling="2026-01-21 11:25:19.273342219 +0000 UTC m=+861.566371511" lastFinishedPulling="2026-01-21 11:25:24.62430195 +0000 UTC m=+866.917331243" observedRunningTime="2026-01-21 11:25:29.694439844 +0000 UTC m=+871.987469136" watchObservedRunningTime="2026-01-21 11:25:29.697515361 +0000 UTC m=+871.990544653" Jan 21 11:25:29 crc kubenswrapper[4824]: W0121 11:25:29.712046 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod292b341f_0f9e_411e_9f85_47bde4dcb2c4.slice/crio-c0704ee36eacd59c7b38561ac33e2d4956e244055bb28267047fee7b43cd9db0 WatchSource:0}: Error finding container c0704ee36eacd59c7b38561ac33e2d4956e244055bb28267047fee7b43cd9db0: Status 404 returned error can't find the container with id c0704ee36eacd59c7b38561ac33e2d4956e244055bb28267047fee7b43cd9db0 Jan 21 11:25:30 crc kubenswrapper[4824]: I0121 11:25:30.034674 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-4xld6" Jan 21 11:25:30 crc kubenswrapper[4824]: I0121 11:25:30.127085 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/56a2624d-7ab8-43ea-967e-e4112c693148-config\") pod \"56a2624d-7ab8-43ea-967e-e4112c693148\" (UID: \"56a2624d-7ab8-43ea-967e-e4112c693148\") " Jan 21 11:25:30 crc kubenswrapper[4824]: I0121 11:25:30.127137 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/56a2624d-7ab8-43ea-967e-e4112c693148-ovsdbserver-nb\") pod \"56a2624d-7ab8-43ea-967e-e4112c693148\" (UID: \"56a2624d-7ab8-43ea-967e-e4112c693148\") " Jan 21 11:25:30 crc kubenswrapper[4824]: I0121 11:25:30.127192 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rlmf6\" (UniqueName: \"kubernetes.io/projected/56a2624d-7ab8-43ea-967e-e4112c693148-kube-api-access-rlmf6\") pod \"56a2624d-7ab8-43ea-967e-e4112c693148\" (UID: \"56a2624d-7ab8-43ea-967e-e4112c693148\") " Jan 21 11:25:30 crc kubenswrapper[4824]: I0121 11:25:30.127231 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/56a2624d-7ab8-43ea-967e-e4112c693148-dns-svc\") pod \"56a2624d-7ab8-43ea-967e-e4112c693148\" (UID: \"56a2624d-7ab8-43ea-967e-e4112c693148\") " Jan 21 11:25:30 crc kubenswrapper[4824]: I0121 11:25:30.131367 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/56a2624d-7ab8-43ea-967e-e4112c693148-kube-api-access-rlmf6" (OuterVolumeSpecName: "kube-api-access-rlmf6") pod "56a2624d-7ab8-43ea-967e-e4112c693148" (UID: "56a2624d-7ab8-43ea-967e-e4112c693148"). InnerVolumeSpecName "kube-api-access-rlmf6". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:25:30 crc kubenswrapper[4824]: I0121 11:25:30.154608 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/56a2624d-7ab8-43ea-967e-e4112c693148-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "56a2624d-7ab8-43ea-967e-e4112c693148" (UID: "56a2624d-7ab8-43ea-967e-e4112c693148"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:25:30 crc kubenswrapper[4824]: I0121 11:25:30.155528 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/56a2624d-7ab8-43ea-967e-e4112c693148-config" (OuterVolumeSpecName: "config") pod "56a2624d-7ab8-43ea-967e-e4112c693148" (UID: "56a2624d-7ab8-43ea-967e-e4112c693148"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:25:30 crc kubenswrapper[4824]: I0121 11:25:30.157172 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/56a2624d-7ab8-43ea-967e-e4112c693148-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "56a2624d-7ab8-43ea-967e-e4112c693148" (UID: "56a2624d-7ab8-43ea-967e-e4112c693148"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:25:30 crc kubenswrapper[4824]: I0121 11:25:30.229027 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/56a2624d-7ab8-43ea-967e-e4112c693148-config\") on node \"crc\" DevicePath \"\"" Jan 21 11:25:30 crc kubenswrapper[4824]: I0121 11:25:30.229061 4824 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/56a2624d-7ab8-43ea-967e-e4112c693148-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Jan 21 11:25:30 crc kubenswrapper[4824]: I0121 11:25:30.229075 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rlmf6\" (UniqueName: \"kubernetes.io/projected/56a2624d-7ab8-43ea-967e-e4112c693148-kube-api-access-rlmf6\") on node \"crc\" DevicePath \"\"" Jan 21 11:25:30 crc kubenswrapper[4824]: I0121 11:25:30.229084 4824 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/56a2624d-7ab8-43ea-967e-e4112c693148-dns-svc\") on node \"crc\" DevicePath \"\"" Jan 21 11:25:30 crc kubenswrapper[4824]: I0121 11:25:30.627701 4824 generic.go:334] "Generic (PLEG): container finished" podID="56a2624d-7ab8-43ea-967e-e4112c693148" containerID="9d5f40499372167e6cfebad13e72e3e53477602cccd684ee9c4ce932d1f266aa" exitCode=0 Jan 21 11:25:30 crc kubenswrapper[4824]: I0121 11:25:30.627773 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-4xld6" Jan 21 11:25:30 crc kubenswrapper[4824]: I0121 11:25:30.627792 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-4xld6" event={"ID":"56a2624d-7ab8-43ea-967e-e4112c693148","Type":"ContainerDied","Data":"9d5f40499372167e6cfebad13e72e3e53477602cccd684ee9c4ce932d1f266aa"} Jan 21 11:25:30 crc kubenswrapper[4824]: I0121 11:25:30.628206 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-4xld6" event={"ID":"56a2624d-7ab8-43ea-967e-e4112c693148","Type":"ContainerDied","Data":"bbfbc490e1ea828e4c1b756158c30f088c089cc696efd16dfed5b06bd0756946"} Jan 21 11:25:30 crc kubenswrapper[4824]: I0121 11:25:30.628224 4824 scope.go:117] "RemoveContainer" containerID="9d5f40499372167e6cfebad13e72e3e53477602cccd684ee9c4ce932d1f266aa" Jan 21 11:25:30 crc kubenswrapper[4824]: I0121 11:25:30.629311 4824 generic.go:334] "Generic (PLEG): container finished" podID="292b341f-0f9e-411e-9f85-47bde4dcb2c4" containerID="b9f66e73f6529373d74d9e19ba52303b84fabf65da4786d896c78db5e897096f" exitCode=0 Jan 21 11:25:30 crc kubenswrapper[4824]: I0121 11:25:30.629359 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-vfbzh" event={"ID":"292b341f-0f9e-411e-9f85-47bde4dcb2c4","Type":"ContainerDied","Data":"b9f66e73f6529373d74d9e19ba52303b84fabf65da4786d896c78db5e897096f"} Jan 21 11:25:30 crc kubenswrapper[4824]: I0121 11:25:30.629382 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-vfbzh" event={"ID":"292b341f-0f9e-411e-9f85-47bde4dcb2c4","Type":"ContainerStarted","Data":"c0704ee36eacd59c7b38561ac33e2d4956e244055bb28267047fee7b43cd9db0"} Jan 21 11:25:30 crc kubenswrapper[4824]: I0121 11:25:30.632079 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-gr9gj" event={"ID":"b27b3f77-8476-4f3e-a946-f4f26fc66cb5","Type":"ContainerStarted","Data":"377f978847f48f8fced112de3ef20125036e82a1968c553b3962be0500956157"} Jan 21 11:25:30 crc kubenswrapper[4824]: I0121 11:25:30.632132 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-gr9gj" event={"ID":"b27b3f77-8476-4f3e-a946-f4f26fc66cb5","Type":"ContainerStarted","Data":"2aa2e9ec59322b23149f826a199fa37ef4c7eedab4b46616599082f2a3764f7c"} Jan 21 11:25:30 crc kubenswrapper[4824]: I0121 11:25:30.632313 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-gr9gj" Jan 21 11:25:30 crc kubenswrapper[4824]: I0121 11:25:30.632451 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-gr9gj" Jan 21 11:25:30 crc kubenswrapper[4824]: I0121 11:25:30.644704 4824 scope.go:117] "RemoveContainer" containerID="7a62ec9112a17b21db69ae3a05c4ff4ba7010a9cc978e58063d41683f7c63e93" Jan 21 11:25:30 crc kubenswrapper[4824]: I0121 11:25:30.647809 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Jan 21 11:25:30 crc kubenswrapper[4824]: I0121 11:25:30.648146 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-gr9gj" podStartSLOduration=4.955803298 podStartE2EDuration="11.648136056s" podCreationTimestamp="2026-01-21 11:25:19 +0000 UTC" firstStartedPulling="2026-01-21 11:25:21.550696857 +0000 UTC m=+863.843726148" lastFinishedPulling="2026-01-21 11:25:28.243029613 +0000 UTC m=+870.536058906" observedRunningTime="2026-01-21 11:25:30.647982939 +0000 UTC m=+872.941012230" watchObservedRunningTime="2026-01-21 11:25:30.648136056 +0000 UTC m=+872.941165348" Jan 21 11:25:30 crc kubenswrapper[4824]: I0121 11:25:30.718697 4824 scope.go:117] "RemoveContainer" containerID="9d5f40499372167e6cfebad13e72e3e53477602cccd684ee9c4ce932d1f266aa" Jan 21 11:25:30 crc kubenswrapper[4824]: E0121 11:25:30.719826 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9d5f40499372167e6cfebad13e72e3e53477602cccd684ee9c4ce932d1f266aa\": container with ID starting with 9d5f40499372167e6cfebad13e72e3e53477602cccd684ee9c4ce932d1f266aa not found: ID does not exist" containerID="9d5f40499372167e6cfebad13e72e3e53477602cccd684ee9c4ce932d1f266aa" Jan 21 11:25:30 crc kubenswrapper[4824]: I0121 11:25:30.719885 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9d5f40499372167e6cfebad13e72e3e53477602cccd684ee9c4ce932d1f266aa"} err="failed to get container status \"9d5f40499372167e6cfebad13e72e3e53477602cccd684ee9c4ce932d1f266aa\": rpc error: code = NotFound desc = could not find container \"9d5f40499372167e6cfebad13e72e3e53477602cccd684ee9c4ce932d1f266aa\": container with ID starting with 9d5f40499372167e6cfebad13e72e3e53477602cccd684ee9c4ce932d1f266aa not found: ID does not exist" Jan 21 11:25:30 crc kubenswrapper[4824]: I0121 11:25:30.719907 4824 scope.go:117] "RemoveContainer" containerID="7a62ec9112a17b21db69ae3a05c4ff4ba7010a9cc978e58063d41683f7c63e93" Jan 21 11:25:30 crc kubenswrapper[4824]: E0121 11:25:30.720308 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7a62ec9112a17b21db69ae3a05c4ff4ba7010a9cc978e58063d41683f7c63e93\": container with ID starting with 7a62ec9112a17b21db69ae3a05c4ff4ba7010a9cc978e58063d41683f7c63e93 not found: ID does not exist" containerID="7a62ec9112a17b21db69ae3a05c4ff4ba7010a9cc978e58063d41683f7c63e93" Jan 21 11:25:30 crc kubenswrapper[4824]: I0121 11:25:30.720341 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7a62ec9112a17b21db69ae3a05c4ff4ba7010a9cc978e58063d41683f7c63e93"} err="failed to get container status \"7a62ec9112a17b21db69ae3a05c4ff4ba7010a9cc978e58063d41683f7c63e93\": rpc error: code = NotFound desc = could not find container \"7a62ec9112a17b21db69ae3a05c4ff4ba7010a9cc978e58063d41683f7c63e93\": container with ID starting with 7a62ec9112a17b21db69ae3a05c4ff4ba7010a9cc978e58063d41683f7c63e93 not found: ID does not exist" Jan 21 11:25:30 crc kubenswrapper[4824]: I0121 11:25:30.761910 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-4xld6"] Jan 21 11:25:30 crc kubenswrapper[4824]: I0121 11:25:30.767278 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-4xld6"] Jan 21 11:25:31 crc kubenswrapper[4824]: I0121 11:25:31.639286 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-vfbzh" event={"ID":"292b341f-0f9e-411e-9f85-47bde4dcb2c4","Type":"ContainerStarted","Data":"095e6c8fd89e1deec2822bbad9ce47d3bc5541769b4bdf66512d4544bcb46405"} Jan 21 11:25:31 crc kubenswrapper[4824]: I0121 11:25:31.656346 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-86db49b7ff-vfbzh" podStartSLOduration=3.656333091 podStartE2EDuration="3.656333091s" podCreationTimestamp="2026-01-21 11:25:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:25:31.651936754 +0000 UTC m=+873.944966046" watchObservedRunningTime="2026-01-21 11:25:31.656333091 +0000 UTC m=+873.949362383" Jan 21 11:25:32 crc kubenswrapper[4824]: I0121 11:25:32.056409 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="56a2624d-7ab8-43ea-967e-e4112c693148" path="/var/lib/kubelet/pods/56a2624d-7ab8-43ea-967e-e4112c693148/volumes" Jan 21 11:25:32 crc kubenswrapper[4824]: I0121 11:25:32.195011 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Jan 21 11:25:32 crc kubenswrapper[4824]: I0121 11:25:32.195418 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Jan 21 11:25:32 crc kubenswrapper[4824]: I0121 11:25:32.644091 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-86db49b7ff-vfbzh" Jan 21 11:25:32 crc kubenswrapper[4824]: I0121 11:25:32.676333 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Jan 21 11:25:32 crc kubenswrapper[4824]: I0121 11:25:32.975784 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Jan 21 11:25:32 crc kubenswrapper[4824]: I0121 11:25:32.975822 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Jan 21 11:25:33 crc kubenswrapper[4824]: I0121 11:25:33.191771 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Jan 21 11:25:33 crc kubenswrapper[4824]: I0121 11:25:33.193120 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Jan 21 11:25:33 crc kubenswrapper[4824]: I0121 11:25:33.695328 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Jan 21 11:25:34 crc kubenswrapper[4824]: I0121 11:25:34.269223 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-jbk7q"] Jan 21 11:25:34 crc kubenswrapper[4824]: E0121 11:25:34.269501 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="56a2624d-7ab8-43ea-967e-e4112c693148" containerName="init" Jan 21 11:25:34 crc kubenswrapper[4824]: I0121 11:25:34.269516 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="56a2624d-7ab8-43ea-967e-e4112c693148" containerName="init" Jan 21 11:25:34 crc kubenswrapper[4824]: E0121 11:25:34.269546 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="56a2624d-7ab8-43ea-967e-e4112c693148" containerName="dnsmasq-dns" Jan 21 11:25:34 crc kubenswrapper[4824]: I0121 11:25:34.269554 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="56a2624d-7ab8-43ea-967e-e4112c693148" containerName="dnsmasq-dns" Jan 21 11:25:34 crc kubenswrapper[4824]: I0121 11:25:34.269705 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="56a2624d-7ab8-43ea-967e-e4112c693148" containerName="dnsmasq-dns" Jan 21 11:25:34 crc kubenswrapper[4824]: I0121 11:25:34.270128 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-jbk7q" Jan 21 11:25:34 crc kubenswrapper[4824]: I0121 11:25:34.275966 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-jbk7q"] Jan 21 11:25:34 crc kubenswrapper[4824]: I0121 11:25:34.319676 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-383b-account-create-update-w99bv"] Jan 21 11:25:34 crc kubenswrapper[4824]: I0121 11:25:34.320508 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-383b-account-create-update-w99bv" Jan 21 11:25:34 crc kubenswrapper[4824]: I0121 11:25:34.322065 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Jan 21 11:25:34 crc kubenswrapper[4824]: I0121 11:25:34.324559 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Jan 21 11:25:34 crc kubenswrapper[4824]: I0121 11:25:34.324595 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Jan 21 11:25:34 crc kubenswrapper[4824]: I0121 11:25:34.327902 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-383b-account-create-update-w99bv"] Jan 21 11:25:34 crc kubenswrapper[4824]: I0121 11:25:34.373502 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Jan 21 11:25:34 crc kubenswrapper[4824]: I0121 11:25:34.382989 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7c01cf79-8a28-4229-9257-90124e66d3cc-operator-scripts\") pod \"keystone-db-create-jbk7q\" (UID: \"7c01cf79-8a28-4229-9257-90124e66d3cc\") " pod="openstack/keystone-db-create-jbk7q" Jan 21 11:25:34 crc kubenswrapper[4824]: I0121 11:25:34.383042 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hcw6h\" (UniqueName: \"kubernetes.io/projected/7c01cf79-8a28-4229-9257-90124e66d3cc-kube-api-access-hcw6h\") pod \"keystone-db-create-jbk7q\" (UID: \"7c01cf79-8a28-4229-9257-90124e66d3cc\") " pod="openstack/keystone-db-create-jbk7q" Jan 21 11:25:34 crc kubenswrapper[4824]: I0121 11:25:34.484786 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7c01cf79-8a28-4229-9257-90124e66d3cc-operator-scripts\") pod \"keystone-db-create-jbk7q\" (UID: \"7c01cf79-8a28-4229-9257-90124e66d3cc\") " pod="openstack/keystone-db-create-jbk7q" Jan 21 11:25:34 crc kubenswrapper[4824]: I0121 11:25:34.484867 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hcw6h\" (UniqueName: \"kubernetes.io/projected/7c01cf79-8a28-4229-9257-90124e66d3cc-kube-api-access-hcw6h\") pod \"keystone-db-create-jbk7q\" (UID: \"7c01cf79-8a28-4229-9257-90124e66d3cc\") " pod="openstack/keystone-db-create-jbk7q" Jan 21 11:25:34 crc kubenswrapper[4824]: I0121 11:25:34.485035 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bc853ab0-e92a-4ada-84fd-b59ac2af573d-operator-scripts\") pod \"keystone-383b-account-create-update-w99bv\" (UID: \"bc853ab0-e92a-4ada-84fd-b59ac2af573d\") " pod="openstack/keystone-383b-account-create-update-w99bv" Jan 21 11:25:34 crc kubenswrapper[4824]: I0121 11:25:34.485228 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t2qxk\" (UniqueName: \"kubernetes.io/projected/bc853ab0-e92a-4ada-84fd-b59ac2af573d-kube-api-access-t2qxk\") pod \"keystone-383b-account-create-update-w99bv\" (UID: \"bc853ab0-e92a-4ada-84fd-b59ac2af573d\") " pod="openstack/keystone-383b-account-create-update-w99bv" Jan 21 11:25:34 crc kubenswrapper[4824]: I0121 11:25:34.485401 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7c01cf79-8a28-4229-9257-90124e66d3cc-operator-scripts\") pod \"keystone-db-create-jbk7q\" (UID: \"7c01cf79-8a28-4229-9257-90124e66d3cc\") " pod="openstack/keystone-db-create-jbk7q" Jan 21 11:25:34 crc kubenswrapper[4824]: I0121 11:25:34.499666 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hcw6h\" (UniqueName: \"kubernetes.io/projected/7c01cf79-8a28-4229-9257-90124e66d3cc-kube-api-access-hcw6h\") pod \"keystone-db-create-jbk7q\" (UID: \"7c01cf79-8a28-4229-9257-90124e66d3cc\") " pod="openstack/keystone-db-create-jbk7q" Jan 21 11:25:34 crc kubenswrapper[4824]: I0121 11:25:34.521130 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-q7wdg"] Jan 21 11:25:34 crc kubenswrapper[4824]: I0121 11:25:34.522211 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-q7wdg" Jan 21 11:25:34 crc kubenswrapper[4824]: I0121 11:25:34.527058 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-q7wdg"] Jan 21 11:25:34 crc kubenswrapper[4824]: I0121 11:25:34.531835 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-a0fd-account-create-update-zxjvg"] Jan 21 11:25:34 crc kubenswrapper[4824]: I0121 11:25:34.533174 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-a0fd-account-create-update-zxjvg" Jan 21 11:25:34 crc kubenswrapper[4824]: I0121 11:25:34.534372 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Jan 21 11:25:34 crc kubenswrapper[4824]: I0121 11:25:34.541032 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-a0fd-account-create-update-zxjvg"] Jan 21 11:25:34 crc kubenswrapper[4824]: I0121 11:25:34.583988 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-jbk7q" Jan 21 11:25:34 crc kubenswrapper[4824]: I0121 11:25:34.586232 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bc853ab0-e92a-4ada-84fd-b59ac2af573d-operator-scripts\") pod \"keystone-383b-account-create-update-w99bv\" (UID: \"bc853ab0-e92a-4ada-84fd-b59ac2af573d\") " pod="openstack/keystone-383b-account-create-update-w99bv" Jan 21 11:25:34 crc kubenswrapper[4824]: I0121 11:25:34.586334 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t2qxk\" (UniqueName: \"kubernetes.io/projected/bc853ab0-e92a-4ada-84fd-b59ac2af573d-kube-api-access-t2qxk\") pod \"keystone-383b-account-create-update-w99bv\" (UID: \"bc853ab0-e92a-4ada-84fd-b59ac2af573d\") " pod="openstack/keystone-383b-account-create-update-w99bv" Jan 21 11:25:34 crc kubenswrapper[4824]: I0121 11:25:34.587127 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bc853ab0-e92a-4ada-84fd-b59ac2af573d-operator-scripts\") pod \"keystone-383b-account-create-update-w99bv\" (UID: \"bc853ab0-e92a-4ada-84fd-b59ac2af573d\") " pod="openstack/keystone-383b-account-create-update-w99bv" Jan 21 11:25:34 crc kubenswrapper[4824]: I0121 11:25:34.599554 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t2qxk\" (UniqueName: \"kubernetes.io/projected/bc853ab0-e92a-4ada-84fd-b59ac2af573d-kube-api-access-t2qxk\") pod \"keystone-383b-account-create-update-w99bv\" (UID: \"bc853ab0-e92a-4ada-84fd-b59ac2af573d\") " pod="openstack/keystone-383b-account-create-update-w99bv" Jan 21 11:25:34 crc kubenswrapper[4824]: I0121 11:25:34.636430 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-383b-account-create-update-w99bv" Jan 21 11:25:34 crc kubenswrapper[4824]: I0121 11:25:34.688376 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a0ca5b98-4b63-4b4e-a80f-d52977ae3d01-operator-scripts\") pod \"placement-a0fd-account-create-update-zxjvg\" (UID: \"a0ca5b98-4b63-4b4e-a80f-d52977ae3d01\") " pod="openstack/placement-a0fd-account-create-update-zxjvg" Jan 21 11:25:34 crc kubenswrapper[4824]: I0121 11:25:34.690159 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2z72x\" (UniqueName: \"kubernetes.io/projected/a0ca5b98-4b63-4b4e-a80f-d52977ae3d01-kube-api-access-2z72x\") pod \"placement-a0fd-account-create-update-zxjvg\" (UID: \"a0ca5b98-4b63-4b4e-a80f-d52977ae3d01\") " pod="openstack/placement-a0fd-account-create-update-zxjvg" Jan 21 11:25:34 crc kubenswrapper[4824]: I0121 11:25:34.690244 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/885db96c-779f-47c5-a383-1821cb53977c-operator-scripts\") pod \"placement-db-create-q7wdg\" (UID: \"885db96c-779f-47c5-a383-1821cb53977c\") " pod="openstack/placement-db-create-q7wdg" Jan 21 11:25:34 crc kubenswrapper[4824]: I0121 11:25:34.690270 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cvg8j\" (UniqueName: \"kubernetes.io/projected/885db96c-779f-47c5-a383-1821cb53977c-kube-api-access-cvg8j\") pod \"placement-db-create-q7wdg\" (UID: \"885db96c-779f-47c5-a383-1821cb53977c\") " pod="openstack/placement-db-create-q7wdg" Jan 21 11:25:34 crc kubenswrapper[4824]: I0121 11:25:34.738474 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Jan 21 11:25:34 crc kubenswrapper[4824]: I0121 11:25:34.791943 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2z72x\" (UniqueName: \"kubernetes.io/projected/a0ca5b98-4b63-4b4e-a80f-d52977ae3d01-kube-api-access-2z72x\") pod \"placement-a0fd-account-create-update-zxjvg\" (UID: \"a0ca5b98-4b63-4b4e-a80f-d52977ae3d01\") " pod="openstack/placement-a0fd-account-create-update-zxjvg" Jan 21 11:25:34 crc kubenswrapper[4824]: I0121 11:25:34.792041 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/885db96c-779f-47c5-a383-1821cb53977c-operator-scripts\") pod \"placement-db-create-q7wdg\" (UID: \"885db96c-779f-47c5-a383-1821cb53977c\") " pod="openstack/placement-db-create-q7wdg" Jan 21 11:25:34 crc kubenswrapper[4824]: I0121 11:25:34.792059 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cvg8j\" (UniqueName: \"kubernetes.io/projected/885db96c-779f-47c5-a383-1821cb53977c-kube-api-access-cvg8j\") pod \"placement-db-create-q7wdg\" (UID: \"885db96c-779f-47c5-a383-1821cb53977c\") " pod="openstack/placement-db-create-q7wdg" Jan 21 11:25:34 crc kubenswrapper[4824]: I0121 11:25:34.792135 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a0ca5b98-4b63-4b4e-a80f-d52977ae3d01-operator-scripts\") pod \"placement-a0fd-account-create-update-zxjvg\" (UID: \"a0ca5b98-4b63-4b4e-a80f-d52977ae3d01\") " pod="openstack/placement-a0fd-account-create-update-zxjvg" Jan 21 11:25:34 crc kubenswrapper[4824]: I0121 11:25:34.794491 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/885db96c-779f-47c5-a383-1821cb53977c-operator-scripts\") pod \"placement-db-create-q7wdg\" (UID: \"885db96c-779f-47c5-a383-1821cb53977c\") " pod="openstack/placement-db-create-q7wdg" Jan 21 11:25:34 crc kubenswrapper[4824]: I0121 11:25:34.794501 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a0ca5b98-4b63-4b4e-a80f-d52977ae3d01-operator-scripts\") pod \"placement-a0fd-account-create-update-zxjvg\" (UID: \"a0ca5b98-4b63-4b4e-a80f-d52977ae3d01\") " pod="openstack/placement-a0fd-account-create-update-zxjvg" Jan 21 11:25:34 crc kubenswrapper[4824]: I0121 11:25:34.807274 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cvg8j\" (UniqueName: \"kubernetes.io/projected/885db96c-779f-47c5-a383-1821cb53977c-kube-api-access-cvg8j\") pod \"placement-db-create-q7wdg\" (UID: \"885db96c-779f-47c5-a383-1821cb53977c\") " pod="openstack/placement-db-create-q7wdg" Jan 21 11:25:34 crc kubenswrapper[4824]: I0121 11:25:34.807505 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2z72x\" (UniqueName: \"kubernetes.io/projected/a0ca5b98-4b63-4b4e-a80f-d52977ae3d01-kube-api-access-2z72x\") pod \"placement-a0fd-account-create-update-zxjvg\" (UID: \"a0ca5b98-4b63-4b4e-a80f-d52977ae3d01\") " pod="openstack/placement-a0fd-account-create-update-zxjvg" Jan 21 11:25:34 crc kubenswrapper[4824]: I0121 11:25:34.848798 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-q7wdg" Jan 21 11:25:34 crc kubenswrapper[4824]: I0121 11:25:34.857212 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-a0fd-account-create-update-zxjvg" Jan 21 11:25:34 crc kubenswrapper[4824]: I0121 11:25:34.947425 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-jbk7q"] Jan 21 11:25:34 crc kubenswrapper[4824]: W0121 11:25:34.948534 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7c01cf79_8a28_4229_9257_90124e66d3cc.slice/crio-1e477713ddb0f2a67fe723732b272384c4ff5196b703dde89021649abdfee9d8 WatchSource:0}: Error finding container 1e477713ddb0f2a67fe723732b272384c4ff5196b703dde89021649abdfee9d8: Status 404 returned error can't find the container with id 1e477713ddb0f2a67fe723732b272384c4ff5196b703dde89021649abdfee9d8 Jan 21 11:25:35 crc kubenswrapper[4824]: I0121 11:25:35.035710 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-383b-account-create-update-w99bv"] Jan 21 11:25:35 crc kubenswrapper[4824]: W0121 11:25:35.042775 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbc853ab0_e92a_4ada_84fd_b59ac2af573d.slice/crio-43eefd9eea7557b24c2304803f43b0c074dd6eee119b0fe3f3750b2f784da5c5 WatchSource:0}: Error finding container 43eefd9eea7557b24c2304803f43b0c074dd6eee119b0fe3f3750b2f784da5c5: Status 404 returned error can't find the container with id 43eefd9eea7557b24c2304803f43b0c074dd6eee119b0fe3f3750b2f784da5c5 Jan 21 11:25:35 crc kubenswrapper[4824]: I0121 11:25:35.222634 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-a0fd-account-create-update-zxjvg"] Jan 21 11:25:35 crc kubenswrapper[4824]: W0121 11:25:35.230574 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda0ca5b98_4b63_4b4e_a80f_d52977ae3d01.slice/crio-a17a042f056b3ff29d8f31835c21afba176775ec0e3832d56d1636a0f187eb66 WatchSource:0}: Error finding container a17a042f056b3ff29d8f31835c21afba176775ec0e3832d56d1636a0f187eb66: Status 404 returned error can't find the container with id a17a042f056b3ff29d8f31835c21afba176775ec0e3832d56d1636a0f187eb66 Jan 21 11:25:35 crc kubenswrapper[4824]: I0121 11:25:35.276939 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-q7wdg"] Jan 21 11:25:35 crc kubenswrapper[4824]: W0121 11:25:35.280745 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod885db96c_779f_47c5_a383_1821cb53977c.slice/crio-a7ec96b60c6cf63256dcb9a6367e913091482fe19b5f87ad070d29ec3152485f WatchSource:0}: Error finding container a7ec96b60c6cf63256dcb9a6367e913091482fe19b5f87ad070d29ec3152485f: Status 404 returned error can't find the container with id a7ec96b60c6cf63256dcb9a6367e913091482fe19b5f87ad070d29ec3152485f Jan 21 11:25:35 crc kubenswrapper[4824]: I0121 11:25:35.660715 4824 generic.go:334] "Generic (PLEG): container finished" podID="bc853ab0-e92a-4ada-84fd-b59ac2af573d" containerID="0811ac195531b85d7ce2dbbbbfe5e4b4eec9c6e973dbfd91e9e942aecc0be98b" exitCode=0 Jan 21 11:25:35 crc kubenswrapper[4824]: I0121 11:25:35.660769 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-383b-account-create-update-w99bv" event={"ID":"bc853ab0-e92a-4ada-84fd-b59ac2af573d","Type":"ContainerDied","Data":"0811ac195531b85d7ce2dbbbbfe5e4b4eec9c6e973dbfd91e9e942aecc0be98b"} Jan 21 11:25:35 crc kubenswrapper[4824]: I0121 11:25:35.660791 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-383b-account-create-update-w99bv" event={"ID":"bc853ab0-e92a-4ada-84fd-b59ac2af573d","Type":"ContainerStarted","Data":"43eefd9eea7557b24c2304803f43b0c074dd6eee119b0fe3f3750b2f784da5c5"} Jan 21 11:25:35 crc kubenswrapper[4824]: I0121 11:25:35.662260 4824 generic.go:334] "Generic (PLEG): container finished" podID="a0ca5b98-4b63-4b4e-a80f-d52977ae3d01" containerID="be822998cfd0d1fe00b165b7bdf2ce2c669d12e3704320cfa573970aa648494d" exitCode=0 Jan 21 11:25:35 crc kubenswrapper[4824]: I0121 11:25:35.662296 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-a0fd-account-create-update-zxjvg" event={"ID":"a0ca5b98-4b63-4b4e-a80f-d52977ae3d01","Type":"ContainerDied","Data":"be822998cfd0d1fe00b165b7bdf2ce2c669d12e3704320cfa573970aa648494d"} Jan 21 11:25:35 crc kubenswrapper[4824]: I0121 11:25:35.662312 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-a0fd-account-create-update-zxjvg" event={"ID":"a0ca5b98-4b63-4b4e-a80f-d52977ae3d01","Type":"ContainerStarted","Data":"a17a042f056b3ff29d8f31835c21afba176775ec0e3832d56d1636a0f187eb66"} Jan 21 11:25:35 crc kubenswrapper[4824]: I0121 11:25:35.665872 4824 generic.go:334] "Generic (PLEG): container finished" podID="7c01cf79-8a28-4229-9257-90124e66d3cc" containerID="e85df67ff5e42c058f160ae767051c544cbf1f8e15561d3884068b5517d68c3a" exitCode=0 Jan 21 11:25:35 crc kubenswrapper[4824]: I0121 11:25:35.665914 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-jbk7q" event={"ID":"7c01cf79-8a28-4229-9257-90124e66d3cc","Type":"ContainerDied","Data":"e85df67ff5e42c058f160ae767051c544cbf1f8e15561d3884068b5517d68c3a"} Jan 21 11:25:35 crc kubenswrapper[4824]: I0121 11:25:35.665930 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-jbk7q" event={"ID":"7c01cf79-8a28-4229-9257-90124e66d3cc","Type":"ContainerStarted","Data":"1e477713ddb0f2a67fe723732b272384c4ff5196b703dde89021649abdfee9d8"} Jan 21 11:25:35 crc kubenswrapper[4824]: I0121 11:25:35.667174 4824 generic.go:334] "Generic (PLEG): container finished" podID="885db96c-779f-47c5-a383-1821cb53977c" containerID="7d54dd5634955be1b99e4810715039653beb466487ffbc563c911444570d3daa" exitCode=0 Jan 21 11:25:35 crc kubenswrapper[4824]: I0121 11:25:35.669779 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-q7wdg" event={"ID":"885db96c-779f-47c5-a383-1821cb53977c","Type":"ContainerDied","Data":"7d54dd5634955be1b99e4810715039653beb466487ffbc563c911444570d3daa"} Jan 21 11:25:35 crc kubenswrapper[4824]: I0121 11:25:35.669820 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-q7wdg" event={"ID":"885db96c-779f-47c5-a383-1821cb53977c","Type":"ContainerStarted","Data":"a7ec96b60c6cf63256dcb9a6367e913091482fe19b5f87ad070d29ec3152485f"} Jan 21 11:25:35 crc kubenswrapper[4824]: I0121 11:25:35.676693 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Jan 21 11:25:35 crc kubenswrapper[4824]: I0121 11:25:35.851398 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Jan 21 11:25:35 crc kubenswrapper[4824]: I0121 11:25:35.852723 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Jan 21 11:25:35 crc kubenswrapper[4824]: I0121 11:25:35.855005 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Jan 21 11:25:35 crc kubenswrapper[4824]: I0121 11:25:35.855176 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Jan 21 11:25:35 crc kubenswrapper[4824]: I0121 11:25:35.855304 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-qc9wz" Jan 21 11:25:35 crc kubenswrapper[4824]: I0121 11:25:35.858047 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Jan 21 11:25:35 crc kubenswrapper[4824]: I0121 11:25:35.874570 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Jan 21 11:25:36 crc kubenswrapper[4824]: I0121 11:25:36.012513 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/8b9c25b1-d0b9-430e-88cf-0aeaf6785461-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"8b9c25b1-d0b9-430e-88cf-0aeaf6785461\") " pod="openstack/ovn-northd-0" Jan 21 11:25:36 crc kubenswrapper[4824]: I0121 11:25:36.012590 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8b9c25b1-d0b9-430e-88cf-0aeaf6785461-config\") pod \"ovn-northd-0\" (UID: \"8b9c25b1-d0b9-430e-88cf-0aeaf6785461\") " pod="openstack/ovn-northd-0" Jan 21 11:25:36 crc kubenswrapper[4824]: I0121 11:25:36.012676 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8b9c25b1-d0b9-430e-88cf-0aeaf6785461-scripts\") pod \"ovn-northd-0\" (UID: \"8b9c25b1-d0b9-430e-88cf-0aeaf6785461\") " pod="openstack/ovn-northd-0" Jan 21 11:25:36 crc kubenswrapper[4824]: I0121 11:25:36.012832 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/8b9c25b1-d0b9-430e-88cf-0aeaf6785461-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"8b9c25b1-d0b9-430e-88cf-0aeaf6785461\") " pod="openstack/ovn-northd-0" Jan 21 11:25:36 crc kubenswrapper[4824]: I0121 11:25:36.012916 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-27w66\" (UniqueName: \"kubernetes.io/projected/8b9c25b1-d0b9-430e-88cf-0aeaf6785461-kube-api-access-27w66\") pod \"ovn-northd-0\" (UID: \"8b9c25b1-d0b9-430e-88cf-0aeaf6785461\") " pod="openstack/ovn-northd-0" Jan 21 11:25:36 crc kubenswrapper[4824]: I0121 11:25:36.013064 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8b9c25b1-d0b9-430e-88cf-0aeaf6785461-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"8b9c25b1-d0b9-430e-88cf-0aeaf6785461\") " pod="openstack/ovn-northd-0" Jan 21 11:25:36 crc kubenswrapper[4824]: I0121 11:25:36.013084 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/8b9c25b1-d0b9-430e-88cf-0aeaf6785461-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"8b9c25b1-d0b9-430e-88cf-0aeaf6785461\") " pod="openstack/ovn-northd-0" Jan 21 11:25:36 crc kubenswrapper[4824]: I0121 11:25:36.113851 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-27w66\" (UniqueName: \"kubernetes.io/projected/8b9c25b1-d0b9-430e-88cf-0aeaf6785461-kube-api-access-27w66\") pod \"ovn-northd-0\" (UID: \"8b9c25b1-d0b9-430e-88cf-0aeaf6785461\") " pod="openstack/ovn-northd-0" Jan 21 11:25:36 crc kubenswrapper[4824]: I0121 11:25:36.113942 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8b9c25b1-d0b9-430e-88cf-0aeaf6785461-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"8b9c25b1-d0b9-430e-88cf-0aeaf6785461\") " pod="openstack/ovn-northd-0" Jan 21 11:25:36 crc kubenswrapper[4824]: I0121 11:25:36.113974 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/8b9c25b1-d0b9-430e-88cf-0aeaf6785461-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"8b9c25b1-d0b9-430e-88cf-0aeaf6785461\") " pod="openstack/ovn-northd-0" Jan 21 11:25:36 crc kubenswrapper[4824]: I0121 11:25:36.114002 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/8b9c25b1-d0b9-430e-88cf-0aeaf6785461-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"8b9c25b1-d0b9-430e-88cf-0aeaf6785461\") " pod="openstack/ovn-northd-0" Jan 21 11:25:36 crc kubenswrapper[4824]: I0121 11:25:36.114054 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8b9c25b1-d0b9-430e-88cf-0aeaf6785461-config\") pod \"ovn-northd-0\" (UID: \"8b9c25b1-d0b9-430e-88cf-0aeaf6785461\") " pod="openstack/ovn-northd-0" Jan 21 11:25:36 crc kubenswrapper[4824]: I0121 11:25:36.114068 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8b9c25b1-d0b9-430e-88cf-0aeaf6785461-scripts\") pod \"ovn-northd-0\" (UID: \"8b9c25b1-d0b9-430e-88cf-0aeaf6785461\") " pod="openstack/ovn-northd-0" Jan 21 11:25:36 crc kubenswrapper[4824]: I0121 11:25:36.114117 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/8b9c25b1-d0b9-430e-88cf-0aeaf6785461-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"8b9c25b1-d0b9-430e-88cf-0aeaf6785461\") " pod="openstack/ovn-northd-0" Jan 21 11:25:36 crc kubenswrapper[4824]: I0121 11:25:36.114523 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/8b9c25b1-d0b9-430e-88cf-0aeaf6785461-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"8b9c25b1-d0b9-430e-88cf-0aeaf6785461\") " pod="openstack/ovn-northd-0" Jan 21 11:25:36 crc kubenswrapper[4824]: I0121 11:25:36.115975 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8b9c25b1-d0b9-430e-88cf-0aeaf6785461-config\") pod \"ovn-northd-0\" (UID: \"8b9c25b1-d0b9-430e-88cf-0aeaf6785461\") " pod="openstack/ovn-northd-0" Jan 21 11:25:36 crc kubenswrapper[4824]: I0121 11:25:36.116489 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8b9c25b1-d0b9-430e-88cf-0aeaf6785461-scripts\") pod \"ovn-northd-0\" (UID: \"8b9c25b1-d0b9-430e-88cf-0aeaf6785461\") " pod="openstack/ovn-northd-0" Jan 21 11:25:36 crc kubenswrapper[4824]: I0121 11:25:36.119057 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8b9c25b1-d0b9-430e-88cf-0aeaf6785461-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"8b9c25b1-d0b9-430e-88cf-0aeaf6785461\") " pod="openstack/ovn-northd-0" Jan 21 11:25:36 crc kubenswrapper[4824]: I0121 11:25:36.121442 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/8b9c25b1-d0b9-430e-88cf-0aeaf6785461-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"8b9c25b1-d0b9-430e-88cf-0aeaf6785461\") " pod="openstack/ovn-northd-0" Jan 21 11:25:36 crc kubenswrapper[4824]: I0121 11:25:36.127800 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-27w66\" (UniqueName: \"kubernetes.io/projected/8b9c25b1-d0b9-430e-88cf-0aeaf6785461-kube-api-access-27w66\") pod \"ovn-northd-0\" (UID: \"8b9c25b1-d0b9-430e-88cf-0aeaf6785461\") " pod="openstack/ovn-northd-0" Jan 21 11:25:36 crc kubenswrapper[4824]: I0121 11:25:36.134897 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/8b9c25b1-d0b9-430e-88cf-0aeaf6785461-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"8b9c25b1-d0b9-430e-88cf-0aeaf6785461\") " pod="openstack/ovn-northd-0" Jan 21 11:25:36 crc kubenswrapper[4824]: I0121 11:25:36.164041 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Jan 21 11:25:36 crc kubenswrapper[4824]: I0121 11:25:36.328429 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-vfbzh"] Jan 21 11:25:36 crc kubenswrapper[4824]: I0121 11:25:36.328823 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-86db49b7ff-vfbzh" podUID="292b341f-0f9e-411e-9f85-47bde4dcb2c4" containerName="dnsmasq-dns" containerID="cri-o://095e6c8fd89e1deec2822bbad9ce47d3bc5541769b4bdf66512d4544bcb46405" gracePeriod=10 Jan 21 11:25:36 crc kubenswrapper[4824]: I0121 11:25:36.333675 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-86db49b7ff-vfbzh" Jan 21 11:25:36 crc kubenswrapper[4824]: I0121 11:25:36.359085 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-698758b865-qzfrf"] Jan 21 11:25:36 crc kubenswrapper[4824]: I0121 11:25:36.374381 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-698758b865-qzfrf"] Jan 21 11:25:36 crc kubenswrapper[4824]: I0121 11:25:36.374468 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-qzfrf" Jan 21 11:25:36 crc kubenswrapper[4824]: I0121 11:25:36.432785 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Jan 21 11:25:36 crc kubenswrapper[4824]: I0121 11:25:36.526063 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e94778b1-d36a-47e9-b201-9b7cd1cdb640-dns-svc\") pod \"dnsmasq-dns-698758b865-qzfrf\" (UID: \"e94778b1-d36a-47e9-b201-9b7cd1cdb640\") " pod="openstack/dnsmasq-dns-698758b865-qzfrf" Jan 21 11:25:36 crc kubenswrapper[4824]: I0121 11:25:36.526131 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q9jhb\" (UniqueName: \"kubernetes.io/projected/e94778b1-d36a-47e9-b201-9b7cd1cdb640-kube-api-access-q9jhb\") pod \"dnsmasq-dns-698758b865-qzfrf\" (UID: \"e94778b1-d36a-47e9-b201-9b7cd1cdb640\") " pod="openstack/dnsmasq-dns-698758b865-qzfrf" Jan 21 11:25:36 crc kubenswrapper[4824]: I0121 11:25:36.526242 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e94778b1-d36a-47e9-b201-9b7cd1cdb640-config\") pod \"dnsmasq-dns-698758b865-qzfrf\" (UID: \"e94778b1-d36a-47e9-b201-9b7cd1cdb640\") " pod="openstack/dnsmasq-dns-698758b865-qzfrf" Jan 21 11:25:36 crc kubenswrapper[4824]: I0121 11:25:36.526279 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e94778b1-d36a-47e9-b201-9b7cd1cdb640-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-qzfrf\" (UID: \"e94778b1-d36a-47e9-b201-9b7cd1cdb640\") " pod="openstack/dnsmasq-dns-698758b865-qzfrf" Jan 21 11:25:36 crc kubenswrapper[4824]: I0121 11:25:36.526319 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e94778b1-d36a-47e9-b201-9b7cd1cdb640-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-qzfrf\" (UID: \"e94778b1-d36a-47e9-b201-9b7cd1cdb640\") " pod="openstack/dnsmasq-dns-698758b865-qzfrf" Jan 21 11:25:36 crc kubenswrapper[4824]: I0121 11:25:36.606218 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Jan 21 11:25:36 crc kubenswrapper[4824]: I0121 11:25:36.627658 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e94778b1-d36a-47e9-b201-9b7cd1cdb640-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-qzfrf\" (UID: \"e94778b1-d36a-47e9-b201-9b7cd1cdb640\") " pod="openstack/dnsmasq-dns-698758b865-qzfrf" Jan 21 11:25:36 crc kubenswrapper[4824]: I0121 11:25:36.627772 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e94778b1-d36a-47e9-b201-9b7cd1cdb640-dns-svc\") pod \"dnsmasq-dns-698758b865-qzfrf\" (UID: \"e94778b1-d36a-47e9-b201-9b7cd1cdb640\") " pod="openstack/dnsmasq-dns-698758b865-qzfrf" Jan 21 11:25:36 crc kubenswrapper[4824]: I0121 11:25:36.627808 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q9jhb\" (UniqueName: \"kubernetes.io/projected/e94778b1-d36a-47e9-b201-9b7cd1cdb640-kube-api-access-q9jhb\") pod \"dnsmasq-dns-698758b865-qzfrf\" (UID: \"e94778b1-d36a-47e9-b201-9b7cd1cdb640\") " pod="openstack/dnsmasq-dns-698758b865-qzfrf" Jan 21 11:25:36 crc kubenswrapper[4824]: I0121 11:25:36.627857 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e94778b1-d36a-47e9-b201-9b7cd1cdb640-config\") pod \"dnsmasq-dns-698758b865-qzfrf\" (UID: \"e94778b1-d36a-47e9-b201-9b7cd1cdb640\") " pod="openstack/dnsmasq-dns-698758b865-qzfrf" Jan 21 11:25:36 crc kubenswrapper[4824]: I0121 11:25:36.627886 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e94778b1-d36a-47e9-b201-9b7cd1cdb640-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-qzfrf\" (UID: \"e94778b1-d36a-47e9-b201-9b7cd1cdb640\") " pod="openstack/dnsmasq-dns-698758b865-qzfrf" Jan 21 11:25:36 crc kubenswrapper[4824]: I0121 11:25:36.628601 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e94778b1-d36a-47e9-b201-9b7cd1cdb640-dns-svc\") pod \"dnsmasq-dns-698758b865-qzfrf\" (UID: \"e94778b1-d36a-47e9-b201-9b7cd1cdb640\") " pod="openstack/dnsmasq-dns-698758b865-qzfrf" Jan 21 11:25:36 crc kubenswrapper[4824]: I0121 11:25:36.629121 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e94778b1-d36a-47e9-b201-9b7cd1cdb640-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-qzfrf\" (UID: \"e94778b1-d36a-47e9-b201-9b7cd1cdb640\") " pod="openstack/dnsmasq-dns-698758b865-qzfrf" Jan 21 11:25:36 crc kubenswrapper[4824]: I0121 11:25:36.629183 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e94778b1-d36a-47e9-b201-9b7cd1cdb640-config\") pod \"dnsmasq-dns-698758b865-qzfrf\" (UID: \"e94778b1-d36a-47e9-b201-9b7cd1cdb640\") " pod="openstack/dnsmasq-dns-698758b865-qzfrf" Jan 21 11:25:36 crc kubenswrapper[4824]: I0121 11:25:36.629536 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e94778b1-d36a-47e9-b201-9b7cd1cdb640-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-qzfrf\" (UID: \"e94778b1-d36a-47e9-b201-9b7cd1cdb640\") " pod="openstack/dnsmasq-dns-698758b865-qzfrf" Jan 21 11:25:36 crc kubenswrapper[4824]: I0121 11:25:36.641860 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q9jhb\" (UniqueName: \"kubernetes.io/projected/e94778b1-d36a-47e9-b201-9b7cd1cdb640-kube-api-access-q9jhb\") pod \"dnsmasq-dns-698758b865-qzfrf\" (UID: \"e94778b1-d36a-47e9-b201-9b7cd1cdb640\") " pod="openstack/dnsmasq-dns-698758b865-qzfrf" Jan 21 11:25:36 crc kubenswrapper[4824]: I0121 11:25:36.678333 4824 generic.go:334] "Generic (PLEG): container finished" podID="292b341f-0f9e-411e-9f85-47bde4dcb2c4" containerID="095e6c8fd89e1deec2822bbad9ce47d3bc5541769b4bdf66512d4544bcb46405" exitCode=0 Jan 21 11:25:36 crc kubenswrapper[4824]: I0121 11:25:36.678409 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-vfbzh" event={"ID":"292b341f-0f9e-411e-9f85-47bde4dcb2c4","Type":"ContainerDied","Data":"095e6c8fd89e1deec2822bbad9ce47d3bc5541769b4bdf66512d4544bcb46405"} Jan 21 11:25:36 crc kubenswrapper[4824]: I0121 11:25:36.679432 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"8b9c25b1-d0b9-430e-88cf-0aeaf6785461","Type":"ContainerStarted","Data":"9247690bf5598a07e7dc894f7f927cbb898894b6d00fd4477369b84588389a75"} Jan 21 11:25:36 crc kubenswrapper[4824]: I0121 11:25:36.694611 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-qzfrf" Jan 21 11:25:36 crc kubenswrapper[4824]: I0121 11:25:36.987202 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-383b-account-create-update-w99bv" Jan 21 11:25:37 crc kubenswrapper[4824]: I0121 11:25:37.135380 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bc853ab0-e92a-4ada-84fd-b59ac2af573d-operator-scripts\") pod \"bc853ab0-e92a-4ada-84fd-b59ac2af573d\" (UID: \"bc853ab0-e92a-4ada-84fd-b59ac2af573d\") " Jan 21 11:25:37 crc kubenswrapper[4824]: I0121 11:25:37.135663 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t2qxk\" (UniqueName: \"kubernetes.io/projected/bc853ab0-e92a-4ada-84fd-b59ac2af573d-kube-api-access-t2qxk\") pod \"bc853ab0-e92a-4ada-84fd-b59ac2af573d\" (UID: \"bc853ab0-e92a-4ada-84fd-b59ac2af573d\") " Jan 21 11:25:37 crc kubenswrapper[4824]: I0121 11:25:37.136088 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bc853ab0-e92a-4ada-84fd-b59ac2af573d-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "bc853ab0-e92a-4ada-84fd-b59ac2af573d" (UID: "bc853ab0-e92a-4ada-84fd-b59ac2af573d"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:25:37 crc kubenswrapper[4824]: I0121 11:25:37.136212 4824 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bc853ab0-e92a-4ada-84fd-b59ac2af573d-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 21 11:25:37 crc kubenswrapper[4824]: I0121 11:25:37.140612 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc853ab0-e92a-4ada-84fd-b59ac2af573d-kube-api-access-t2qxk" (OuterVolumeSpecName: "kube-api-access-t2qxk") pod "bc853ab0-e92a-4ada-84fd-b59ac2af573d" (UID: "bc853ab0-e92a-4ada-84fd-b59ac2af573d"). InnerVolumeSpecName "kube-api-access-t2qxk". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:25:37 crc kubenswrapper[4824]: I0121 11:25:37.167709 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-a0fd-account-create-update-zxjvg" Jan 21 11:25:37 crc kubenswrapper[4824]: I0121 11:25:37.181205 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-jbk7q" Jan 21 11:25:37 crc kubenswrapper[4824]: I0121 11:25:37.187931 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-q7wdg" Jan 21 11:25:37 crc kubenswrapper[4824]: I0121 11:25:37.237402 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t2qxk\" (UniqueName: \"kubernetes.io/projected/bc853ab0-e92a-4ada-84fd-b59ac2af573d-kube-api-access-t2qxk\") on node \"crc\" DevicePath \"\"" Jan 21 11:25:37 crc kubenswrapper[4824]: I0121 11:25:37.268031 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-698758b865-qzfrf"] Jan 21 11:25:37 crc kubenswrapper[4824]: W0121 11:25:37.269498 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode94778b1_d36a_47e9_b201_9b7cd1cdb640.slice/crio-48fb0631b23867cac173ec7a9df005de8edd1cb0df987f8dceb763f2434dd5b7 WatchSource:0}: Error finding container 48fb0631b23867cac173ec7a9df005de8edd1cb0df987f8dceb763f2434dd5b7: Status 404 returned error can't find the container with id 48fb0631b23867cac173ec7a9df005de8edd1cb0df987f8dceb763f2434dd5b7 Jan 21 11:25:37 crc kubenswrapper[4824]: I0121 11:25:37.338292 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cvg8j\" (UniqueName: \"kubernetes.io/projected/885db96c-779f-47c5-a383-1821cb53977c-kube-api-access-cvg8j\") pod \"885db96c-779f-47c5-a383-1821cb53977c\" (UID: \"885db96c-779f-47c5-a383-1821cb53977c\") " Jan 21 11:25:37 crc kubenswrapper[4824]: I0121 11:25:37.338358 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/885db96c-779f-47c5-a383-1821cb53977c-operator-scripts\") pod \"885db96c-779f-47c5-a383-1821cb53977c\" (UID: \"885db96c-779f-47c5-a383-1821cb53977c\") " Jan 21 11:25:37 crc kubenswrapper[4824]: I0121 11:25:37.338393 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a0ca5b98-4b63-4b4e-a80f-d52977ae3d01-operator-scripts\") pod \"a0ca5b98-4b63-4b4e-a80f-d52977ae3d01\" (UID: \"a0ca5b98-4b63-4b4e-a80f-d52977ae3d01\") " Jan 21 11:25:37 crc kubenswrapper[4824]: I0121 11:25:37.338429 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7c01cf79-8a28-4229-9257-90124e66d3cc-operator-scripts\") pod \"7c01cf79-8a28-4229-9257-90124e66d3cc\" (UID: \"7c01cf79-8a28-4229-9257-90124e66d3cc\") " Jan 21 11:25:37 crc kubenswrapper[4824]: I0121 11:25:37.338503 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2z72x\" (UniqueName: \"kubernetes.io/projected/a0ca5b98-4b63-4b4e-a80f-d52977ae3d01-kube-api-access-2z72x\") pod \"a0ca5b98-4b63-4b4e-a80f-d52977ae3d01\" (UID: \"a0ca5b98-4b63-4b4e-a80f-d52977ae3d01\") " Jan 21 11:25:37 crc kubenswrapper[4824]: I0121 11:25:37.338530 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hcw6h\" (UniqueName: \"kubernetes.io/projected/7c01cf79-8a28-4229-9257-90124e66d3cc-kube-api-access-hcw6h\") pod \"7c01cf79-8a28-4229-9257-90124e66d3cc\" (UID: \"7c01cf79-8a28-4229-9257-90124e66d3cc\") " Jan 21 11:25:37 crc kubenswrapper[4824]: I0121 11:25:37.339140 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a0ca5b98-4b63-4b4e-a80f-d52977ae3d01-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "a0ca5b98-4b63-4b4e-a80f-d52977ae3d01" (UID: "a0ca5b98-4b63-4b4e-a80f-d52977ae3d01"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:25:37 crc kubenswrapper[4824]: I0121 11:25:37.339159 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7c01cf79-8a28-4229-9257-90124e66d3cc-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "7c01cf79-8a28-4229-9257-90124e66d3cc" (UID: "7c01cf79-8a28-4229-9257-90124e66d3cc"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:25:37 crc kubenswrapper[4824]: I0121 11:25:37.339195 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/885db96c-779f-47c5-a383-1821cb53977c-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "885db96c-779f-47c5-a383-1821cb53977c" (UID: "885db96c-779f-47c5-a383-1821cb53977c"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:25:37 crc kubenswrapper[4824]: I0121 11:25:37.341099 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0ca5b98-4b63-4b4e-a80f-d52977ae3d01-kube-api-access-2z72x" (OuterVolumeSpecName: "kube-api-access-2z72x") pod "a0ca5b98-4b63-4b4e-a80f-d52977ae3d01" (UID: "a0ca5b98-4b63-4b4e-a80f-d52977ae3d01"). InnerVolumeSpecName "kube-api-access-2z72x". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:25:37 crc kubenswrapper[4824]: I0121 11:25:37.342394 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7c01cf79-8a28-4229-9257-90124e66d3cc-kube-api-access-hcw6h" (OuterVolumeSpecName: "kube-api-access-hcw6h") pod "7c01cf79-8a28-4229-9257-90124e66d3cc" (UID: "7c01cf79-8a28-4229-9257-90124e66d3cc"). InnerVolumeSpecName "kube-api-access-hcw6h". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:25:37 crc kubenswrapper[4824]: I0121 11:25:37.343359 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/885db96c-779f-47c5-a383-1821cb53977c-kube-api-access-cvg8j" (OuterVolumeSpecName: "kube-api-access-cvg8j") pod "885db96c-779f-47c5-a383-1821cb53977c" (UID: "885db96c-779f-47c5-a383-1821cb53977c"). InnerVolumeSpecName "kube-api-access-cvg8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:25:37 crc kubenswrapper[4824]: I0121 11:25:37.440740 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cvg8j\" (UniqueName: \"kubernetes.io/projected/885db96c-779f-47c5-a383-1821cb53977c-kube-api-access-cvg8j\") on node \"crc\" DevicePath \"\"" Jan 21 11:25:37 crc kubenswrapper[4824]: I0121 11:25:37.440770 4824 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/885db96c-779f-47c5-a383-1821cb53977c-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 21 11:25:37 crc kubenswrapper[4824]: I0121 11:25:37.440779 4824 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a0ca5b98-4b63-4b4e-a80f-d52977ae3d01-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 21 11:25:37 crc kubenswrapper[4824]: I0121 11:25:37.440788 4824 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7c01cf79-8a28-4229-9257-90124e66d3cc-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 21 11:25:37 crc kubenswrapper[4824]: I0121 11:25:37.440801 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2z72x\" (UniqueName: \"kubernetes.io/projected/a0ca5b98-4b63-4b4e-a80f-d52977ae3d01-kube-api-access-2z72x\") on node \"crc\" DevicePath \"\"" Jan 21 11:25:37 crc kubenswrapper[4824]: I0121 11:25:37.440808 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hcw6h\" (UniqueName: \"kubernetes.io/projected/7c01cf79-8a28-4229-9257-90124e66d3cc-kube-api-access-hcw6h\") on node \"crc\" DevicePath \"\"" Jan 21 11:25:37 crc kubenswrapper[4824]: I0121 11:25:37.510585 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-storage-0"] Jan 21 11:25:37 crc kubenswrapper[4824]: E0121 11:25:37.510879 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="885db96c-779f-47c5-a383-1821cb53977c" containerName="mariadb-database-create" Jan 21 11:25:37 crc kubenswrapper[4824]: I0121 11:25:37.510896 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="885db96c-779f-47c5-a383-1821cb53977c" containerName="mariadb-database-create" Jan 21 11:25:37 crc kubenswrapper[4824]: E0121 11:25:37.510918 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c01cf79-8a28-4229-9257-90124e66d3cc" containerName="mariadb-database-create" Jan 21 11:25:37 crc kubenswrapper[4824]: I0121 11:25:37.510924 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c01cf79-8a28-4229-9257-90124e66d3cc" containerName="mariadb-database-create" Jan 21 11:25:37 crc kubenswrapper[4824]: E0121 11:25:37.510952 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a0ca5b98-4b63-4b4e-a80f-d52977ae3d01" containerName="mariadb-account-create-update" Jan 21 11:25:37 crc kubenswrapper[4824]: I0121 11:25:37.510972 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="a0ca5b98-4b63-4b4e-a80f-d52977ae3d01" containerName="mariadb-account-create-update" Jan 21 11:25:37 crc kubenswrapper[4824]: E0121 11:25:37.510987 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc853ab0-e92a-4ada-84fd-b59ac2af573d" containerName="mariadb-account-create-update" Jan 21 11:25:37 crc kubenswrapper[4824]: I0121 11:25:37.510992 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc853ab0-e92a-4ada-84fd-b59ac2af573d" containerName="mariadb-account-create-update" Jan 21 11:25:37 crc kubenswrapper[4824]: I0121 11:25:37.511124 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="7c01cf79-8a28-4229-9257-90124e66d3cc" containerName="mariadb-database-create" Jan 21 11:25:37 crc kubenswrapper[4824]: I0121 11:25:37.511143 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="bc853ab0-e92a-4ada-84fd-b59ac2af573d" containerName="mariadb-account-create-update" Jan 21 11:25:37 crc kubenswrapper[4824]: I0121 11:25:37.511151 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="885db96c-779f-47c5-a383-1821cb53977c" containerName="mariadb-database-create" Jan 21 11:25:37 crc kubenswrapper[4824]: I0121 11:25:37.511158 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="a0ca5b98-4b63-4b4e-a80f-d52977ae3d01" containerName="mariadb-account-create-update" Jan 21 11:25:37 crc kubenswrapper[4824]: I0121 11:25:37.514744 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Jan 21 11:25:37 crc kubenswrapper[4824]: I0121 11:25:37.516392 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-lk8t8" Jan 21 11:25:37 crc kubenswrapper[4824]: I0121 11:25:37.517136 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Jan 21 11:25:37 crc kubenswrapper[4824]: I0121 11:25:37.517144 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Jan 21 11:25:37 crc kubenswrapper[4824]: I0121 11:25:37.517218 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-storage-config-data" Jan 21 11:25:37 crc kubenswrapper[4824]: I0121 11:25:37.528089 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Jan 21 11:25:37 crc kubenswrapper[4824]: I0121 11:25:37.643139 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/f980e493-ca74-47f1-8f6f-aca9c9011bf5-lock\") pod \"swift-storage-0\" (UID: \"f980e493-ca74-47f1-8f6f-aca9c9011bf5\") " pod="openstack/swift-storage-0" Jan 21 11:25:37 crc kubenswrapper[4824]: I0121 11:25:37.643180 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/f980e493-ca74-47f1-8f6f-aca9c9011bf5-cache\") pod \"swift-storage-0\" (UID: \"f980e493-ca74-47f1-8f6f-aca9c9011bf5\") " pod="openstack/swift-storage-0" Jan 21 11:25:37 crc kubenswrapper[4824]: I0121 11:25:37.643197 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5hwq7\" (UniqueName: \"kubernetes.io/projected/f980e493-ca74-47f1-8f6f-aca9c9011bf5-kube-api-access-5hwq7\") pod \"swift-storage-0\" (UID: \"f980e493-ca74-47f1-8f6f-aca9c9011bf5\") " pod="openstack/swift-storage-0" Jan 21 11:25:37 crc kubenswrapper[4824]: I0121 11:25:37.643226 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"swift-storage-0\" (UID: \"f980e493-ca74-47f1-8f6f-aca9c9011bf5\") " pod="openstack/swift-storage-0" Jan 21 11:25:37 crc kubenswrapper[4824]: I0121 11:25:37.643354 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/f980e493-ca74-47f1-8f6f-aca9c9011bf5-etc-swift\") pod \"swift-storage-0\" (UID: \"f980e493-ca74-47f1-8f6f-aca9c9011bf5\") " pod="openstack/swift-storage-0" Jan 21 11:25:37 crc kubenswrapper[4824]: I0121 11:25:37.685953 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-a0fd-account-create-update-zxjvg" Jan 21 11:25:37 crc kubenswrapper[4824]: I0121 11:25:37.685950 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-a0fd-account-create-update-zxjvg" event={"ID":"a0ca5b98-4b63-4b4e-a80f-d52977ae3d01","Type":"ContainerDied","Data":"a17a042f056b3ff29d8f31835c21afba176775ec0e3832d56d1636a0f187eb66"} Jan 21 11:25:37 crc kubenswrapper[4824]: I0121 11:25:37.686262 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a17a042f056b3ff29d8f31835c21afba176775ec0e3832d56d1636a0f187eb66" Jan 21 11:25:37 crc kubenswrapper[4824]: I0121 11:25:37.687351 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-jbk7q" event={"ID":"7c01cf79-8a28-4229-9257-90124e66d3cc","Type":"ContainerDied","Data":"1e477713ddb0f2a67fe723732b272384c4ff5196b703dde89021649abdfee9d8"} Jan 21 11:25:37 crc kubenswrapper[4824]: I0121 11:25:37.687383 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-jbk7q" Jan 21 11:25:37 crc kubenswrapper[4824]: I0121 11:25:37.687385 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1e477713ddb0f2a67fe723732b272384c4ff5196b703dde89021649abdfee9d8" Jan 21 11:25:37 crc kubenswrapper[4824]: I0121 11:25:37.688486 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-q7wdg" event={"ID":"885db96c-779f-47c5-a383-1821cb53977c","Type":"ContainerDied","Data":"a7ec96b60c6cf63256dcb9a6367e913091482fe19b5f87ad070d29ec3152485f"} Jan 21 11:25:37 crc kubenswrapper[4824]: I0121 11:25:37.688514 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a7ec96b60c6cf63256dcb9a6367e913091482fe19b5f87ad070d29ec3152485f" Jan 21 11:25:37 crc kubenswrapper[4824]: I0121 11:25:37.688556 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-q7wdg" Jan 21 11:25:37 crc kubenswrapper[4824]: I0121 11:25:37.690465 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-qzfrf" event={"ID":"e94778b1-d36a-47e9-b201-9b7cd1cdb640","Type":"ContainerStarted","Data":"48fb0631b23867cac173ec7a9df005de8edd1cb0df987f8dceb763f2434dd5b7"} Jan 21 11:25:37 crc kubenswrapper[4824]: I0121 11:25:37.691612 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-383b-account-create-update-w99bv" event={"ID":"bc853ab0-e92a-4ada-84fd-b59ac2af573d","Type":"ContainerDied","Data":"43eefd9eea7557b24c2304803f43b0c074dd6eee119b0fe3f3750b2f784da5c5"} Jan 21 11:25:37 crc kubenswrapper[4824]: I0121 11:25:37.691646 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="43eefd9eea7557b24c2304803f43b0c074dd6eee119b0fe3f3750b2f784da5c5" Jan 21 11:25:37 crc kubenswrapper[4824]: I0121 11:25:37.691697 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-383b-account-create-update-w99bv" Jan 21 11:25:37 crc kubenswrapper[4824]: I0121 11:25:37.745171 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/f980e493-ca74-47f1-8f6f-aca9c9011bf5-etc-swift\") pod \"swift-storage-0\" (UID: \"f980e493-ca74-47f1-8f6f-aca9c9011bf5\") " pod="openstack/swift-storage-0" Jan 21 11:25:37 crc kubenswrapper[4824]: E0121 11:25:37.745352 4824 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Jan 21 11:25:37 crc kubenswrapper[4824]: E0121 11:25:37.745376 4824 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Jan 21 11:25:37 crc kubenswrapper[4824]: E0121 11:25:37.745419 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/f980e493-ca74-47f1-8f6f-aca9c9011bf5-etc-swift podName:f980e493-ca74-47f1-8f6f-aca9c9011bf5 nodeName:}" failed. No retries permitted until 2026-01-21 11:25:38.24540553 +0000 UTC m=+880.538434821 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/f980e493-ca74-47f1-8f6f-aca9c9011bf5-etc-swift") pod "swift-storage-0" (UID: "f980e493-ca74-47f1-8f6f-aca9c9011bf5") : configmap "swift-ring-files" not found Jan 21 11:25:37 crc kubenswrapper[4824]: I0121 11:25:37.745487 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/f980e493-ca74-47f1-8f6f-aca9c9011bf5-lock\") pod \"swift-storage-0\" (UID: \"f980e493-ca74-47f1-8f6f-aca9c9011bf5\") " pod="openstack/swift-storage-0" Jan 21 11:25:37 crc kubenswrapper[4824]: I0121 11:25:37.745513 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/f980e493-ca74-47f1-8f6f-aca9c9011bf5-cache\") pod \"swift-storage-0\" (UID: \"f980e493-ca74-47f1-8f6f-aca9c9011bf5\") " pod="openstack/swift-storage-0" Jan 21 11:25:37 crc kubenswrapper[4824]: I0121 11:25:37.745532 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5hwq7\" (UniqueName: \"kubernetes.io/projected/f980e493-ca74-47f1-8f6f-aca9c9011bf5-kube-api-access-5hwq7\") pod \"swift-storage-0\" (UID: \"f980e493-ca74-47f1-8f6f-aca9c9011bf5\") " pod="openstack/swift-storage-0" Jan 21 11:25:37 crc kubenswrapper[4824]: I0121 11:25:37.745557 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"swift-storage-0\" (UID: \"f980e493-ca74-47f1-8f6f-aca9c9011bf5\") " pod="openstack/swift-storage-0" Jan 21 11:25:37 crc kubenswrapper[4824]: I0121 11:25:37.745815 4824 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"swift-storage-0\" (UID: \"f980e493-ca74-47f1-8f6f-aca9c9011bf5\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/swift-storage-0" Jan 21 11:25:37 crc kubenswrapper[4824]: I0121 11:25:37.745864 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/f980e493-ca74-47f1-8f6f-aca9c9011bf5-lock\") pod \"swift-storage-0\" (UID: \"f980e493-ca74-47f1-8f6f-aca9c9011bf5\") " pod="openstack/swift-storage-0" Jan 21 11:25:37 crc kubenswrapper[4824]: I0121 11:25:37.745892 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/f980e493-ca74-47f1-8f6f-aca9c9011bf5-cache\") pod \"swift-storage-0\" (UID: \"f980e493-ca74-47f1-8f6f-aca9c9011bf5\") " pod="openstack/swift-storage-0" Jan 21 11:25:37 crc kubenswrapper[4824]: I0121 11:25:37.759598 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5hwq7\" (UniqueName: \"kubernetes.io/projected/f980e493-ca74-47f1-8f6f-aca9c9011bf5-kube-api-access-5hwq7\") pod \"swift-storage-0\" (UID: \"f980e493-ca74-47f1-8f6f-aca9c9011bf5\") " pod="openstack/swift-storage-0" Jan 21 11:25:37 crc kubenswrapper[4824]: I0121 11:25:37.765361 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"swift-storage-0\" (UID: \"f980e493-ca74-47f1-8f6f-aca9c9011bf5\") " pod="openstack/swift-storage-0" Jan 21 11:25:38 crc kubenswrapper[4824]: I0121 11:25:38.253101 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/f980e493-ca74-47f1-8f6f-aca9c9011bf5-etc-swift\") pod \"swift-storage-0\" (UID: \"f980e493-ca74-47f1-8f6f-aca9c9011bf5\") " pod="openstack/swift-storage-0" Jan 21 11:25:38 crc kubenswrapper[4824]: E0121 11:25:38.253294 4824 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Jan 21 11:25:38 crc kubenswrapper[4824]: E0121 11:25:38.253317 4824 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Jan 21 11:25:38 crc kubenswrapper[4824]: E0121 11:25:38.253366 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/f980e493-ca74-47f1-8f6f-aca9c9011bf5-etc-swift podName:f980e493-ca74-47f1-8f6f-aca9c9011bf5 nodeName:}" failed. No retries permitted until 2026-01-21 11:25:39.253351654 +0000 UTC m=+881.546380946 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/f980e493-ca74-47f1-8f6f-aca9c9011bf5-etc-swift") pod "swift-storage-0" (UID: "f980e493-ca74-47f1-8f6f-aca9c9011bf5") : configmap "swift-ring-files" not found Jan 21 11:25:39 crc kubenswrapper[4824]: I0121 11:25:39.266830 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/f980e493-ca74-47f1-8f6f-aca9c9011bf5-etc-swift\") pod \"swift-storage-0\" (UID: \"f980e493-ca74-47f1-8f6f-aca9c9011bf5\") " pod="openstack/swift-storage-0" Jan 21 11:25:39 crc kubenswrapper[4824]: E0121 11:25:39.267002 4824 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Jan 21 11:25:39 crc kubenswrapper[4824]: E0121 11:25:39.267021 4824 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Jan 21 11:25:39 crc kubenswrapper[4824]: E0121 11:25:39.267063 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/f980e493-ca74-47f1-8f6f-aca9c9011bf5-etc-swift podName:f980e493-ca74-47f1-8f6f-aca9c9011bf5 nodeName:}" failed. No retries permitted until 2026-01-21 11:25:41.267051492 +0000 UTC m=+883.560080784 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/f980e493-ca74-47f1-8f6f-aca9c9011bf5-etc-swift") pod "swift-storage-0" (UID: "f980e493-ca74-47f1-8f6f-aca9c9011bf5") : configmap "swift-ring-files" not found Jan 21 11:25:39 crc kubenswrapper[4824]: I0121 11:25:39.287525 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-86db49b7ff-vfbzh" podUID="292b341f-0f9e-411e-9f85-47bde4dcb2c4" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.106:5353: connect: connection refused" Jan 21 11:25:39 crc kubenswrapper[4824]: I0121 11:25:39.775142 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-rlxbb"] Jan 21 11:25:39 crc kubenswrapper[4824]: I0121 11:25:39.776196 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-rlxbb" Jan 21 11:25:39 crc kubenswrapper[4824]: I0121 11:25:39.784807 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-rlxbb"] Jan 21 11:25:39 crc kubenswrapper[4824]: I0121 11:25:39.875023 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-0697-account-create-update-9nvnh"] Jan 21 11:25:39 crc kubenswrapper[4824]: I0121 11:25:39.875796 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6fbc8b12-d965-44e9-822f-b95a5c5e88ab-operator-scripts\") pod \"glance-db-create-rlxbb\" (UID: \"6fbc8b12-d965-44e9-822f-b95a5c5e88ab\") " pod="openstack/glance-db-create-rlxbb" Jan 21 11:25:39 crc kubenswrapper[4824]: I0121 11:25:39.875849 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hxqbh\" (UniqueName: \"kubernetes.io/projected/6fbc8b12-d965-44e9-822f-b95a5c5e88ab-kube-api-access-hxqbh\") pod \"glance-db-create-rlxbb\" (UID: \"6fbc8b12-d965-44e9-822f-b95a5c5e88ab\") " pod="openstack/glance-db-create-rlxbb" Jan 21 11:25:39 crc kubenswrapper[4824]: I0121 11:25:39.875891 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-0697-account-create-update-9nvnh" Jan 21 11:25:39 crc kubenswrapper[4824]: I0121 11:25:39.877854 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Jan 21 11:25:39 crc kubenswrapper[4824]: I0121 11:25:39.882320 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-0697-account-create-update-9nvnh"] Jan 21 11:25:39 crc kubenswrapper[4824]: I0121 11:25:39.977830 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g4wtz\" (UniqueName: \"kubernetes.io/projected/72452c85-3787-4bd5-b605-af5555c1fdaf-kube-api-access-g4wtz\") pod \"glance-0697-account-create-update-9nvnh\" (UID: \"72452c85-3787-4bd5-b605-af5555c1fdaf\") " pod="openstack/glance-0697-account-create-update-9nvnh" Jan 21 11:25:39 crc kubenswrapper[4824]: I0121 11:25:39.977877 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/72452c85-3787-4bd5-b605-af5555c1fdaf-operator-scripts\") pod \"glance-0697-account-create-update-9nvnh\" (UID: \"72452c85-3787-4bd5-b605-af5555c1fdaf\") " pod="openstack/glance-0697-account-create-update-9nvnh" Jan 21 11:25:39 crc kubenswrapper[4824]: I0121 11:25:39.977927 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6fbc8b12-d965-44e9-822f-b95a5c5e88ab-operator-scripts\") pod \"glance-db-create-rlxbb\" (UID: \"6fbc8b12-d965-44e9-822f-b95a5c5e88ab\") " pod="openstack/glance-db-create-rlxbb" Jan 21 11:25:39 crc kubenswrapper[4824]: I0121 11:25:39.978013 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hxqbh\" (UniqueName: \"kubernetes.io/projected/6fbc8b12-d965-44e9-822f-b95a5c5e88ab-kube-api-access-hxqbh\") pod \"glance-db-create-rlxbb\" (UID: \"6fbc8b12-d965-44e9-822f-b95a5c5e88ab\") " pod="openstack/glance-db-create-rlxbb" Jan 21 11:25:39 crc kubenswrapper[4824]: I0121 11:25:39.978541 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6fbc8b12-d965-44e9-822f-b95a5c5e88ab-operator-scripts\") pod \"glance-db-create-rlxbb\" (UID: \"6fbc8b12-d965-44e9-822f-b95a5c5e88ab\") " pod="openstack/glance-db-create-rlxbb" Jan 21 11:25:39 crc kubenswrapper[4824]: I0121 11:25:39.993273 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hxqbh\" (UniqueName: \"kubernetes.io/projected/6fbc8b12-d965-44e9-822f-b95a5c5e88ab-kube-api-access-hxqbh\") pod \"glance-db-create-rlxbb\" (UID: \"6fbc8b12-d965-44e9-822f-b95a5c5e88ab\") " pod="openstack/glance-db-create-rlxbb" Jan 21 11:25:40 crc kubenswrapper[4824]: I0121 11:25:40.078973 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g4wtz\" (UniqueName: \"kubernetes.io/projected/72452c85-3787-4bd5-b605-af5555c1fdaf-kube-api-access-g4wtz\") pod \"glance-0697-account-create-update-9nvnh\" (UID: \"72452c85-3787-4bd5-b605-af5555c1fdaf\") " pod="openstack/glance-0697-account-create-update-9nvnh" Jan 21 11:25:40 crc kubenswrapper[4824]: I0121 11:25:40.079017 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/72452c85-3787-4bd5-b605-af5555c1fdaf-operator-scripts\") pod \"glance-0697-account-create-update-9nvnh\" (UID: \"72452c85-3787-4bd5-b605-af5555c1fdaf\") " pod="openstack/glance-0697-account-create-update-9nvnh" Jan 21 11:25:40 crc kubenswrapper[4824]: I0121 11:25:40.079616 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/72452c85-3787-4bd5-b605-af5555c1fdaf-operator-scripts\") pod \"glance-0697-account-create-update-9nvnh\" (UID: \"72452c85-3787-4bd5-b605-af5555c1fdaf\") " pod="openstack/glance-0697-account-create-update-9nvnh" Jan 21 11:25:40 crc kubenswrapper[4824]: I0121 11:25:40.091871 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-rlxbb" Jan 21 11:25:40 crc kubenswrapper[4824]: I0121 11:25:40.092100 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g4wtz\" (UniqueName: \"kubernetes.io/projected/72452c85-3787-4bd5-b605-af5555c1fdaf-kube-api-access-g4wtz\") pod \"glance-0697-account-create-update-9nvnh\" (UID: \"72452c85-3787-4bd5-b605-af5555c1fdaf\") " pod="openstack/glance-0697-account-create-update-9nvnh" Jan 21 11:25:40 crc kubenswrapper[4824]: I0121 11:25:40.187135 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-0697-account-create-update-9nvnh" Jan 21 11:25:40 crc kubenswrapper[4824]: I0121 11:25:40.457594 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-rlxbb"] Jan 21 11:25:40 crc kubenswrapper[4824]: W0121 11:25:40.460428 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6fbc8b12_d965_44e9_822f_b95a5c5e88ab.slice/crio-13850974d8bbf3ce2cfe084bd90bf6f50c0ac5f205f33608417019e7c200e2cf WatchSource:0}: Error finding container 13850974d8bbf3ce2cfe084bd90bf6f50c0ac5f205f33608417019e7c200e2cf: Status 404 returned error can't find the container with id 13850974d8bbf3ce2cfe084bd90bf6f50c0ac5f205f33608417019e7c200e2cf Jan 21 11:25:40 crc kubenswrapper[4824]: I0121 11:25:40.551783 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-0697-account-create-update-9nvnh"] Jan 21 11:25:40 crc kubenswrapper[4824]: W0121 11:25:40.554622 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod72452c85_3787_4bd5_b605_af5555c1fdaf.slice/crio-87203445239d34009800a67402ce4e349e19b952aad7441faa2589ed0b72b8fa WatchSource:0}: Error finding container 87203445239d34009800a67402ce4e349e19b952aad7441faa2589ed0b72b8fa: Status 404 returned error can't find the container with id 87203445239d34009800a67402ce4e349e19b952aad7441faa2589ed0b72b8fa Jan 21 11:25:40 crc kubenswrapper[4824]: I0121 11:25:40.708419 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-rlxbb" event={"ID":"6fbc8b12-d965-44e9-822f-b95a5c5e88ab","Type":"ContainerStarted","Data":"13850974d8bbf3ce2cfe084bd90bf6f50c0ac5f205f33608417019e7c200e2cf"} Jan 21 11:25:40 crc kubenswrapper[4824]: I0121 11:25:40.709409 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-0697-account-create-update-9nvnh" event={"ID":"72452c85-3787-4bd5-b605-af5555c1fdaf","Type":"ContainerStarted","Data":"87203445239d34009800a67402ce4e349e19b952aad7441faa2589ed0b72b8fa"} Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.296953 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/f980e493-ca74-47f1-8f6f-aca9c9011bf5-etc-swift\") pod \"swift-storage-0\" (UID: \"f980e493-ca74-47f1-8f6f-aca9c9011bf5\") " pod="openstack/swift-storage-0" Jan 21 11:25:41 crc kubenswrapper[4824]: E0121 11:25:41.297086 4824 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Jan 21 11:25:41 crc kubenswrapper[4824]: E0121 11:25:41.297220 4824 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Jan 21 11:25:41 crc kubenswrapper[4824]: E0121 11:25:41.297470 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/f980e493-ca74-47f1-8f6f-aca9c9011bf5-etc-swift podName:f980e493-ca74-47f1-8f6f-aca9c9011bf5 nodeName:}" failed. No retries permitted until 2026-01-21 11:25:45.297453385 +0000 UTC m=+887.590482678 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/f980e493-ca74-47f1-8f6f-aca9c9011bf5-etc-swift") pod "swift-storage-0" (UID: "f980e493-ca74-47f1-8f6f-aca9c9011bf5") : configmap "swift-ring-files" not found Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.368072 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-vfbzh" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.458610 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-w2dc9"] Jan 21 11:25:41 crc kubenswrapper[4824]: E0121 11:25:41.458937 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="292b341f-0f9e-411e-9f85-47bde4dcb2c4" containerName="init" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.458950 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="292b341f-0f9e-411e-9f85-47bde4dcb2c4" containerName="init" Jan 21 11:25:41 crc kubenswrapper[4824]: E0121 11:25:41.459011 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="292b341f-0f9e-411e-9f85-47bde4dcb2c4" containerName="dnsmasq-dns" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.459017 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="292b341f-0f9e-411e-9f85-47bde4dcb2c4" containerName="dnsmasq-dns" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.459161 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="292b341f-0f9e-411e-9f85-47bde4dcb2c4" containerName="dnsmasq-dns" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.459619 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-w2dc9" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.463087 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.463096 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.463130 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.501419 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/292b341f-0f9e-411e-9f85-47bde4dcb2c4-ovsdbserver-sb\") pod \"292b341f-0f9e-411e-9f85-47bde4dcb2c4\" (UID: \"292b341f-0f9e-411e-9f85-47bde4dcb2c4\") " Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.501500 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/292b341f-0f9e-411e-9f85-47bde4dcb2c4-dns-svc\") pod \"292b341f-0f9e-411e-9f85-47bde4dcb2c4\" (UID: \"292b341f-0f9e-411e-9f85-47bde4dcb2c4\") " Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.501551 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hg5nl\" (UniqueName: \"kubernetes.io/projected/292b341f-0f9e-411e-9f85-47bde4dcb2c4-kube-api-access-hg5nl\") pod \"292b341f-0f9e-411e-9f85-47bde4dcb2c4\" (UID: \"292b341f-0f9e-411e-9f85-47bde4dcb2c4\") " Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.501673 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/292b341f-0f9e-411e-9f85-47bde4dcb2c4-config\") pod \"292b341f-0f9e-411e-9f85-47bde4dcb2c4\" (UID: \"292b341f-0f9e-411e-9f85-47bde4dcb2c4\") " Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.501722 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/292b341f-0f9e-411e-9f85-47bde4dcb2c4-ovsdbserver-nb\") pod \"292b341f-0f9e-411e-9f85-47bde4dcb2c4\" (UID: \"292b341f-0f9e-411e-9f85-47bde4dcb2c4\") " Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.507415 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/292b341f-0f9e-411e-9f85-47bde4dcb2c4-kube-api-access-hg5nl" (OuterVolumeSpecName: "kube-api-access-hg5nl") pod "292b341f-0f9e-411e-9f85-47bde4dcb2c4" (UID: "292b341f-0f9e-411e-9f85-47bde4dcb2c4"). InnerVolumeSpecName "kube-api-access-hg5nl". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.507492 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-w2dc9"] Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.510677 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-nc6vf"] Jan 21 11:25:41 crc kubenswrapper[4824]: E0121 11:25:41.511080 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[combined-ca-bundle dispersionconf etc-swift kube-api-access-ssxbb ring-data-devices scripts swiftconf], unattached volumes=[], failed to process volumes=[combined-ca-bundle dispersionconf etc-swift kube-api-access-ssxbb ring-data-devices scripts swiftconf]: context canceled" pod="openstack/swift-ring-rebalance-w2dc9" podUID="d65692e9-ea67-4312-b382-dbe3442c3a32" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.511825 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-nc6vf" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.514754 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-w2dc9"] Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.519321 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-nc6vf"] Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.548144 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/292b341f-0f9e-411e-9f85-47bde4dcb2c4-config" (OuterVolumeSpecName: "config") pod "292b341f-0f9e-411e-9f85-47bde4dcb2c4" (UID: "292b341f-0f9e-411e-9f85-47bde4dcb2c4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.550347 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/292b341f-0f9e-411e-9f85-47bde4dcb2c4-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "292b341f-0f9e-411e-9f85-47bde4dcb2c4" (UID: "292b341f-0f9e-411e-9f85-47bde4dcb2c4"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.550555 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/292b341f-0f9e-411e-9f85-47bde4dcb2c4-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "292b341f-0f9e-411e-9f85-47bde4dcb2c4" (UID: "292b341f-0f9e-411e-9f85-47bde4dcb2c4"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.554360 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/292b341f-0f9e-411e-9f85-47bde4dcb2c4-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "292b341f-0f9e-411e-9f85-47bde4dcb2c4" (UID: "292b341f-0f9e-411e-9f85-47bde4dcb2c4"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.603065 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/29473359-3517-4b86-bd5e-80e25706ff27-dispersionconf\") pod \"swift-ring-rebalance-nc6vf\" (UID: \"29473359-3517-4b86-bd5e-80e25706ff27\") " pod="openstack/swift-ring-rebalance-nc6vf" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.603102 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/29473359-3517-4b86-bd5e-80e25706ff27-swiftconf\") pod \"swift-ring-rebalance-nc6vf\" (UID: \"29473359-3517-4b86-bd5e-80e25706ff27\") " pod="openstack/swift-ring-rebalance-nc6vf" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.603151 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ssxbb\" (UniqueName: \"kubernetes.io/projected/d65692e9-ea67-4312-b382-dbe3442c3a32-kube-api-access-ssxbb\") pod \"swift-ring-rebalance-w2dc9\" (UID: \"d65692e9-ea67-4312-b382-dbe3442c3a32\") " pod="openstack/swift-ring-rebalance-w2dc9" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.603217 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cffv8\" (UniqueName: \"kubernetes.io/projected/29473359-3517-4b86-bd5e-80e25706ff27-kube-api-access-cffv8\") pod \"swift-ring-rebalance-nc6vf\" (UID: \"29473359-3517-4b86-bd5e-80e25706ff27\") " pod="openstack/swift-ring-rebalance-nc6vf" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.603237 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d65692e9-ea67-4312-b382-dbe3442c3a32-combined-ca-bundle\") pod \"swift-ring-rebalance-w2dc9\" (UID: \"d65692e9-ea67-4312-b382-dbe3442c3a32\") " pod="openstack/swift-ring-rebalance-w2dc9" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.603259 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/29473359-3517-4b86-bd5e-80e25706ff27-scripts\") pod \"swift-ring-rebalance-nc6vf\" (UID: \"29473359-3517-4b86-bd5e-80e25706ff27\") " pod="openstack/swift-ring-rebalance-nc6vf" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.603273 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/d65692e9-ea67-4312-b382-dbe3442c3a32-ring-data-devices\") pod \"swift-ring-rebalance-w2dc9\" (UID: \"d65692e9-ea67-4312-b382-dbe3442c3a32\") " pod="openstack/swift-ring-rebalance-w2dc9" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.603307 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/29473359-3517-4b86-bd5e-80e25706ff27-etc-swift\") pod \"swift-ring-rebalance-nc6vf\" (UID: \"29473359-3517-4b86-bd5e-80e25706ff27\") " pod="openstack/swift-ring-rebalance-nc6vf" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.603322 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/29473359-3517-4b86-bd5e-80e25706ff27-ring-data-devices\") pod \"swift-ring-rebalance-nc6vf\" (UID: \"29473359-3517-4b86-bd5e-80e25706ff27\") " pod="openstack/swift-ring-rebalance-nc6vf" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.603348 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/d65692e9-ea67-4312-b382-dbe3442c3a32-etc-swift\") pod \"swift-ring-rebalance-w2dc9\" (UID: \"d65692e9-ea67-4312-b382-dbe3442c3a32\") " pod="openstack/swift-ring-rebalance-w2dc9" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.603374 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/d65692e9-ea67-4312-b382-dbe3442c3a32-swiftconf\") pod \"swift-ring-rebalance-w2dc9\" (UID: \"d65692e9-ea67-4312-b382-dbe3442c3a32\") " pod="openstack/swift-ring-rebalance-w2dc9" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.603427 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/d65692e9-ea67-4312-b382-dbe3442c3a32-dispersionconf\") pod \"swift-ring-rebalance-w2dc9\" (UID: \"d65692e9-ea67-4312-b382-dbe3442c3a32\") " pod="openstack/swift-ring-rebalance-w2dc9" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.603489 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29473359-3517-4b86-bd5e-80e25706ff27-combined-ca-bundle\") pod \"swift-ring-rebalance-nc6vf\" (UID: \"29473359-3517-4b86-bd5e-80e25706ff27\") " pod="openstack/swift-ring-rebalance-nc6vf" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.603530 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d65692e9-ea67-4312-b382-dbe3442c3a32-scripts\") pod \"swift-ring-rebalance-w2dc9\" (UID: \"d65692e9-ea67-4312-b382-dbe3442c3a32\") " pod="openstack/swift-ring-rebalance-w2dc9" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.603597 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/292b341f-0f9e-411e-9f85-47bde4dcb2c4-config\") on node \"crc\" DevicePath \"\"" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.603612 4824 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/292b341f-0f9e-411e-9f85-47bde4dcb2c4-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.603621 4824 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/292b341f-0f9e-411e-9f85-47bde4dcb2c4-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.603629 4824 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/292b341f-0f9e-411e-9f85-47bde4dcb2c4-dns-svc\") on node \"crc\" DevicePath \"\"" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.603637 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hg5nl\" (UniqueName: \"kubernetes.io/projected/292b341f-0f9e-411e-9f85-47bde4dcb2c4-kube-api-access-hg5nl\") on node \"crc\" DevicePath \"\"" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.626043 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/root-account-create-update-n665r"] Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.626913 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-n665r" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.629042 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-mariadb-root-db-secret" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.633267 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-n665r"] Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.704528 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ssxbb\" (UniqueName: \"kubernetes.io/projected/d65692e9-ea67-4312-b382-dbe3442c3a32-kube-api-access-ssxbb\") pod \"swift-ring-rebalance-w2dc9\" (UID: \"d65692e9-ea67-4312-b382-dbe3442c3a32\") " pod="openstack/swift-ring-rebalance-w2dc9" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.704589 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cffv8\" (UniqueName: \"kubernetes.io/projected/29473359-3517-4b86-bd5e-80e25706ff27-kube-api-access-cffv8\") pod \"swift-ring-rebalance-nc6vf\" (UID: \"29473359-3517-4b86-bd5e-80e25706ff27\") " pod="openstack/swift-ring-rebalance-nc6vf" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.704614 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d65692e9-ea67-4312-b382-dbe3442c3a32-combined-ca-bundle\") pod \"swift-ring-rebalance-w2dc9\" (UID: \"d65692e9-ea67-4312-b382-dbe3442c3a32\") " pod="openstack/swift-ring-rebalance-w2dc9" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.704636 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/29473359-3517-4b86-bd5e-80e25706ff27-scripts\") pod \"swift-ring-rebalance-nc6vf\" (UID: \"29473359-3517-4b86-bd5e-80e25706ff27\") " pod="openstack/swift-ring-rebalance-nc6vf" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.704652 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/d65692e9-ea67-4312-b382-dbe3442c3a32-ring-data-devices\") pod \"swift-ring-rebalance-w2dc9\" (UID: \"d65692e9-ea67-4312-b382-dbe3442c3a32\") " pod="openstack/swift-ring-rebalance-w2dc9" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.704667 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/29473359-3517-4b86-bd5e-80e25706ff27-etc-swift\") pod \"swift-ring-rebalance-nc6vf\" (UID: \"29473359-3517-4b86-bd5e-80e25706ff27\") " pod="openstack/swift-ring-rebalance-nc6vf" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.704695 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/29473359-3517-4b86-bd5e-80e25706ff27-ring-data-devices\") pod \"swift-ring-rebalance-nc6vf\" (UID: \"29473359-3517-4b86-bd5e-80e25706ff27\") " pod="openstack/swift-ring-rebalance-nc6vf" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.704724 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/d65692e9-ea67-4312-b382-dbe3442c3a32-etc-swift\") pod \"swift-ring-rebalance-w2dc9\" (UID: \"d65692e9-ea67-4312-b382-dbe3442c3a32\") " pod="openstack/swift-ring-rebalance-w2dc9" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.704749 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/d65692e9-ea67-4312-b382-dbe3442c3a32-swiftconf\") pod \"swift-ring-rebalance-w2dc9\" (UID: \"d65692e9-ea67-4312-b382-dbe3442c3a32\") " pod="openstack/swift-ring-rebalance-w2dc9" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.704794 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/d65692e9-ea67-4312-b382-dbe3442c3a32-dispersionconf\") pod \"swift-ring-rebalance-w2dc9\" (UID: \"d65692e9-ea67-4312-b382-dbe3442c3a32\") " pod="openstack/swift-ring-rebalance-w2dc9" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.704857 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29473359-3517-4b86-bd5e-80e25706ff27-combined-ca-bundle\") pod \"swift-ring-rebalance-nc6vf\" (UID: \"29473359-3517-4b86-bd5e-80e25706ff27\") " pod="openstack/swift-ring-rebalance-nc6vf" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.704902 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d65692e9-ea67-4312-b382-dbe3442c3a32-scripts\") pod \"swift-ring-rebalance-w2dc9\" (UID: \"d65692e9-ea67-4312-b382-dbe3442c3a32\") " pod="openstack/swift-ring-rebalance-w2dc9" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.704937 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/29473359-3517-4b86-bd5e-80e25706ff27-dispersionconf\") pod \"swift-ring-rebalance-nc6vf\" (UID: \"29473359-3517-4b86-bd5e-80e25706ff27\") " pod="openstack/swift-ring-rebalance-nc6vf" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.704951 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/29473359-3517-4b86-bd5e-80e25706ff27-swiftconf\") pod \"swift-ring-rebalance-nc6vf\" (UID: \"29473359-3517-4b86-bd5e-80e25706ff27\") " pod="openstack/swift-ring-rebalance-nc6vf" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.704990 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/af5c64ab-b4a4-43f5-81c3-416194e40b9a-operator-scripts\") pod \"root-account-create-update-n665r\" (UID: \"af5c64ab-b4a4-43f5-81c3-416194e40b9a\") " pod="openstack/root-account-create-update-n665r" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.705009 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8m496\" (UniqueName: \"kubernetes.io/projected/af5c64ab-b4a4-43f5-81c3-416194e40b9a-kube-api-access-8m496\") pod \"root-account-create-update-n665r\" (UID: \"af5c64ab-b4a4-43f5-81c3-416194e40b9a\") " pod="openstack/root-account-create-update-n665r" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.705412 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/29473359-3517-4b86-bd5e-80e25706ff27-etc-swift\") pod \"swift-ring-rebalance-nc6vf\" (UID: \"29473359-3517-4b86-bd5e-80e25706ff27\") " pod="openstack/swift-ring-rebalance-nc6vf" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.705420 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/d65692e9-ea67-4312-b382-dbe3442c3a32-ring-data-devices\") pod \"swift-ring-rebalance-w2dc9\" (UID: \"d65692e9-ea67-4312-b382-dbe3442c3a32\") " pod="openstack/swift-ring-rebalance-w2dc9" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.705914 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/d65692e9-ea67-4312-b382-dbe3442c3a32-etc-swift\") pod \"swift-ring-rebalance-w2dc9\" (UID: \"d65692e9-ea67-4312-b382-dbe3442c3a32\") " pod="openstack/swift-ring-rebalance-w2dc9" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.705931 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/29473359-3517-4b86-bd5e-80e25706ff27-scripts\") pod \"swift-ring-rebalance-nc6vf\" (UID: \"29473359-3517-4b86-bd5e-80e25706ff27\") " pod="openstack/swift-ring-rebalance-nc6vf" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.705946 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d65692e9-ea67-4312-b382-dbe3442c3a32-scripts\") pod \"swift-ring-rebalance-w2dc9\" (UID: \"d65692e9-ea67-4312-b382-dbe3442c3a32\") " pod="openstack/swift-ring-rebalance-w2dc9" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.706460 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/29473359-3517-4b86-bd5e-80e25706ff27-ring-data-devices\") pod \"swift-ring-rebalance-nc6vf\" (UID: \"29473359-3517-4b86-bd5e-80e25706ff27\") " pod="openstack/swift-ring-rebalance-nc6vf" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.708704 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/29473359-3517-4b86-bd5e-80e25706ff27-swiftconf\") pod \"swift-ring-rebalance-nc6vf\" (UID: \"29473359-3517-4b86-bd5e-80e25706ff27\") " pod="openstack/swift-ring-rebalance-nc6vf" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.709224 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29473359-3517-4b86-bd5e-80e25706ff27-combined-ca-bundle\") pod \"swift-ring-rebalance-nc6vf\" (UID: \"29473359-3517-4b86-bd5e-80e25706ff27\") " pod="openstack/swift-ring-rebalance-nc6vf" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.709791 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/d65692e9-ea67-4312-b382-dbe3442c3a32-swiftconf\") pod \"swift-ring-rebalance-w2dc9\" (UID: \"d65692e9-ea67-4312-b382-dbe3442c3a32\") " pod="openstack/swift-ring-rebalance-w2dc9" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.709807 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d65692e9-ea67-4312-b382-dbe3442c3a32-combined-ca-bundle\") pod \"swift-ring-rebalance-w2dc9\" (UID: \"d65692e9-ea67-4312-b382-dbe3442c3a32\") " pod="openstack/swift-ring-rebalance-w2dc9" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.710622 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/d65692e9-ea67-4312-b382-dbe3442c3a32-dispersionconf\") pod \"swift-ring-rebalance-w2dc9\" (UID: \"d65692e9-ea67-4312-b382-dbe3442c3a32\") " pod="openstack/swift-ring-rebalance-w2dc9" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.711235 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/29473359-3517-4b86-bd5e-80e25706ff27-dispersionconf\") pod \"swift-ring-rebalance-nc6vf\" (UID: \"29473359-3517-4b86-bd5e-80e25706ff27\") " pod="openstack/swift-ring-rebalance-nc6vf" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.718186 4824 generic.go:334] "Generic (PLEG): container finished" podID="e94778b1-d36a-47e9-b201-9b7cd1cdb640" containerID="a517f23f781efca10654bdff86d5e2efc32071ea09d5af5829095603d962c629" exitCode=0 Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.718242 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-qzfrf" event={"ID":"e94778b1-d36a-47e9-b201-9b7cd1cdb640","Type":"ContainerDied","Data":"a517f23f781efca10654bdff86d5e2efc32071ea09d5af5829095603d962c629"} Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.719328 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ssxbb\" (UniqueName: \"kubernetes.io/projected/d65692e9-ea67-4312-b382-dbe3442c3a32-kube-api-access-ssxbb\") pod \"swift-ring-rebalance-w2dc9\" (UID: \"d65692e9-ea67-4312-b382-dbe3442c3a32\") " pod="openstack/swift-ring-rebalance-w2dc9" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.719884 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cffv8\" (UniqueName: \"kubernetes.io/projected/29473359-3517-4b86-bd5e-80e25706ff27-kube-api-access-cffv8\") pod \"swift-ring-rebalance-nc6vf\" (UID: \"29473359-3517-4b86-bd5e-80e25706ff27\") " pod="openstack/swift-ring-rebalance-nc6vf" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.720482 4824 generic.go:334] "Generic (PLEG): container finished" podID="72452c85-3787-4bd5-b605-af5555c1fdaf" containerID="aece7515cad2f167dd4b5a5e58d5aa5c0df426c12c8bbc743e03dad13feaedf7" exitCode=0 Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.720513 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-0697-account-create-update-9nvnh" event={"ID":"72452c85-3787-4bd5-b605-af5555c1fdaf","Type":"ContainerDied","Data":"aece7515cad2f167dd4b5a5e58d5aa5c0df426c12c8bbc743e03dad13feaedf7"} Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.723781 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-vfbzh" event={"ID":"292b341f-0f9e-411e-9f85-47bde4dcb2c4","Type":"ContainerDied","Data":"c0704ee36eacd59c7b38561ac33e2d4956e244055bb28267047fee7b43cd9db0"} Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.723820 4824 scope.go:117] "RemoveContainer" containerID="095e6c8fd89e1deec2822bbad9ce47d3bc5541769b4bdf66512d4544bcb46405" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.723940 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-vfbzh" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.726227 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"8b9c25b1-d0b9-430e-88cf-0aeaf6785461","Type":"ContainerStarted","Data":"301fce5be639a4edea0791738f7bedb2f7586b53e5f9101d69276726401d0a51"} Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.726258 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"8b9c25b1-d0b9-430e-88cf-0aeaf6785461","Type":"ContainerStarted","Data":"6dceef68019a4aed59819b21a85c3314c7f535000cc74c2484dafc2002f1d644"} Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.727313 4824 generic.go:334] "Generic (PLEG): container finished" podID="6fbc8b12-d965-44e9-822f-b95a5c5e88ab" containerID="091c53575308b0937e0789468c0edd3eb80a63b653591b4cc1559689a8b4b286" exitCode=0 Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.727463 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-w2dc9" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.728227 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-rlxbb" event={"ID":"6fbc8b12-d965-44e9-822f-b95a5c5e88ab","Type":"ContainerDied","Data":"091c53575308b0937e0789468c0edd3eb80a63b653591b4cc1559689a8b4b286"} Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.741895 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-w2dc9" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.749903 4824 scope.go:117] "RemoveContainer" containerID="b9f66e73f6529373d74d9e19ba52303b84fabf65da4786d896c78db5e897096f" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.755538 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=2.106391518 podStartE2EDuration="6.755523996s" podCreationTimestamp="2026-01-21 11:25:35 +0000 UTC" firstStartedPulling="2026-01-21 11:25:36.613513837 +0000 UTC m=+878.906543129" lastFinishedPulling="2026-01-21 11:25:41.262646314 +0000 UTC m=+883.555675607" observedRunningTime="2026-01-21 11:25:41.751742098 +0000 UTC m=+884.044771391" watchObservedRunningTime="2026-01-21 11:25:41.755523996 +0000 UTC m=+884.048553288" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.788726 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-vfbzh"] Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.793864 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-vfbzh"] Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.806738 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/af5c64ab-b4a4-43f5-81c3-416194e40b9a-operator-scripts\") pod \"root-account-create-update-n665r\" (UID: \"af5c64ab-b4a4-43f5-81c3-416194e40b9a\") " pod="openstack/root-account-create-update-n665r" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.806893 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8m496\" (UniqueName: \"kubernetes.io/projected/af5c64ab-b4a4-43f5-81c3-416194e40b9a-kube-api-access-8m496\") pod \"root-account-create-update-n665r\" (UID: \"af5c64ab-b4a4-43f5-81c3-416194e40b9a\") " pod="openstack/root-account-create-update-n665r" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.807420 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/af5c64ab-b4a4-43f5-81c3-416194e40b9a-operator-scripts\") pod \"root-account-create-update-n665r\" (UID: \"af5c64ab-b4a4-43f5-81c3-416194e40b9a\") " pod="openstack/root-account-create-update-n665r" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.821487 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8m496\" (UniqueName: \"kubernetes.io/projected/af5c64ab-b4a4-43f5-81c3-416194e40b9a-kube-api-access-8m496\") pod \"root-account-create-update-n665r\" (UID: \"af5c64ab-b4a4-43f5-81c3-416194e40b9a\") " pod="openstack/root-account-create-update-n665r" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.838616 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-nc6vf" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.910068 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/d65692e9-ea67-4312-b382-dbe3442c3a32-swiftconf\") pod \"d65692e9-ea67-4312-b382-dbe3442c3a32\" (UID: \"d65692e9-ea67-4312-b382-dbe3442c3a32\") " Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.910251 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d65692e9-ea67-4312-b382-dbe3442c3a32-scripts\") pod \"d65692e9-ea67-4312-b382-dbe3442c3a32\" (UID: \"d65692e9-ea67-4312-b382-dbe3442c3a32\") " Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.910292 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/d65692e9-ea67-4312-b382-dbe3442c3a32-ring-data-devices\") pod \"d65692e9-ea67-4312-b382-dbe3442c3a32\" (UID: \"d65692e9-ea67-4312-b382-dbe3442c3a32\") " Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.910326 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ssxbb\" (UniqueName: \"kubernetes.io/projected/d65692e9-ea67-4312-b382-dbe3442c3a32-kube-api-access-ssxbb\") pod \"d65692e9-ea67-4312-b382-dbe3442c3a32\" (UID: \"d65692e9-ea67-4312-b382-dbe3442c3a32\") " Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.910409 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/d65692e9-ea67-4312-b382-dbe3442c3a32-dispersionconf\") pod \"d65692e9-ea67-4312-b382-dbe3442c3a32\" (UID: \"d65692e9-ea67-4312-b382-dbe3442c3a32\") " Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.910425 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/d65692e9-ea67-4312-b382-dbe3442c3a32-etc-swift\") pod \"d65692e9-ea67-4312-b382-dbe3442c3a32\" (UID: \"d65692e9-ea67-4312-b382-dbe3442c3a32\") " Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.910461 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d65692e9-ea67-4312-b382-dbe3442c3a32-combined-ca-bundle\") pod \"d65692e9-ea67-4312-b382-dbe3442c3a32\" (UID: \"d65692e9-ea67-4312-b382-dbe3442c3a32\") " Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.910702 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d65692e9-ea67-4312-b382-dbe3442c3a32-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "d65692e9-ea67-4312-b382-dbe3442c3a32" (UID: "d65692e9-ea67-4312-b382-dbe3442c3a32"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.910758 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d65692e9-ea67-4312-b382-dbe3442c3a32-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "d65692e9-ea67-4312-b382-dbe3442c3a32" (UID: "d65692e9-ea67-4312-b382-dbe3442c3a32"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.910785 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d65692e9-ea67-4312-b382-dbe3442c3a32-scripts" (OuterVolumeSpecName: "scripts") pod "d65692e9-ea67-4312-b382-dbe3442c3a32" (UID: "d65692e9-ea67-4312-b382-dbe3442c3a32"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.910991 4824 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/d65692e9-ea67-4312-b382-dbe3442c3a32-etc-swift\") on node \"crc\" DevicePath \"\"" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.911004 4824 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d65692e9-ea67-4312-b382-dbe3442c3a32-scripts\") on node \"crc\" DevicePath \"\"" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.911013 4824 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/d65692e9-ea67-4312-b382-dbe3442c3a32-ring-data-devices\") on node \"crc\" DevicePath \"\"" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.912730 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d65692e9-ea67-4312-b382-dbe3442c3a32-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "d65692e9-ea67-4312-b382-dbe3442c3a32" (UID: "d65692e9-ea67-4312-b382-dbe3442c3a32"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.912830 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d65692e9-ea67-4312-b382-dbe3442c3a32-kube-api-access-ssxbb" (OuterVolumeSpecName: "kube-api-access-ssxbb") pod "d65692e9-ea67-4312-b382-dbe3442c3a32" (UID: "d65692e9-ea67-4312-b382-dbe3442c3a32"). InnerVolumeSpecName "kube-api-access-ssxbb". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.913053 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d65692e9-ea67-4312-b382-dbe3442c3a32-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d65692e9-ea67-4312-b382-dbe3442c3a32" (UID: "d65692e9-ea67-4312-b382-dbe3442c3a32"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.914294 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d65692e9-ea67-4312-b382-dbe3442c3a32-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "d65692e9-ea67-4312-b382-dbe3442c3a32" (UID: "d65692e9-ea67-4312-b382-dbe3442c3a32"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:25:41 crc kubenswrapper[4824]: I0121 11:25:41.940281 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-n665r" Jan 21 11:25:42 crc kubenswrapper[4824]: I0121 11:25:42.012147 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ssxbb\" (UniqueName: \"kubernetes.io/projected/d65692e9-ea67-4312-b382-dbe3442c3a32-kube-api-access-ssxbb\") on node \"crc\" DevicePath \"\"" Jan 21 11:25:42 crc kubenswrapper[4824]: I0121 11:25:42.012173 4824 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/d65692e9-ea67-4312-b382-dbe3442c3a32-dispersionconf\") on node \"crc\" DevicePath \"\"" Jan 21 11:25:42 crc kubenswrapper[4824]: I0121 11:25:42.012183 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d65692e9-ea67-4312-b382-dbe3442c3a32-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 21 11:25:42 crc kubenswrapper[4824]: I0121 11:25:42.012191 4824 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/d65692e9-ea67-4312-b382-dbe3442c3a32-swiftconf\") on node \"crc\" DevicePath \"\"" Jan 21 11:25:42 crc kubenswrapper[4824]: I0121 11:25:42.058603 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="292b341f-0f9e-411e-9f85-47bde4dcb2c4" path="/var/lib/kubelet/pods/292b341f-0f9e-411e-9f85-47bde4dcb2c4/volumes" Jan 21 11:25:42 crc kubenswrapper[4824]: I0121 11:25:42.190551 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-nc6vf"] Jan 21 11:25:42 crc kubenswrapper[4824]: W0121 11:25:42.194483 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod29473359_3517_4b86_bd5e_80e25706ff27.slice/crio-be8ab46a5292e712c0ba8d486b130fdd73752ef5b1cc0ad9789ef7c0c22e7dd0 WatchSource:0}: Error finding container be8ab46a5292e712c0ba8d486b130fdd73752ef5b1cc0ad9789ef7c0c22e7dd0: Status 404 returned error can't find the container with id be8ab46a5292e712c0ba8d486b130fdd73752ef5b1cc0ad9789ef7c0c22e7dd0 Jan 21 11:25:42 crc kubenswrapper[4824]: I0121 11:25:42.294381 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-n665r"] Jan 21 11:25:42 crc kubenswrapper[4824]: I0121 11:25:42.734917 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-qzfrf" event={"ID":"e94778b1-d36a-47e9-b201-9b7cd1cdb640","Type":"ContainerStarted","Data":"8a0b04dc02e5d8c871ec744b980efd9b4e195f4024e4ef8187ce38c2f8284265"} Jan 21 11:25:42 crc kubenswrapper[4824]: I0121 11:25:42.735146 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-698758b865-qzfrf" Jan 21 11:25:42 crc kubenswrapper[4824]: I0121 11:25:42.735778 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-nc6vf" event={"ID":"29473359-3517-4b86-bd5e-80e25706ff27","Type":"ContainerStarted","Data":"be8ab46a5292e712c0ba8d486b130fdd73752ef5b1cc0ad9789ef7c0c22e7dd0"} Jan 21 11:25:42 crc kubenswrapper[4824]: I0121 11:25:42.736830 4824 generic.go:334] "Generic (PLEG): container finished" podID="af5c64ab-b4a4-43f5-81c3-416194e40b9a" containerID="26ba4c23450a1625fb62c54cdc290126e5f1371f38bf091a20d00cb1c946e58f" exitCode=0 Jan 21 11:25:42 crc kubenswrapper[4824]: I0121 11:25:42.736868 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-n665r" event={"ID":"af5c64ab-b4a4-43f5-81c3-416194e40b9a","Type":"ContainerDied","Data":"26ba4c23450a1625fb62c54cdc290126e5f1371f38bf091a20d00cb1c946e58f"} Jan 21 11:25:42 crc kubenswrapper[4824]: I0121 11:25:42.736886 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-n665r" event={"ID":"af5c64ab-b4a4-43f5-81c3-416194e40b9a","Type":"ContainerStarted","Data":"ad7dafc5bf4c4bfe5a803af9f930e427405aeb6acf9bb0bca8e9c155247c42ed"} Jan 21 11:25:42 crc kubenswrapper[4824]: I0121 11:25:42.736895 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-w2dc9" Jan 21 11:25:42 crc kubenswrapper[4824]: I0121 11:25:42.737271 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Jan 21 11:25:42 crc kubenswrapper[4824]: I0121 11:25:42.756060 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-698758b865-qzfrf" podStartSLOduration=6.756046675 podStartE2EDuration="6.756046675s" podCreationTimestamp="2026-01-21 11:25:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:25:42.752804905 +0000 UTC m=+885.045834197" watchObservedRunningTime="2026-01-21 11:25:42.756046675 +0000 UTC m=+885.049075967" Jan 21 11:25:42 crc kubenswrapper[4824]: I0121 11:25:42.777267 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-w2dc9"] Jan 21 11:25:42 crc kubenswrapper[4824]: I0121 11:25:42.783666 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-ring-rebalance-w2dc9"] Jan 21 11:25:43 crc kubenswrapper[4824]: I0121 11:25:43.102481 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-rlxbb" Jan 21 11:25:43 crc kubenswrapper[4824]: I0121 11:25:43.159123 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-0697-account-create-update-9nvnh" Jan 21 11:25:43 crc kubenswrapper[4824]: I0121 11:25:43.231043 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hxqbh\" (UniqueName: \"kubernetes.io/projected/6fbc8b12-d965-44e9-822f-b95a5c5e88ab-kube-api-access-hxqbh\") pod \"6fbc8b12-d965-44e9-822f-b95a5c5e88ab\" (UID: \"6fbc8b12-d965-44e9-822f-b95a5c5e88ab\") " Jan 21 11:25:43 crc kubenswrapper[4824]: I0121 11:25:43.231115 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g4wtz\" (UniqueName: \"kubernetes.io/projected/72452c85-3787-4bd5-b605-af5555c1fdaf-kube-api-access-g4wtz\") pod \"72452c85-3787-4bd5-b605-af5555c1fdaf\" (UID: \"72452c85-3787-4bd5-b605-af5555c1fdaf\") " Jan 21 11:25:43 crc kubenswrapper[4824]: I0121 11:25:43.231159 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/72452c85-3787-4bd5-b605-af5555c1fdaf-operator-scripts\") pod \"72452c85-3787-4bd5-b605-af5555c1fdaf\" (UID: \"72452c85-3787-4bd5-b605-af5555c1fdaf\") " Jan 21 11:25:43 crc kubenswrapper[4824]: I0121 11:25:43.231237 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6fbc8b12-d965-44e9-822f-b95a5c5e88ab-operator-scripts\") pod \"6fbc8b12-d965-44e9-822f-b95a5c5e88ab\" (UID: \"6fbc8b12-d965-44e9-822f-b95a5c5e88ab\") " Jan 21 11:25:43 crc kubenswrapper[4824]: I0121 11:25:43.231699 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/72452c85-3787-4bd5-b605-af5555c1fdaf-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "72452c85-3787-4bd5-b605-af5555c1fdaf" (UID: "72452c85-3787-4bd5-b605-af5555c1fdaf"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:25:43 crc kubenswrapper[4824]: I0121 11:25:43.232121 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6fbc8b12-d965-44e9-822f-b95a5c5e88ab-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "6fbc8b12-d965-44e9-822f-b95a5c5e88ab" (UID: "6fbc8b12-d965-44e9-822f-b95a5c5e88ab"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:25:43 crc kubenswrapper[4824]: I0121 11:25:43.235696 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6fbc8b12-d965-44e9-822f-b95a5c5e88ab-kube-api-access-hxqbh" (OuterVolumeSpecName: "kube-api-access-hxqbh") pod "6fbc8b12-d965-44e9-822f-b95a5c5e88ab" (UID: "6fbc8b12-d965-44e9-822f-b95a5c5e88ab"). InnerVolumeSpecName "kube-api-access-hxqbh". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:25:43 crc kubenswrapper[4824]: I0121 11:25:43.235739 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/72452c85-3787-4bd5-b605-af5555c1fdaf-kube-api-access-g4wtz" (OuterVolumeSpecName: "kube-api-access-g4wtz") pod "72452c85-3787-4bd5-b605-af5555c1fdaf" (UID: "72452c85-3787-4bd5-b605-af5555c1fdaf"). InnerVolumeSpecName "kube-api-access-g4wtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:25:43 crc kubenswrapper[4824]: I0121 11:25:43.295854 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-gnlcj"] Jan 21 11:25:43 crc kubenswrapper[4824]: E0121 11:25:43.296270 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6fbc8b12-d965-44e9-822f-b95a5c5e88ab" containerName="mariadb-database-create" Jan 21 11:25:43 crc kubenswrapper[4824]: I0121 11:25:43.296299 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="6fbc8b12-d965-44e9-822f-b95a5c5e88ab" containerName="mariadb-database-create" Jan 21 11:25:43 crc kubenswrapper[4824]: E0121 11:25:43.296319 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="72452c85-3787-4bd5-b605-af5555c1fdaf" containerName="mariadb-account-create-update" Jan 21 11:25:43 crc kubenswrapper[4824]: I0121 11:25:43.296324 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="72452c85-3787-4bd5-b605-af5555c1fdaf" containerName="mariadb-account-create-update" Jan 21 11:25:43 crc kubenswrapper[4824]: I0121 11:25:43.296535 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="6fbc8b12-d965-44e9-822f-b95a5c5e88ab" containerName="mariadb-database-create" Jan 21 11:25:43 crc kubenswrapper[4824]: I0121 11:25:43.296547 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="72452c85-3787-4bd5-b605-af5555c1fdaf" containerName="mariadb-account-create-update" Jan 21 11:25:43 crc kubenswrapper[4824]: I0121 11:25:43.297817 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gnlcj" Jan 21 11:25:43 crc kubenswrapper[4824]: I0121 11:25:43.305178 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-gnlcj"] Jan 21 11:25:43 crc kubenswrapper[4824]: I0121 11:25:43.333476 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g4wtz\" (UniqueName: \"kubernetes.io/projected/72452c85-3787-4bd5-b605-af5555c1fdaf-kube-api-access-g4wtz\") on node \"crc\" DevicePath \"\"" Jan 21 11:25:43 crc kubenswrapper[4824]: I0121 11:25:43.333502 4824 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/72452c85-3787-4bd5-b605-af5555c1fdaf-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 21 11:25:43 crc kubenswrapper[4824]: I0121 11:25:43.334603 4824 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6fbc8b12-d965-44e9-822f-b95a5c5e88ab-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 21 11:25:43 crc kubenswrapper[4824]: I0121 11:25:43.334614 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hxqbh\" (UniqueName: \"kubernetes.io/projected/6fbc8b12-d965-44e9-822f-b95a5c5e88ab-kube-api-access-hxqbh\") on node \"crc\" DevicePath \"\"" Jan 21 11:25:43 crc kubenswrapper[4824]: I0121 11:25:43.435828 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bb4180b2-5d3e-486d-8197-348d01d2286e-catalog-content\") pod \"community-operators-gnlcj\" (UID: \"bb4180b2-5d3e-486d-8197-348d01d2286e\") " pod="openshift-marketplace/community-operators-gnlcj" Jan 21 11:25:43 crc kubenswrapper[4824]: I0121 11:25:43.435999 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6mmv9\" (UniqueName: \"kubernetes.io/projected/bb4180b2-5d3e-486d-8197-348d01d2286e-kube-api-access-6mmv9\") pod \"community-operators-gnlcj\" (UID: \"bb4180b2-5d3e-486d-8197-348d01d2286e\") " pod="openshift-marketplace/community-operators-gnlcj" Jan 21 11:25:43 crc kubenswrapper[4824]: I0121 11:25:43.436040 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bb4180b2-5d3e-486d-8197-348d01d2286e-utilities\") pod \"community-operators-gnlcj\" (UID: \"bb4180b2-5d3e-486d-8197-348d01d2286e\") " pod="openshift-marketplace/community-operators-gnlcj" Jan 21 11:25:43 crc kubenswrapper[4824]: I0121 11:25:43.537763 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6mmv9\" (UniqueName: \"kubernetes.io/projected/bb4180b2-5d3e-486d-8197-348d01d2286e-kube-api-access-6mmv9\") pod \"community-operators-gnlcj\" (UID: \"bb4180b2-5d3e-486d-8197-348d01d2286e\") " pod="openshift-marketplace/community-operators-gnlcj" Jan 21 11:25:43 crc kubenswrapper[4824]: I0121 11:25:43.537831 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bb4180b2-5d3e-486d-8197-348d01d2286e-utilities\") pod \"community-operators-gnlcj\" (UID: \"bb4180b2-5d3e-486d-8197-348d01d2286e\") " pod="openshift-marketplace/community-operators-gnlcj" Jan 21 11:25:43 crc kubenswrapper[4824]: I0121 11:25:43.537950 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bb4180b2-5d3e-486d-8197-348d01d2286e-catalog-content\") pod \"community-operators-gnlcj\" (UID: \"bb4180b2-5d3e-486d-8197-348d01d2286e\") " pod="openshift-marketplace/community-operators-gnlcj" Jan 21 11:25:43 crc kubenswrapper[4824]: I0121 11:25:43.538324 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bb4180b2-5d3e-486d-8197-348d01d2286e-catalog-content\") pod \"community-operators-gnlcj\" (UID: \"bb4180b2-5d3e-486d-8197-348d01d2286e\") " pod="openshift-marketplace/community-operators-gnlcj" Jan 21 11:25:43 crc kubenswrapper[4824]: I0121 11:25:43.538558 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bb4180b2-5d3e-486d-8197-348d01d2286e-utilities\") pod \"community-operators-gnlcj\" (UID: \"bb4180b2-5d3e-486d-8197-348d01d2286e\") " pod="openshift-marketplace/community-operators-gnlcj" Jan 21 11:25:43 crc kubenswrapper[4824]: I0121 11:25:43.559698 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6mmv9\" (UniqueName: \"kubernetes.io/projected/bb4180b2-5d3e-486d-8197-348d01d2286e-kube-api-access-6mmv9\") pod \"community-operators-gnlcj\" (UID: \"bb4180b2-5d3e-486d-8197-348d01d2286e\") " pod="openshift-marketplace/community-operators-gnlcj" Jan 21 11:25:43 crc kubenswrapper[4824]: I0121 11:25:43.622781 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gnlcj" Jan 21 11:25:43 crc kubenswrapper[4824]: I0121 11:25:43.744728 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-0697-account-create-update-9nvnh" event={"ID":"72452c85-3787-4bd5-b605-af5555c1fdaf","Type":"ContainerDied","Data":"87203445239d34009800a67402ce4e349e19b952aad7441faa2589ed0b72b8fa"} Jan 21 11:25:43 crc kubenswrapper[4824]: I0121 11:25:43.744771 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="87203445239d34009800a67402ce4e349e19b952aad7441faa2589ed0b72b8fa" Jan 21 11:25:43 crc kubenswrapper[4824]: I0121 11:25:43.744745 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-0697-account-create-update-9nvnh" Jan 21 11:25:43 crc kubenswrapper[4824]: I0121 11:25:43.745938 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-rlxbb" Jan 21 11:25:43 crc kubenswrapper[4824]: I0121 11:25:43.746418 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-rlxbb" event={"ID":"6fbc8b12-d965-44e9-822f-b95a5c5e88ab","Type":"ContainerDied","Data":"13850974d8bbf3ce2cfe084bd90bf6f50c0ac5f205f33608417019e7c200e2cf"} Jan 21 11:25:43 crc kubenswrapper[4824]: I0121 11:25:43.746470 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="13850974d8bbf3ce2cfe084bd90bf6f50c0ac5f205f33608417019e7c200e2cf" Jan 21 11:25:44 crc kubenswrapper[4824]: I0121 11:25:44.056237 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d65692e9-ea67-4312-b382-dbe3442c3a32" path="/var/lib/kubelet/pods/d65692e9-ea67-4312-b382-dbe3442c3a32/volumes" Jan 21 11:25:44 crc kubenswrapper[4824]: I0121 11:25:44.753828 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-n665r" event={"ID":"af5c64ab-b4a4-43f5-81c3-416194e40b9a","Type":"ContainerDied","Data":"ad7dafc5bf4c4bfe5a803af9f930e427405aeb6acf9bb0bca8e9c155247c42ed"} Jan 21 11:25:44 crc kubenswrapper[4824]: I0121 11:25:44.753867 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ad7dafc5bf4c4bfe5a803af9f930e427405aeb6acf9bb0bca8e9c155247c42ed" Jan 21 11:25:44 crc kubenswrapper[4824]: I0121 11:25:44.868864 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-n665r" Jan 21 11:25:44 crc kubenswrapper[4824]: I0121 11:25:44.961851 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8m496\" (UniqueName: \"kubernetes.io/projected/af5c64ab-b4a4-43f5-81c3-416194e40b9a-kube-api-access-8m496\") pod \"af5c64ab-b4a4-43f5-81c3-416194e40b9a\" (UID: \"af5c64ab-b4a4-43f5-81c3-416194e40b9a\") " Jan 21 11:25:44 crc kubenswrapper[4824]: I0121 11:25:44.961968 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/af5c64ab-b4a4-43f5-81c3-416194e40b9a-operator-scripts\") pod \"af5c64ab-b4a4-43f5-81c3-416194e40b9a\" (UID: \"af5c64ab-b4a4-43f5-81c3-416194e40b9a\") " Jan 21 11:25:44 crc kubenswrapper[4824]: I0121 11:25:44.962892 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/af5c64ab-b4a4-43f5-81c3-416194e40b9a-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "af5c64ab-b4a4-43f5-81c3-416194e40b9a" (UID: "af5c64ab-b4a4-43f5-81c3-416194e40b9a"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:25:44 crc kubenswrapper[4824]: I0121 11:25:44.966069 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/af5c64ab-b4a4-43f5-81c3-416194e40b9a-kube-api-access-8m496" (OuterVolumeSpecName: "kube-api-access-8m496") pod "af5c64ab-b4a4-43f5-81c3-416194e40b9a" (UID: "af5c64ab-b4a4-43f5-81c3-416194e40b9a"). InnerVolumeSpecName "kube-api-access-8m496". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:25:44 crc kubenswrapper[4824]: I0121 11:25:44.998476 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-sx9lb"] Jan 21 11:25:44 crc kubenswrapper[4824]: E0121 11:25:44.998796 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af5c64ab-b4a4-43f5-81c3-416194e40b9a" containerName="mariadb-account-create-update" Jan 21 11:25:44 crc kubenswrapper[4824]: I0121 11:25:44.998813 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="af5c64ab-b4a4-43f5-81c3-416194e40b9a" containerName="mariadb-account-create-update" Jan 21 11:25:44 crc kubenswrapper[4824]: I0121 11:25:44.998981 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="af5c64ab-b4a4-43f5-81c3-416194e40b9a" containerName="mariadb-account-create-update" Jan 21 11:25:44 crc kubenswrapper[4824]: I0121 11:25:44.999437 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-sx9lb" Jan 21 11:25:45 crc kubenswrapper[4824]: I0121 11:25:45.001336 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-4bzft" Jan 21 11:25:45 crc kubenswrapper[4824]: I0121 11:25:45.001536 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Jan 21 11:25:45 crc kubenswrapper[4824]: I0121 11:25:45.006546 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-sx9lb"] Jan 21 11:25:45 crc kubenswrapper[4824]: I0121 11:25:45.063355 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8m496\" (UniqueName: \"kubernetes.io/projected/af5c64ab-b4a4-43f5-81c3-416194e40b9a-kube-api-access-8m496\") on node \"crc\" DevicePath \"\"" Jan 21 11:25:45 crc kubenswrapper[4824]: I0121 11:25:45.063381 4824 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/af5c64ab-b4a4-43f5-81c3-416194e40b9a-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 21 11:25:45 crc kubenswrapper[4824]: I0121 11:25:45.128543 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-gnlcj"] Jan 21 11:25:45 crc kubenswrapper[4824]: W0121 11:25:45.135154 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbb4180b2_5d3e_486d_8197_348d01d2286e.slice/crio-f00978733ebd4e9d3763dd20389763b97d4ebd2a17854c4cfba91b3e800bb66e WatchSource:0}: Error finding container f00978733ebd4e9d3763dd20389763b97d4ebd2a17854c4cfba91b3e800bb66e: Status 404 returned error can't find the container with id f00978733ebd4e9d3763dd20389763b97d4ebd2a17854c4cfba91b3e800bb66e Jan 21 11:25:45 crc kubenswrapper[4824]: I0121 11:25:45.164537 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee7a30d4-dcec-4719-a629-dfaf588d4169-combined-ca-bundle\") pod \"glance-db-sync-sx9lb\" (UID: \"ee7a30d4-dcec-4719-a629-dfaf588d4169\") " pod="openstack/glance-db-sync-sx9lb" Jan 21 11:25:45 crc kubenswrapper[4824]: I0121 11:25:45.164841 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ee7a30d4-dcec-4719-a629-dfaf588d4169-config-data\") pod \"glance-db-sync-sx9lb\" (UID: \"ee7a30d4-dcec-4719-a629-dfaf588d4169\") " pod="openstack/glance-db-sync-sx9lb" Jan 21 11:25:45 crc kubenswrapper[4824]: I0121 11:25:45.164897 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cz8ld\" (UniqueName: \"kubernetes.io/projected/ee7a30d4-dcec-4719-a629-dfaf588d4169-kube-api-access-cz8ld\") pod \"glance-db-sync-sx9lb\" (UID: \"ee7a30d4-dcec-4719-a629-dfaf588d4169\") " pod="openstack/glance-db-sync-sx9lb" Jan 21 11:25:45 crc kubenswrapper[4824]: I0121 11:25:45.164930 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ee7a30d4-dcec-4719-a629-dfaf588d4169-db-sync-config-data\") pod \"glance-db-sync-sx9lb\" (UID: \"ee7a30d4-dcec-4719-a629-dfaf588d4169\") " pod="openstack/glance-db-sync-sx9lb" Jan 21 11:25:45 crc kubenswrapper[4824]: I0121 11:25:45.266595 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee7a30d4-dcec-4719-a629-dfaf588d4169-combined-ca-bundle\") pod \"glance-db-sync-sx9lb\" (UID: \"ee7a30d4-dcec-4719-a629-dfaf588d4169\") " pod="openstack/glance-db-sync-sx9lb" Jan 21 11:25:45 crc kubenswrapper[4824]: I0121 11:25:45.266631 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ee7a30d4-dcec-4719-a629-dfaf588d4169-config-data\") pod \"glance-db-sync-sx9lb\" (UID: \"ee7a30d4-dcec-4719-a629-dfaf588d4169\") " pod="openstack/glance-db-sync-sx9lb" Jan 21 11:25:45 crc kubenswrapper[4824]: I0121 11:25:45.266694 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cz8ld\" (UniqueName: \"kubernetes.io/projected/ee7a30d4-dcec-4719-a629-dfaf588d4169-kube-api-access-cz8ld\") pod \"glance-db-sync-sx9lb\" (UID: \"ee7a30d4-dcec-4719-a629-dfaf588d4169\") " pod="openstack/glance-db-sync-sx9lb" Jan 21 11:25:45 crc kubenswrapper[4824]: I0121 11:25:45.266725 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ee7a30d4-dcec-4719-a629-dfaf588d4169-db-sync-config-data\") pod \"glance-db-sync-sx9lb\" (UID: \"ee7a30d4-dcec-4719-a629-dfaf588d4169\") " pod="openstack/glance-db-sync-sx9lb" Jan 21 11:25:45 crc kubenswrapper[4824]: I0121 11:25:45.271805 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ee7a30d4-dcec-4719-a629-dfaf588d4169-config-data\") pod \"glance-db-sync-sx9lb\" (UID: \"ee7a30d4-dcec-4719-a629-dfaf588d4169\") " pod="openstack/glance-db-sync-sx9lb" Jan 21 11:25:45 crc kubenswrapper[4824]: I0121 11:25:45.271830 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ee7a30d4-dcec-4719-a629-dfaf588d4169-db-sync-config-data\") pod \"glance-db-sync-sx9lb\" (UID: \"ee7a30d4-dcec-4719-a629-dfaf588d4169\") " pod="openstack/glance-db-sync-sx9lb" Jan 21 11:25:45 crc kubenswrapper[4824]: I0121 11:25:45.271844 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee7a30d4-dcec-4719-a629-dfaf588d4169-combined-ca-bundle\") pod \"glance-db-sync-sx9lb\" (UID: \"ee7a30d4-dcec-4719-a629-dfaf588d4169\") " pod="openstack/glance-db-sync-sx9lb" Jan 21 11:25:45 crc kubenswrapper[4824]: I0121 11:25:45.280123 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cz8ld\" (UniqueName: \"kubernetes.io/projected/ee7a30d4-dcec-4719-a629-dfaf588d4169-kube-api-access-cz8ld\") pod \"glance-db-sync-sx9lb\" (UID: \"ee7a30d4-dcec-4719-a629-dfaf588d4169\") " pod="openstack/glance-db-sync-sx9lb" Jan 21 11:25:45 crc kubenswrapper[4824]: I0121 11:25:45.312848 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-sx9lb" Jan 21 11:25:45 crc kubenswrapper[4824]: I0121 11:25:45.367935 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/f980e493-ca74-47f1-8f6f-aca9c9011bf5-etc-swift\") pod \"swift-storage-0\" (UID: \"f980e493-ca74-47f1-8f6f-aca9c9011bf5\") " pod="openstack/swift-storage-0" Jan 21 11:25:45 crc kubenswrapper[4824]: E0121 11:25:45.368106 4824 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Jan 21 11:25:45 crc kubenswrapper[4824]: E0121 11:25:45.368121 4824 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Jan 21 11:25:45 crc kubenswrapper[4824]: E0121 11:25:45.368158 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/f980e493-ca74-47f1-8f6f-aca9c9011bf5-etc-swift podName:f980e493-ca74-47f1-8f6f-aca9c9011bf5 nodeName:}" failed. No retries permitted until 2026-01-21 11:25:53.36814564 +0000 UTC m=+895.661174931 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/f980e493-ca74-47f1-8f6f-aca9c9011bf5-etc-swift") pod "swift-storage-0" (UID: "f980e493-ca74-47f1-8f6f-aca9c9011bf5") : configmap "swift-ring-files" not found Jan 21 11:25:45 crc kubenswrapper[4824]: I0121 11:25:45.747316 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-sx9lb"] Jan 21 11:25:45 crc kubenswrapper[4824]: W0121 11:25:45.749094 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podee7a30d4_dcec_4719_a629_dfaf588d4169.slice/crio-0094c2d2871ffe548bcd39b394fb34d9654678fd14fa5a472f33e3e6e434588c WatchSource:0}: Error finding container 0094c2d2871ffe548bcd39b394fb34d9654678fd14fa5a472f33e3e6e434588c: Status 404 returned error can't find the container with id 0094c2d2871ffe548bcd39b394fb34d9654678fd14fa5a472f33e3e6e434588c Jan 21 11:25:45 crc kubenswrapper[4824]: I0121 11:25:45.767102 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-nc6vf" event={"ID":"29473359-3517-4b86-bd5e-80e25706ff27","Type":"ContainerStarted","Data":"8b9dec71c98fa449139cdcf21064eb0b0fb8a6897189e13bedcb78be35a24b3d"} Jan 21 11:25:45 crc kubenswrapper[4824]: I0121 11:25:45.771251 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-sx9lb" event={"ID":"ee7a30d4-dcec-4719-a629-dfaf588d4169","Type":"ContainerStarted","Data":"0094c2d2871ffe548bcd39b394fb34d9654678fd14fa5a472f33e3e6e434588c"} Jan 21 11:25:45 crc kubenswrapper[4824]: I0121 11:25:45.773466 4824 generic.go:334] "Generic (PLEG): container finished" podID="bb4180b2-5d3e-486d-8197-348d01d2286e" containerID="e050c4db8b50c4f7ba6c6253d674cec54ab71b556af9b68dbd86558b0523c4ab" exitCode=0 Jan 21 11:25:45 crc kubenswrapper[4824]: I0121 11:25:45.773668 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-n665r" Jan 21 11:25:45 crc kubenswrapper[4824]: I0121 11:25:45.774251 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gnlcj" event={"ID":"bb4180b2-5d3e-486d-8197-348d01d2286e","Type":"ContainerDied","Data":"e050c4db8b50c4f7ba6c6253d674cec54ab71b556af9b68dbd86558b0523c4ab"} Jan 21 11:25:45 crc kubenswrapper[4824]: I0121 11:25:45.774351 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gnlcj" event={"ID":"bb4180b2-5d3e-486d-8197-348d01d2286e","Type":"ContainerStarted","Data":"f00978733ebd4e9d3763dd20389763b97d4ebd2a17854c4cfba91b3e800bb66e"} Jan 21 11:25:45 crc kubenswrapper[4824]: I0121 11:25:45.787480 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-nc6vf" podStartSLOduration=2.217822141 podStartE2EDuration="4.787461561s" podCreationTimestamp="2026-01-21 11:25:41 +0000 UTC" firstStartedPulling="2026-01-21 11:25:42.196502349 +0000 UTC m=+884.489531642" lastFinishedPulling="2026-01-21 11:25:44.76614177 +0000 UTC m=+887.059171062" observedRunningTime="2026-01-21 11:25:45.783125788 +0000 UTC m=+888.076155080" watchObservedRunningTime="2026-01-21 11:25:45.787461561 +0000 UTC m=+888.080490853" Jan 21 11:25:46 crc kubenswrapper[4824]: I0121 11:25:46.087157 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-5s6wj"] Jan 21 11:25:46 crc kubenswrapper[4824]: I0121 11:25:46.095275 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5s6wj" Jan 21 11:25:46 crc kubenswrapper[4824]: I0121 11:25:46.099284 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-5s6wj"] Jan 21 11:25:46 crc kubenswrapper[4824]: I0121 11:25:46.179662 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6426f379-b3d5-4589-b5e1-659aed932410-catalog-content\") pod \"redhat-operators-5s6wj\" (UID: \"6426f379-b3d5-4589-b5e1-659aed932410\") " pod="openshift-marketplace/redhat-operators-5s6wj" Jan 21 11:25:46 crc kubenswrapper[4824]: I0121 11:25:46.180011 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6426f379-b3d5-4589-b5e1-659aed932410-utilities\") pod \"redhat-operators-5s6wj\" (UID: \"6426f379-b3d5-4589-b5e1-659aed932410\") " pod="openshift-marketplace/redhat-operators-5s6wj" Jan 21 11:25:46 crc kubenswrapper[4824]: I0121 11:25:46.180222 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rqdmn\" (UniqueName: \"kubernetes.io/projected/6426f379-b3d5-4589-b5e1-659aed932410-kube-api-access-rqdmn\") pod \"redhat-operators-5s6wj\" (UID: \"6426f379-b3d5-4589-b5e1-659aed932410\") " pod="openshift-marketplace/redhat-operators-5s6wj" Jan 21 11:25:46 crc kubenswrapper[4824]: I0121 11:25:46.280989 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6426f379-b3d5-4589-b5e1-659aed932410-catalog-content\") pod \"redhat-operators-5s6wj\" (UID: \"6426f379-b3d5-4589-b5e1-659aed932410\") " pod="openshift-marketplace/redhat-operators-5s6wj" Jan 21 11:25:46 crc kubenswrapper[4824]: I0121 11:25:46.281075 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6426f379-b3d5-4589-b5e1-659aed932410-utilities\") pod \"redhat-operators-5s6wj\" (UID: \"6426f379-b3d5-4589-b5e1-659aed932410\") " pod="openshift-marketplace/redhat-operators-5s6wj" Jan 21 11:25:46 crc kubenswrapper[4824]: I0121 11:25:46.281174 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rqdmn\" (UniqueName: \"kubernetes.io/projected/6426f379-b3d5-4589-b5e1-659aed932410-kube-api-access-rqdmn\") pod \"redhat-operators-5s6wj\" (UID: \"6426f379-b3d5-4589-b5e1-659aed932410\") " pod="openshift-marketplace/redhat-operators-5s6wj" Jan 21 11:25:46 crc kubenswrapper[4824]: I0121 11:25:46.281742 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6426f379-b3d5-4589-b5e1-659aed932410-catalog-content\") pod \"redhat-operators-5s6wj\" (UID: \"6426f379-b3d5-4589-b5e1-659aed932410\") " pod="openshift-marketplace/redhat-operators-5s6wj" Jan 21 11:25:46 crc kubenswrapper[4824]: I0121 11:25:46.281798 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6426f379-b3d5-4589-b5e1-659aed932410-utilities\") pod \"redhat-operators-5s6wj\" (UID: \"6426f379-b3d5-4589-b5e1-659aed932410\") " pod="openshift-marketplace/redhat-operators-5s6wj" Jan 21 11:25:46 crc kubenswrapper[4824]: I0121 11:25:46.297090 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rqdmn\" (UniqueName: \"kubernetes.io/projected/6426f379-b3d5-4589-b5e1-659aed932410-kube-api-access-rqdmn\") pod \"redhat-operators-5s6wj\" (UID: \"6426f379-b3d5-4589-b5e1-659aed932410\") " pod="openshift-marketplace/redhat-operators-5s6wj" Jan 21 11:25:46 crc kubenswrapper[4824]: I0121 11:25:46.416675 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5s6wj" Jan 21 11:25:46 crc kubenswrapper[4824]: I0121 11:25:46.701309 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-698758b865-qzfrf" Jan 21 11:25:46 crc kubenswrapper[4824]: I0121 11:25:46.766768 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-nrqwx"] Jan 21 11:25:46 crc kubenswrapper[4824]: I0121 11:25:46.766937 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-57d769cc4f-nrqwx" podUID="71b408e9-2ba6-4b0e-9b4c-22faf4ef4fd4" containerName="dnsmasq-dns" containerID="cri-o://bdaacbc24f5cf64fe4e60d7403f3ba4bc3f083322e15a6cc88509aa65a02a4b8" gracePeriod=10 Jan 21 11:25:46 crc kubenswrapper[4824]: I0121 11:25:46.793499 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gnlcj" event={"ID":"bb4180b2-5d3e-486d-8197-348d01d2286e","Type":"ContainerStarted","Data":"32f9fcde162e01284d5f8705d4630a345221a7c38111d1bb434ad870a2be305d"} Jan 21 11:25:46 crc kubenswrapper[4824]: I0121 11:25:46.942415 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-5s6wj"] Jan 21 11:25:46 crc kubenswrapper[4824]: W0121 11:25:46.954562 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6426f379_b3d5_4589_b5e1_659aed932410.slice/crio-bc78d731286fa1edc4820267f7da4e9a25da4496ce988c15be5ce2476d00d97c WatchSource:0}: Error finding container bc78d731286fa1edc4820267f7da4e9a25da4496ce988c15be5ce2476d00d97c: Status 404 returned error can't find the container with id bc78d731286fa1edc4820267f7da4e9a25da4496ce988c15be5ce2476d00d97c Jan 21 11:25:47 crc kubenswrapper[4824]: I0121 11:25:47.246598 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-nrqwx" Jan 21 11:25:47 crc kubenswrapper[4824]: I0121 11:25:47.406412 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wcgfn\" (UniqueName: \"kubernetes.io/projected/71b408e9-2ba6-4b0e-9b4c-22faf4ef4fd4-kube-api-access-wcgfn\") pod \"71b408e9-2ba6-4b0e-9b4c-22faf4ef4fd4\" (UID: \"71b408e9-2ba6-4b0e-9b4c-22faf4ef4fd4\") " Jan 21 11:25:47 crc kubenswrapper[4824]: I0121 11:25:47.406772 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/71b408e9-2ba6-4b0e-9b4c-22faf4ef4fd4-dns-svc\") pod \"71b408e9-2ba6-4b0e-9b4c-22faf4ef4fd4\" (UID: \"71b408e9-2ba6-4b0e-9b4c-22faf4ef4fd4\") " Jan 21 11:25:47 crc kubenswrapper[4824]: I0121 11:25:47.406798 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/71b408e9-2ba6-4b0e-9b4c-22faf4ef4fd4-config\") pod \"71b408e9-2ba6-4b0e-9b4c-22faf4ef4fd4\" (UID: \"71b408e9-2ba6-4b0e-9b4c-22faf4ef4fd4\") " Jan 21 11:25:47 crc kubenswrapper[4824]: I0121 11:25:47.412637 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/71b408e9-2ba6-4b0e-9b4c-22faf4ef4fd4-kube-api-access-wcgfn" (OuterVolumeSpecName: "kube-api-access-wcgfn") pod "71b408e9-2ba6-4b0e-9b4c-22faf4ef4fd4" (UID: "71b408e9-2ba6-4b0e-9b4c-22faf4ef4fd4"). InnerVolumeSpecName "kube-api-access-wcgfn". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:25:47 crc kubenswrapper[4824]: I0121 11:25:47.444488 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/71b408e9-2ba6-4b0e-9b4c-22faf4ef4fd4-config" (OuterVolumeSpecName: "config") pod "71b408e9-2ba6-4b0e-9b4c-22faf4ef4fd4" (UID: "71b408e9-2ba6-4b0e-9b4c-22faf4ef4fd4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:25:47 crc kubenswrapper[4824]: I0121 11:25:47.450888 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/71b408e9-2ba6-4b0e-9b4c-22faf4ef4fd4-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "71b408e9-2ba6-4b0e-9b4c-22faf4ef4fd4" (UID: "71b408e9-2ba6-4b0e-9b4c-22faf4ef4fd4"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:25:47 crc kubenswrapper[4824]: I0121 11:25:47.508404 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wcgfn\" (UniqueName: \"kubernetes.io/projected/71b408e9-2ba6-4b0e-9b4c-22faf4ef4fd4-kube-api-access-wcgfn\") on node \"crc\" DevicePath \"\"" Jan 21 11:25:47 crc kubenswrapper[4824]: I0121 11:25:47.508433 4824 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/71b408e9-2ba6-4b0e-9b4c-22faf4ef4fd4-dns-svc\") on node \"crc\" DevicePath \"\"" Jan 21 11:25:47 crc kubenswrapper[4824]: I0121 11:25:47.508442 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/71b408e9-2ba6-4b0e-9b4c-22faf4ef4fd4-config\") on node \"crc\" DevicePath \"\"" Jan 21 11:25:47 crc kubenswrapper[4824]: I0121 11:25:47.799778 4824 generic.go:334] "Generic (PLEG): container finished" podID="6426f379-b3d5-4589-b5e1-659aed932410" containerID="7d27efc53a3e07270c497cc3025ae83bbe01ff00a96faf06c726ec8704aea5d3" exitCode=0 Jan 21 11:25:47 crc kubenswrapper[4824]: I0121 11:25:47.799850 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5s6wj" event={"ID":"6426f379-b3d5-4589-b5e1-659aed932410","Type":"ContainerDied","Data":"7d27efc53a3e07270c497cc3025ae83bbe01ff00a96faf06c726ec8704aea5d3"} Jan 21 11:25:47 crc kubenswrapper[4824]: I0121 11:25:47.799876 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5s6wj" event={"ID":"6426f379-b3d5-4589-b5e1-659aed932410","Type":"ContainerStarted","Data":"bc78d731286fa1edc4820267f7da4e9a25da4496ce988c15be5ce2476d00d97c"} Jan 21 11:25:47 crc kubenswrapper[4824]: I0121 11:25:47.802020 4824 generic.go:334] "Generic (PLEG): container finished" podID="71b408e9-2ba6-4b0e-9b4c-22faf4ef4fd4" containerID="bdaacbc24f5cf64fe4e60d7403f3ba4bc3f083322e15a6cc88509aa65a02a4b8" exitCode=0 Jan 21 11:25:47 crc kubenswrapper[4824]: I0121 11:25:47.802128 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-nrqwx" Jan 21 11:25:47 crc kubenswrapper[4824]: I0121 11:25:47.802147 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-nrqwx" event={"ID":"71b408e9-2ba6-4b0e-9b4c-22faf4ef4fd4","Type":"ContainerDied","Data":"bdaacbc24f5cf64fe4e60d7403f3ba4bc3f083322e15a6cc88509aa65a02a4b8"} Jan 21 11:25:47 crc kubenswrapper[4824]: I0121 11:25:47.802200 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-nrqwx" event={"ID":"71b408e9-2ba6-4b0e-9b4c-22faf4ef4fd4","Type":"ContainerDied","Data":"9f481192efadf26388108ffc156d5c70e082a7630e4e6aa574c82729d2674bf6"} Jan 21 11:25:47 crc kubenswrapper[4824]: I0121 11:25:47.802219 4824 scope.go:117] "RemoveContainer" containerID="bdaacbc24f5cf64fe4e60d7403f3ba4bc3f083322e15a6cc88509aa65a02a4b8" Jan 21 11:25:47 crc kubenswrapper[4824]: I0121 11:25:47.810141 4824 generic.go:334] "Generic (PLEG): container finished" podID="bb4180b2-5d3e-486d-8197-348d01d2286e" containerID="32f9fcde162e01284d5f8705d4630a345221a7c38111d1bb434ad870a2be305d" exitCode=0 Jan 21 11:25:47 crc kubenswrapper[4824]: I0121 11:25:47.810186 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gnlcj" event={"ID":"bb4180b2-5d3e-486d-8197-348d01d2286e","Type":"ContainerDied","Data":"32f9fcde162e01284d5f8705d4630a345221a7c38111d1bb434ad870a2be305d"} Jan 21 11:25:47 crc kubenswrapper[4824]: I0121 11:25:47.830775 4824 scope.go:117] "RemoveContainer" containerID="cd51f4db187c81d9d459bc61d2b5fff2b45f07ab86bc1d1f49ffc846a1036059" Jan 21 11:25:47 crc kubenswrapper[4824]: I0121 11:25:47.856465 4824 scope.go:117] "RemoveContainer" containerID="bdaacbc24f5cf64fe4e60d7403f3ba4bc3f083322e15a6cc88509aa65a02a4b8" Jan 21 11:25:47 crc kubenswrapper[4824]: E0121 11:25:47.857027 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bdaacbc24f5cf64fe4e60d7403f3ba4bc3f083322e15a6cc88509aa65a02a4b8\": container with ID starting with bdaacbc24f5cf64fe4e60d7403f3ba4bc3f083322e15a6cc88509aa65a02a4b8 not found: ID does not exist" containerID="bdaacbc24f5cf64fe4e60d7403f3ba4bc3f083322e15a6cc88509aa65a02a4b8" Jan 21 11:25:47 crc kubenswrapper[4824]: I0121 11:25:47.857063 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bdaacbc24f5cf64fe4e60d7403f3ba4bc3f083322e15a6cc88509aa65a02a4b8"} err="failed to get container status \"bdaacbc24f5cf64fe4e60d7403f3ba4bc3f083322e15a6cc88509aa65a02a4b8\": rpc error: code = NotFound desc = could not find container \"bdaacbc24f5cf64fe4e60d7403f3ba4bc3f083322e15a6cc88509aa65a02a4b8\": container with ID starting with bdaacbc24f5cf64fe4e60d7403f3ba4bc3f083322e15a6cc88509aa65a02a4b8 not found: ID does not exist" Jan 21 11:25:47 crc kubenswrapper[4824]: I0121 11:25:47.857084 4824 scope.go:117] "RemoveContainer" containerID="cd51f4db187c81d9d459bc61d2b5fff2b45f07ab86bc1d1f49ffc846a1036059" Jan 21 11:25:47 crc kubenswrapper[4824]: E0121 11:25:47.857430 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cd51f4db187c81d9d459bc61d2b5fff2b45f07ab86bc1d1f49ffc846a1036059\": container with ID starting with cd51f4db187c81d9d459bc61d2b5fff2b45f07ab86bc1d1f49ffc846a1036059 not found: ID does not exist" containerID="cd51f4db187c81d9d459bc61d2b5fff2b45f07ab86bc1d1f49ffc846a1036059" Jan 21 11:25:47 crc kubenswrapper[4824]: I0121 11:25:47.857456 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cd51f4db187c81d9d459bc61d2b5fff2b45f07ab86bc1d1f49ffc846a1036059"} err="failed to get container status \"cd51f4db187c81d9d459bc61d2b5fff2b45f07ab86bc1d1f49ffc846a1036059\": rpc error: code = NotFound desc = could not find container \"cd51f4db187c81d9d459bc61d2b5fff2b45f07ab86bc1d1f49ffc846a1036059\": container with ID starting with cd51f4db187c81d9d459bc61d2b5fff2b45f07ab86bc1d1f49ffc846a1036059 not found: ID does not exist" Jan 21 11:25:47 crc kubenswrapper[4824]: I0121 11:25:47.859414 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-nrqwx"] Jan 21 11:25:47 crc kubenswrapper[4824]: I0121 11:25:47.864986 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-nrqwx"] Jan 21 11:25:48 crc kubenswrapper[4824]: I0121 11:25:48.000209 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/root-account-create-update-n665r"] Jan 21 11:25:48 crc kubenswrapper[4824]: I0121 11:25:48.004652 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/root-account-create-update-n665r"] Jan 21 11:25:48 crc kubenswrapper[4824]: I0121 11:25:48.056567 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="71b408e9-2ba6-4b0e-9b4c-22faf4ef4fd4" path="/var/lib/kubelet/pods/71b408e9-2ba6-4b0e-9b4c-22faf4ef4fd4/volumes" Jan 21 11:25:48 crc kubenswrapper[4824]: I0121 11:25:48.057138 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="af5c64ab-b4a4-43f5-81c3-416194e40b9a" path="/var/lib/kubelet/pods/af5c64ab-b4a4-43f5-81c3-416194e40b9a/volumes" Jan 21 11:25:48 crc kubenswrapper[4824]: I0121 11:25:48.820156 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gnlcj" event={"ID":"bb4180b2-5d3e-486d-8197-348d01d2286e","Type":"ContainerStarted","Data":"39ae973d44d7313b6a5dc308f01a2d8f8528e558a2117324b0a478bdce629279"} Jan 21 11:25:48 crc kubenswrapper[4824]: I0121 11:25:48.822414 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5s6wj" event={"ID":"6426f379-b3d5-4589-b5e1-659aed932410","Type":"ContainerStarted","Data":"44fe649276096a6e22d13f5809a21cfaa7a7d7fb3e877b6985edd5b4017b012c"} Jan 21 11:25:48 crc kubenswrapper[4824]: I0121 11:25:48.843196 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-gnlcj" podStartSLOduration=3.259282191 podStartE2EDuration="5.843179837s" podCreationTimestamp="2026-01-21 11:25:43 +0000 UTC" firstStartedPulling="2026-01-21 11:25:45.775043593 +0000 UTC m=+888.068072885" lastFinishedPulling="2026-01-21 11:25:48.358941239 +0000 UTC m=+890.651970531" observedRunningTime="2026-01-21 11:25:48.840207994 +0000 UTC m=+891.133237286" watchObservedRunningTime="2026-01-21 11:25:48.843179837 +0000 UTC m=+891.136209129" Jan 21 11:25:49 crc kubenswrapper[4824]: I0121 11:25:49.840837 4824 generic.go:334] "Generic (PLEG): container finished" podID="6426f379-b3d5-4589-b5e1-659aed932410" containerID="44fe649276096a6e22d13f5809a21cfaa7a7d7fb3e877b6985edd5b4017b012c" exitCode=0 Jan 21 11:25:49 crc kubenswrapper[4824]: I0121 11:25:49.840993 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5s6wj" event={"ID":"6426f379-b3d5-4589-b5e1-659aed932410","Type":"ContainerDied","Data":"44fe649276096a6e22d13f5809a21cfaa7a7d7fb3e877b6985edd5b4017b012c"} Jan 21 11:25:50 crc kubenswrapper[4824]: I0121 11:25:50.848664 4824 generic.go:334] "Generic (PLEG): container finished" podID="29473359-3517-4b86-bd5e-80e25706ff27" containerID="8b9dec71c98fa449139cdcf21064eb0b0fb8a6897189e13bedcb78be35a24b3d" exitCode=0 Jan 21 11:25:50 crc kubenswrapper[4824]: I0121 11:25:50.848772 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-nc6vf" event={"ID":"29473359-3517-4b86-bd5e-80e25706ff27","Type":"ContainerDied","Data":"8b9dec71c98fa449139cdcf21064eb0b0fb8a6897189e13bedcb78be35a24b3d"} Jan 21 11:25:50 crc kubenswrapper[4824]: I0121 11:25:50.856625 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5s6wj" event={"ID":"6426f379-b3d5-4589-b5e1-659aed932410","Type":"ContainerStarted","Data":"828858ab38aa4d0e3c2ec1fe8be4d8444b2d96a181e3ed3dd9412225aeb16e0e"} Jan 21 11:25:50 crc kubenswrapper[4824]: I0121 11:25:50.877486 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-5s6wj" podStartSLOduration=2.37313704 podStartE2EDuration="4.877472504s" podCreationTimestamp="2026-01-21 11:25:46 +0000 UTC" firstStartedPulling="2026-01-21 11:25:47.801641395 +0000 UTC m=+890.094670688" lastFinishedPulling="2026-01-21 11:25:50.305976861 +0000 UTC m=+892.599006152" observedRunningTime="2026-01-21 11:25:50.87291368 +0000 UTC m=+893.165942973" watchObservedRunningTime="2026-01-21 11:25:50.877472504 +0000 UTC m=+893.170501795" Jan 21 11:25:51 crc kubenswrapper[4824]: I0121 11:25:51.210202 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Jan 21 11:25:53 crc kubenswrapper[4824]: I0121 11:25:53.008700 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/root-account-create-update-m6rpv"] Jan 21 11:25:53 crc kubenswrapper[4824]: E0121 11:25:53.009206 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="71b408e9-2ba6-4b0e-9b4c-22faf4ef4fd4" containerName="dnsmasq-dns" Jan 21 11:25:53 crc kubenswrapper[4824]: I0121 11:25:53.009218 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="71b408e9-2ba6-4b0e-9b4c-22faf4ef4fd4" containerName="dnsmasq-dns" Jan 21 11:25:53 crc kubenswrapper[4824]: E0121 11:25:53.009229 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="71b408e9-2ba6-4b0e-9b4c-22faf4ef4fd4" containerName="init" Jan 21 11:25:53 crc kubenswrapper[4824]: I0121 11:25:53.009235 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="71b408e9-2ba6-4b0e-9b4c-22faf4ef4fd4" containerName="init" Jan 21 11:25:53 crc kubenswrapper[4824]: I0121 11:25:53.009394 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="71b408e9-2ba6-4b0e-9b4c-22faf4ef4fd4" containerName="dnsmasq-dns" Jan 21 11:25:53 crc kubenswrapper[4824]: I0121 11:25:53.009871 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-m6rpv" Jan 21 11:25:53 crc kubenswrapper[4824]: I0121 11:25:53.011619 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-mariadb-root-db-secret" Jan 21 11:25:53 crc kubenswrapper[4824]: I0121 11:25:53.023341 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-m6rpv"] Jan 21 11:25:53 crc kubenswrapper[4824]: I0121 11:25:53.097270 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2p2xd\" (UniqueName: \"kubernetes.io/projected/76e33763-1a41-4640-8e4a-492ab92009e8-kube-api-access-2p2xd\") pod \"root-account-create-update-m6rpv\" (UID: \"76e33763-1a41-4640-8e4a-492ab92009e8\") " pod="openstack/root-account-create-update-m6rpv" Jan 21 11:25:53 crc kubenswrapper[4824]: I0121 11:25:53.097437 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/76e33763-1a41-4640-8e4a-492ab92009e8-operator-scripts\") pod \"root-account-create-update-m6rpv\" (UID: \"76e33763-1a41-4640-8e4a-492ab92009e8\") " pod="openstack/root-account-create-update-m6rpv" Jan 21 11:25:53 crc kubenswrapper[4824]: I0121 11:25:53.200988 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2p2xd\" (UniqueName: \"kubernetes.io/projected/76e33763-1a41-4640-8e4a-492ab92009e8-kube-api-access-2p2xd\") pod \"root-account-create-update-m6rpv\" (UID: \"76e33763-1a41-4640-8e4a-492ab92009e8\") " pod="openstack/root-account-create-update-m6rpv" Jan 21 11:25:53 crc kubenswrapper[4824]: I0121 11:25:53.201119 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/76e33763-1a41-4640-8e4a-492ab92009e8-operator-scripts\") pod \"root-account-create-update-m6rpv\" (UID: \"76e33763-1a41-4640-8e4a-492ab92009e8\") " pod="openstack/root-account-create-update-m6rpv" Jan 21 11:25:53 crc kubenswrapper[4824]: I0121 11:25:53.203508 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/76e33763-1a41-4640-8e4a-492ab92009e8-operator-scripts\") pod \"root-account-create-update-m6rpv\" (UID: \"76e33763-1a41-4640-8e4a-492ab92009e8\") " pod="openstack/root-account-create-update-m6rpv" Jan 21 11:25:53 crc kubenswrapper[4824]: I0121 11:25:53.223180 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2p2xd\" (UniqueName: \"kubernetes.io/projected/76e33763-1a41-4640-8e4a-492ab92009e8-kube-api-access-2p2xd\") pod \"root-account-create-update-m6rpv\" (UID: \"76e33763-1a41-4640-8e4a-492ab92009e8\") " pod="openstack/root-account-create-update-m6rpv" Jan 21 11:25:53 crc kubenswrapper[4824]: I0121 11:25:53.331142 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-m6rpv" Jan 21 11:25:53 crc kubenswrapper[4824]: I0121 11:25:53.403526 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/f980e493-ca74-47f1-8f6f-aca9c9011bf5-etc-swift\") pod \"swift-storage-0\" (UID: \"f980e493-ca74-47f1-8f6f-aca9c9011bf5\") " pod="openstack/swift-storage-0" Jan 21 11:25:53 crc kubenswrapper[4824]: I0121 11:25:53.408515 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/f980e493-ca74-47f1-8f6f-aca9c9011bf5-etc-swift\") pod \"swift-storage-0\" (UID: \"f980e493-ca74-47f1-8f6f-aca9c9011bf5\") " pod="openstack/swift-storage-0" Jan 21 11:25:53 crc kubenswrapper[4824]: I0121 11:25:53.494589 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Jan 21 11:25:53 crc kubenswrapper[4824]: I0121 11:25:53.624121 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-gnlcj" Jan 21 11:25:53 crc kubenswrapper[4824]: I0121 11:25:53.624163 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-gnlcj" Jan 21 11:25:53 crc kubenswrapper[4824]: I0121 11:25:53.658143 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-gnlcj" Jan 21 11:25:53 crc kubenswrapper[4824]: I0121 11:25:53.908016 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-gnlcj" Jan 21 11:25:55 crc kubenswrapper[4824]: I0121 11:25:55.476308 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-gnlcj"] Jan 21 11:25:55 crc kubenswrapper[4824]: I0121 11:25:55.887746 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-gnlcj" podUID="bb4180b2-5d3e-486d-8197-348d01d2286e" containerName="registry-server" containerID="cri-o://39ae973d44d7313b6a5dc308f01a2d8f8528e558a2117324b0a478bdce629279" gracePeriod=2 Jan 21 11:25:56 crc kubenswrapper[4824]: I0121 11:25:56.132736 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-nc6vf" Jan 21 11:25:56 crc kubenswrapper[4824]: I0121 11:25:56.254895 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/29473359-3517-4b86-bd5e-80e25706ff27-dispersionconf\") pod \"29473359-3517-4b86-bd5e-80e25706ff27\" (UID: \"29473359-3517-4b86-bd5e-80e25706ff27\") " Jan 21 11:25:56 crc kubenswrapper[4824]: I0121 11:25:56.254938 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cffv8\" (UniqueName: \"kubernetes.io/projected/29473359-3517-4b86-bd5e-80e25706ff27-kube-api-access-cffv8\") pod \"29473359-3517-4b86-bd5e-80e25706ff27\" (UID: \"29473359-3517-4b86-bd5e-80e25706ff27\") " Jan 21 11:25:56 crc kubenswrapper[4824]: I0121 11:25:56.254993 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/29473359-3517-4b86-bd5e-80e25706ff27-ring-data-devices\") pod \"29473359-3517-4b86-bd5e-80e25706ff27\" (UID: \"29473359-3517-4b86-bd5e-80e25706ff27\") " Jan 21 11:25:56 crc kubenswrapper[4824]: I0121 11:25:56.255016 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/29473359-3517-4b86-bd5e-80e25706ff27-etc-swift\") pod \"29473359-3517-4b86-bd5e-80e25706ff27\" (UID: \"29473359-3517-4b86-bd5e-80e25706ff27\") " Jan 21 11:25:56 crc kubenswrapper[4824]: I0121 11:25:56.255046 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/29473359-3517-4b86-bd5e-80e25706ff27-swiftconf\") pod \"29473359-3517-4b86-bd5e-80e25706ff27\" (UID: \"29473359-3517-4b86-bd5e-80e25706ff27\") " Jan 21 11:25:56 crc kubenswrapper[4824]: I0121 11:25:56.255061 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/29473359-3517-4b86-bd5e-80e25706ff27-scripts\") pod \"29473359-3517-4b86-bd5e-80e25706ff27\" (UID: \"29473359-3517-4b86-bd5e-80e25706ff27\") " Jan 21 11:25:56 crc kubenswrapper[4824]: I0121 11:25:56.255136 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29473359-3517-4b86-bd5e-80e25706ff27-combined-ca-bundle\") pod \"29473359-3517-4b86-bd5e-80e25706ff27\" (UID: \"29473359-3517-4b86-bd5e-80e25706ff27\") " Jan 21 11:25:56 crc kubenswrapper[4824]: I0121 11:25:56.256536 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/29473359-3517-4b86-bd5e-80e25706ff27-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "29473359-3517-4b86-bd5e-80e25706ff27" (UID: "29473359-3517-4b86-bd5e-80e25706ff27"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:25:56 crc kubenswrapper[4824]: I0121 11:25:56.259306 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/29473359-3517-4b86-bd5e-80e25706ff27-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "29473359-3517-4b86-bd5e-80e25706ff27" (UID: "29473359-3517-4b86-bd5e-80e25706ff27"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:25:56 crc kubenswrapper[4824]: I0121 11:25:56.261996 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/29473359-3517-4b86-bd5e-80e25706ff27-kube-api-access-cffv8" (OuterVolumeSpecName: "kube-api-access-cffv8") pod "29473359-3517-4b86-bd5e-80e25706ff27" (UID: "29473359-3517-4b86-bd5e-80e25706ff27"). InnerVolumeSpecName "kube-api-access-cffv8". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:25:56 crc kubenswrapper[4824]: I0121 11:25:56.262971 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/29473359-3517-4b86-bd5e-80e25706ff27-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "29473359-3517-4b86-bd5e-80e25706ff27" (UID: "29473359-3517-4b86-bd5e-80e25706ff27"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:25:56 crc kubenswrapper[4824]: I0121 11:25:56.274363 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/29473359-3517-4b86-bd5e-80e25706ff27-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "29473359-3517-4b86-bd5e-80e25706ff27" (UID: "29473359-3517-4b86-bd5e-80e25706ff27"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:25:56 crc kubenswrapper[4824]: I0121 11:25:56.274604 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/29473359-3517-4b86-bd5e-80e25706ff27-scripts" (OuterVolumeSpecName: "scripts") pod "29473359-3517-4b86-bd5e-80e25706ff27" (UID: "29473359-3517-4b86-bd5e-80e25706ff27"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:25:56 crc kubenswrapper[4824]: I0121 11:25:56.278934 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gnlcj" Jan 21 11:25:56 crc kubenswrapper[4824]: I0121 11:25:56.281399 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/29473359-3517-4b86-bd5e-80e25706ff27-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "29473359-3517-4b86-bd5e-80e25706ff27" (UID: "29473359-3517-4b86-bd5e-80e25706ff27"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:25:56 crc kubenswrapper[4824]: I0121 11:25:56.356734 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6mmv9\" (UniqueName: \"kubernetes.io/projected/bb4180b2-5d3e-486d-8197-348d01d2286e-kube-api-access-6mmv9\") pod \"bb4180b2-5d3e-486d-8197-348d01d2286e\" (UID: \"bb4180b2-5d3e-486d-8197-348d01d2286e\") " Jan 21 11:25:56 crc kubenswrapper[4824]: I0121 11:25:56.356827 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bb4180b2-5d3e-486d-8197-348d01d2286e-catalog-content\") pod \"bb4180b2-5d3e-486d-8197-348d01d2286e\" (UID: \"bb4180b2-5d3e-486d-8197-348d01d2286e\") " Jan 21 11:25:56 crc kubenswrapper[4824]: I0121 11:25:56.356853 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bb4180b2-5d3e-486d-8197-348d01d2286e-utilities\") pod \"bb4180b2-5d3e-486d-8197-348d01d2286e\" (UID: \"bb4180b2-5d3e-486d-8197-348d01d2286e\") " Jan 21 11:25:56 crc kubenswrapper[4824]: I0121 11:25:56.357309 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29473359-3517-4b86-bd5e-80e25706ff27-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 21 11:25:56 crc kubenswrapper[4824]: I0121 11:25:56.357326 4824 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/29473359-3517-4b86-bd5e-80e25706ff27-dispersionconf\") on node \"crc\" DevicePath \"\"" Jan 21 11:25:56 crc kubenswrapper[4824]: I0121 11:25:56.357335 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cffv8\" (UniqueName: \"kubernetes.io/projected/29473359-3517-4b86-bd5e-80e25706ff27-kube-api-access-cffv8\") on node \"crc\" DevicePath \"\"" Jan 21 11:25:56 crc kubenswrapper[4824]: I0121 11:25:56.357343 4824 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/29473359-3517-4b86-bd5e-80e25706ff27-ring-data-devices\") on node \"crc\" DevicePath \"\"" Jan 21 11:25:56 crc kubenswrapper[4824]: I0121 11:25:56.357351 4824 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/29473359-3517-4b86-bd5e-80e25706ff27-etc-swift\") on node \"crc\" DevicePath \"\"" Jan 21 11:25:56 crc kubenswrapper[4824]: I0121 11:25:56.357359 4824 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/29473359-3517-4b86-bd5e-80e25706ff27-swiftconf\") on node \"crc\" DevicePath \"\"" Jan 21 11:25:56 crc kubenswrapper[4824]: I0121 11:25:56.357367 4824 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/29473359-3517-4b86-bd5e-80e25706ff27-scripts\") on node \"crc\" DevicePath \"\"" Jan 21 11:25:56 crc kubenswrapper[4824]: I0121 11:25:56.357930 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bb4180b2-5d3e-486d-8197-348d01d2286e-utilities" (OuterVolumeSpecName: "utilities") pod "bb4180b2-5d3e-486d-8197-348d01d2286e" (UID: "bb4180b2-5d3e-486d-8197-348d01d2286e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:25:56 crc kubenswrapper[4824]: I0121 11:25:56.359366 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bb4180b2-5d3e-486d-8197-348d01d2286e-kube-api-access-6mmv9" (OuterVolumeSpecName: "kube-api-access-6mmv9") pod "bb4180b2-5d3e-486d-8197-348d01d2286e" (UID: "bb4180b2-5d3e-486d-8197-348d01d2286e"). InnerVolumeSpecName "kube-api-access-6mmv9". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:25:56 crc kubenswrapper[4824]: I0121 11:25:56.396377 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-m6rpv"] Jan 21 11:25:56 crc kubenswrapper[4824]: I0121 11:25:56.403339 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bb4180b2-5d3e-486d-8197-348d01d2286e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "bb4180b2-5d3e-486d-8197-348d01d2286e" (UID: "bb4180b2-5d3e-486d-8197-348d01d2286e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:25:56 crc kubenswrapper[4824]: W0121 11:25:56.408484 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod76e33763_1a41_4640_8e4a_492ab92009e8.slice/crio-78f8c0c168ea13f1ae2d548e6e02a4143f0dde836f9d6fb9f2378253323b9a7a WatchSource:0}: Error finding container 78f8c0c168ea13f1ae2d548e6e02a4143f0dde836f9d6fb9f2378253323b9a7a: Status 404 returned error can't find the container with id 78f8c0c168ea13f1ae2d548e6e02a4143f0dde836f9d6fb9f2378253323b9a7a Jan 21 11:25:56 crc kubenswrapper[4824]: I0121 11:25:56.417830 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-5s6wj" Jan 21 11:25:56 crc kubenswrapper[4824]: I0121 11:25:56.417867 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-5s6wj" Jan 21 11:25:56 crc kubenswrapper[4824]: I0121 11:25:56.459231 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bb4180b2-5d3e-486d-8197-348d01d2286e-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 21 11:25:56 crc kubenswrapper[4824]: I0121 11:25:56.459406 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bb4180b2-5d3e-486d-8197-348d01d2286e-utilities\") on node \"crc\" DevicePath \"\"" Jan 21 11:25:56 crc kubenswrapper[4824]: I0121 11:25:56.459417 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6mmv9\" (UniqueName: \"kubernetes.io/projected/bb4180b2-5d3e-486d-8197-348d01d2286e-kube-api-access-6mmv9\") on node \"crc\" DevicePath \"\"" Jan 21 11:25:56 crc kubenswrapper[4824]: I0121 11:25:56.460433 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-5s6wj" Jan 21 11:25:56 crc kubenswrapper[4824]: I0121 11:25:56.480925 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Jan 21 11:25:56 crc kubenswrapper[4824]: I0121 11:25:56.493613 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-b795q"] Jan 21 11:25:56 crc kubenswrapper[4824]: E0121 11:25:56.493913 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb4180b2-5d3e-486d-8197-348d01d2286e" containerName="extract-utilities" Jan 21 11:25:56 crc kubenswrapper[4824]: I0121 11:25:56.493934 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb4180b2-5d3e-486d-8197-348d01d2286e" containerName="extract-utilities" Jan 21 11:25:56 crc kubenswrapper[4824]: E0121 11:25:56.493949 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="29473359-3517-4b86-bd5e-80e25706ff27" containerName="swift-ring-rebalance" Jan 21 11:25:56 crc kubenswrapper[4824]: I0121 11:25:56.493968 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="29473359-3517-4b86-bd5e-80e25706ff27" containerName="swift-ring-rebalance" Jan 21 11:25:56 crc kubenswrapper[4824]: E0121 11:25:56.493978 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb4180b2-5d3e-486d-8197-348d01d2286e" containerName="extract-content" Jan 21 11:25:56 crc kubenswrapper[4824]: I0121 11:25:56.493984 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb4180b2-5d3e-486d-8197-348d01d2286e" containerName="extract-content" Jan 21 11:25:56 crc kubenswrapper[4824]: E0121 11:25:56.494006 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb4180b2-5d3e-486d-8197-348d01d2286e" containerName="registry-server" Jan 21 11:25:56 crc kubenswrapper[4824]: I0121 11:25:56.494012 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb4180b2-5d3e-486d-8197-348d01d2286e" containerName="registry-server" Jan 21 11:25:56 crc kubenswrapper[4824]: I0121 11:25:56.494145 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="bb4180b2-5d3e-486d-8197-348d01d2286e" containerName="registry-server" Jan 21 11:25:56 crc kubenswrapper[4824]: I0121 11:25:56.494163 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="29473359-3517-4b86-bd5e-80e25706ff27" containerName="swift-ring-rebalance" Jan 21 11:25:56 crc kubenswrapper[4824]: I0121 11:25:56.496432 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-b795q" Jan 21 11:25:56 crc kubenswrapper[4824]: I0121 11:25:56.499309 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-b795q"] Jan 21 11:25:56 crc kubenswrapper[4824]: I0121 11:25:56.662210 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1c75f0f9-53e1-4a64-9211-559ef51c1b53-catalog-content\") pod \"redhat-marketplace-b795q\" (UID: \"1c75f0f9-53e1-4a64-9211-559ef51c1b53\") " pod="openshift-marketplace/redhat-marketplace-b795q" Jan 21 11:25:56 crc kubenswrapper[4824]: I0121 11:25:56.662340 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hd85x\" (UniqueName: \"kubernetes.io/projected/1c75f0f9-53e1-4a64-9211-559ef51c1b53-kube-api-access-hd85x\") pod \"redhat-marketplace-b795q\" (UID: \"1c75f0f9-53e1-4a64-9211-559ef51c1b53\") " pod="openshift-marketplace/redhat-marketplace-b795q" Jan 21 11:25:56 crc kubenswrapper[4824]: I0121 11:25:56.662497 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1c75f0f9-53e1-4a64-9211-559ef51c1b53-utilities\") pod \"redhat-marketplace-b795q\" (UID: \"1c75f0f9-53e1-4a64-9211-559ef51c1b53\") " pod="openshift-marketplace/redhat-marketplace-b795q" Jan 21 11:25:56 crc kubenswrapper[4824]: I0121 11:25:56.763811 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1c75f0f9-53e1-4a64-9211-559ef51c1b53-catalog-content\") pod \"redhat-marketplace-b795q\" (UID: \"1c75f0f9-53e1-4a64-9211-559ef51c1b53\") " pod="openshift-marketplace/redhat-marketplace-b795q" Jan 21 11:25:56 crc kubenswrapper[4824]: I0121 11:25:56.763921 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hd85x\" (UniqueName: \"kubernetes.io/projected/1c75f0f9-53e1-4a64-9211-559ef51c1b53-kube-api-access-hd85x\") pod \"redhat-marketplace-b795q\" (UID: \"1c75f0f9-53e1-4a64-9211-559ef51c1b53\") " pod="openshift-marketplace/redhat-marketplace-b795q" Jan 21 11:25:56 crc kubenswrapper[4824]: I0121 11:25:56.764037 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1c75f0f9-53e1-4a64-9211-559ef51c1b53-utilities\") pod \"redhat-marketplace-b795q\" (UID: \"1c75f0f9-53e1-4a64-9211-559ef51c1b53\") " pod="openshift-marketplace/redhat-marketplace-b795q" Jan 21 11:25:56 crc kubenswrapper[4824]: I0121 11:25:56.764311 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1c75f0f9-53e1-4a64-9211-559ef51c1b53-catalog-content\") pod \"redhat-marketplace-b795q\" (UID: \"1c75f0f9-53e1-4a64-9211-559ef51c1b53\") " pod="openshift-marketplace/redhat-marketplace-b795q" Jan 21 11:25:56 crc kubenswrapper[4824]: I0121 11:25:56.764414 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1c75f0f9-53e1-4a64-9211-559ef51c1b53-utilities\") pod \"redhat-marketplace-b795q\" (UID: \"1c75f0f9-53e1-4a64-9211-559ef51c1b53\") " pod="openshift-marketplace/redhat-marketplace-b795q" Jan 21 11:25:56 crc kubenswrapper[4824]: I0121 11:25:56.781746 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hd85x\" (UniqueName: \"kubernetes.io/projected/1c75f0f9-53e1-4a64-9211-559ef51c1b53-kube-api-access-hd85x\") pod \"redhat-marketplace-b795q\" (UID: \"1c75f0f9-53e1-4a64-9211-559ef51c1b53\") " pod="openshift-marketplace/redhat-marketplace-b795q" Jan 21 11:25:56 crc kubenswrapper[4824]: I0121 11:25:56.829888 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-b795q" Jan 21 11:25:56 crc kubenswrapper[4824]: I0121 11:25:56.917313 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-nc6vf" Jan 21 11:25:56 crc kubenswrapper[4824]: I0121 11:25:56.918704 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-nc6vf" event={"ID":"29473359-3517-4b86-bd5e-80e25706ff27","Type":"ContainerDied","Data":"be8ab46a5292e712c0ba8d486b130fdd73752ef5b1cc0ad9789ef7c0c22e7dd0"} Jan 21 11:25:56 crc kubenswrapper[4824]: I0121 11:25:56.918741 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="be8ab46a5292e712c0ba8d486b130fdd73752ef5b1cc0ad9789ef7c0c22e7dd0" Jan 21 11:25:56 crc kubenswrapper[4824]: I0121 11:25:56.925326 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-sx9lb" event={"ID":"ee7a30d4-dcec-4719-a629-dfaf588d4169","Type":"ContainerStarted","Data":"2b4aa0578cb659bd1fe3cb0351e682b347beae954c60ccf215edaf9c2da1abd4"} Jan 21 11:25:56 crc kubenswrapper[4824]: I0121 11:25:56.935777 4824 generic.go:334] "Generic (PLEG): container finished" podID="bb4180b2-5d3e-486d-8197-348d01d2286e" containerID="39ae973d44d7313b6a5dc308f01a2d8f8528e558a2117324b0a478bdce629279" exitCode=0 Jan 21 11:25:56 crc kubenswrapper[4824]: I0121 11:25:56.935829 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gnlcj" event={"ID":"bb4180b2-5d3e-486d-8197-348d01d2286e","Type":"ContainerDied","Data":"39ae973d44d7313b6a5dc308f01a2d8f8528e558a2117324b0a478bdce629279"} Jan 21 11:25:56 crc kubenswrapper[4824]: I0121 11:25:56.935852 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gnlcj" event={"ID":"bb4180b2-5d3e-486d-8197-348d01d2286e","Type":"ContainerDied","Data":"f00978733ebd4e9d3763dd20389763b97d4ebd2a17854c4cfba91b3e800bb66e"} Jan 21 11:25:56 crc kubenswrapper[4824]: I0121 11:25:56.935867 4824 scope.go:117] "RemoveContainer" containerID="39ae973d44d7313b6a5dc308f01a2d8f8528e558a2117324b0a478bdce629279" Jan 21 11:25:56 crc kubenswrapper[4824]: I0121 11:25:56.936242 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gnlcj" Jan 21 11:25:56 crc kubenswrapper[4824]: I0121 11:25:56.942879 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-sx9lb" podStartSLOduration=2.685972681 podStartE2EDuration="12.942863451s" podCreationTimestamp="2026-01-21 11:25:44 +0000 UTC" firstStartedPulling="2026-01-21 11:25:45.752781961 +0000 UTC m=+888.045811253" lastFinishedPulling="2026-01-21 11:25:56.009672731 +0000 UTC m=+898.302702023" observedRunningTime="2026-01-21 11:25:56.940300941 +0000 UTC m=+899.233330233" watchObservedRunningTime="2026-01-21 11:25:56.942863451 +0000 UTC m=+899.235892743" Jan 21 11:25:56 crc kubenswrapper[4824]: I0121 11:25:56.945578 4824 generic.go:334] "Generic (PLEG): container finished" podID="76e33763-1a41-4640-8e4a-492ab92009e8" containerID="25e2c38241236d848b515e556d4747f5cb7cbca9e5eba1926722e4f2e38b424d" exitCode=0 Jan 21 11:25:56 crc kubenswrapper[4824]: I0121 11:25:56.945626 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-m6rpv" event={"ID":"76e33763-1a41-4640-8e4a-492ab92009e8","Type":"ContainerDied","Data":"25e2c38241236d848b515e556d4747f5cb7cbca9e5eba1926722e4f2e38b424d"} Jan 21 11:25:56 crc kubenswrapper[4824]: I0121 11:25:56.945648 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-m6rpv" event={"ID":"76e33763-1a41-4640-8e4a-492ab92009e8","Type":"ContainerStarted","Data":"78f8c0c168ea13f1ae2d548e6e02a4143f0dde836f9d6fb9f2378253323b9a7a"} Jan 21 11:25:56 crc kubenswrapper[4824]: I0121 11:25:56.949051 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"f980e493-ca74-47f1-8f6f-aca9c9011bf5","Type":"ContainerStarted","Data":"5ce1bb7e7ea8646e76b785d5f245fdc4e2efc04e0360410adebd6bb17d781b28"} Jan 21 11:25:56 crc kubenswrapper[4824]: I0121 11:25:56.990190 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-5s6wj" Jan 21 11:25:57 crc kubenswrapper[4824]: I0121 11:25:57.116500 4824 scope.go:117] "RemoveContainer" containerID="32f9fcde162e01284d5f8705d4630a345221a7c38111d1bb434ad870a2be305d" Jan 21 11:25:57 crc kubenswrapper[4824]: I0121 11:25:57.124784 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-gnlcj"] Jan 21 11:25:57 crc kubenswrapper[4824]: I0121 11:25:57.131811 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-gnlcj"] Jan 21 11:25:57 crc kubenswrapper[4824]: I0121 11:25:57.147709 4824 scope.go:117] "RemoveContainer" containerID="e050c4db8b50c4f7ba6c6253d674cec54ab71b556af9b68dbd86558b0523c4ab" Jan 21 11:25:57 crc kubenswrapper[4824]: I0121 11:25:57.166689 4824 scope.go:117] "RemoveContainer" containerID="39ae973d44d7313b6a5dc308f01a2d8f8528e558a2117324b0a478bdce629279" Jan 21 11:25:57 crc kubenswrapper[4824]: E0121 11:25:57.167010 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"39ae973d44d7313b6a5dc308f01a2d8f8528e558a2117324b0a478bdce629279\": container with ID starting with 39ae973d44d7313b6a5dc308f01a2d8f8528e558a2117324b0a478bdce629279 not found: ID does not exist" containerID="39ae973d44d7313b6a5dc308f01a2d8f8528e558a2117324b0a478bdce629279" Jan 21 11:25:57 crc kubenswrapper[4824]: I0121 11:25:57.167048 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"39ae973d44d7313b6a5dc308f01a2d8f8528e558a2117324b0a478bdce629279"} err="failed to get container status \"39ae973d44d7313b6a5dc308f01a2d8f8528e558a2117324b0a478bdce629279\": rpc error: code = NotFound desc = could not find container \"39ae973d44d7313b6a5dc308f01a2d8f8528e558a2117324b0a478bdce629279\": container with ID starting with 39ae973d44d7313b6a5dc308f01a2d8f8528e558a2117324b0a478bdce629279 not found: ID does not exist" Jan 21 11:25:57 crc kubenswrapper[4824]: I0121 11:25:57.167073 4824 scope.go:117] "RemoveContainer" containerID="32f9fcde162e01284d5f8705d4630a345221a7c38111d1bb434ad870a2be305d" Jan 21 11:25:57 crc kubenswrapper[4824]: E0121 11:25:57.167370 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"32f9fcde162e01284d5f8705d4630a345221a7c38111d1bb434ad870a2be305d\": container with ID starting with 32f9fcde162e01284d5f8705d4630a345221a7c38111d1bb434ad870a2be305d not found: ID does not exist" containerID="32f9fcde162e01284d5f8705d4630a345221a7c38111d1bb434ad870a2be305d" Jan 21 11:25:57 crc kubenswrapper[4824]: I0121 11:25:57.167392 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"32f9fcde162e01284d5f8705d4630a345221a7c38111d1bb434ad870a2be305d"} err="failed to get container status \"32f9fcde162e01284d5f8705d4630a345221a7c38111d1bb434ad870a2be305d\": rpc error: code = NotFound desc = could not find container \"32f9fcde162e01284d5f8705d4630a345221a7c38111d1bb434ad870a2be305d\": container with ID starting with 32f9fcde162e01284d5f8705d4630a345221a7c38111d1bb434ad870a2be305d not found: ID does not exist" Jan 21 11:25:57 crc kubenswrapper[4824]: I0121 11:25:57.167415 4824 scope.go:117] "RemoveContainer" containerID="e050c4db8b50c4f7ba6c6253d674cec54ab71b556af9b68dbd86558b0523c4ab" Jan 21 11:25:57 crc kubenswrapper[4824]: E0121 11:25:57.169857 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e050c4db8b50c4f7ba6c6253d674cec54ab71b556af9b68dbd86558b0523c4ab\": container with ID starting with e050c4db8b50c4f7ba6c6253d674cec54ab71b556af9b68dbd86558b0523c4ab not found: ID does not exist" containerID="e050c4db8b50c4f7ba6c6253d674cec54ab71b556af9b68dbd86558b0523c4ab" Jan 21 11:25:57 crc kubenswrapper[4824]: I0121 11:25:57.169887 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e050c4db8b50c4f7ba6c6253d674cec54ab71b556af9b68dbd86558b0523c4ab"} err="failed to get container status \"e050c4db8b50c4f7ba6c6253d674cec54ab71b556af9b68dbd86558b0523c4ab\": rpc error: code = NotFound desc = could not find container \"e050c4db8b50c4f7ba6c6253d674cec54ab71b556af9b68dbd86558b0523c4ab\": container with ID starting with e050c4db8b50c4f7ba6c6253d674cec54ab71b556af9b68dbd86558b0523c4ab not found: ID does not exist" Jan 21 11:25:57 crc kubenswrapper[4824]: I0121 11:25:57.231495 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-b795q"] Jan 21 11:25:57 crc kubenswrapper[4824]: W0121 11:25:57.238445 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1c75f0f9_53e1_4a64_9211_559ef51c1b53.slice/crio-799120a9a7a20b359b616c2030579c280c96e8a87a304704ae580c5acba60e25 WatchSource:0}: Error finding container 799120a9a7a20b359b616c2030579c280c96e8a87a304704ae580c5acba60e25: Status 404 returned error can't find the container with id 799120a9a7a20b359b616c2030579c280c96e8a87a304704ae580c5acba60e25 Jan 21 11:25:57 crc kubenswrapper[4824]: I0121 11:25:57.956876 4824 generic.go:334] "Generic (PLEG): container finished" podID="84e74d7b-18a0-4a3f-8680-6246ac538a6e" containerID="12b90320ffa88a78c84d62cf8b65c324fb6c4b884b2714e0a27b33b29cca2e17" exitCode=0 Jan 21 11:25:57 crc kubenswrapper[4824]: I0121 11:25:57.956906 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"84e74d7b-18a0-4a3f-8680-6246ac538a6e","Type":"ContainerDied","Data":"12b90320ffa88a78c84d62cf8b65c324fb6c4b884b2714e0a27b33b29cca2e17"} Jan 21 11:25:57 crc kubenswrapper[4824]: I0121 11:25:57.958354 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b795q" event={"ID":"1c75f0f9-53e1-4a64-9211-559ef51c1b53","Type":"ContainerStarted","Data":"799120a9a7a20b359b616c2030579c280c96e8a87a304704ae580c5acba60e25"} Jan 21 11:25:57 crc kubenswrapper[4824]: I0121 11:25:57.959993 4824 generic.go:334] "Generic (PLEG): container finished" podID="156fec50-b486-4e84-a7bf-b40491a863c7" containerID="57e1badc440cdd38a60828271554d42b57947d40b89d5290cb4d30f5fcb54487" exitCode=0 Jan 21 11:25:57 crc kubenswrapper[4824]: I0121 11:25:57.960020 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"156fec50-b486-4e84-a7bf-b40491a863c7","Type":"ContainerDied","Data":"57e1badc440cdd38a60828271554d42b57947d40b89d5290cb4d30f5fcb54487"} Jan 21 11:25:58 crc kubenswrapper[4824]: I0121 11:25:58.062585 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bb4180b2-5d3e-486d-8197-348d01d2286e" path="/var/lib/kubelet/pods/bb4180b2-5d3e-486d-8197-348d01d2286e/volumes" Jan 21 11:25:58 crc kubenswrapper[4824]: I0121 11:25:58.213108 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-m6rpv" Jan 21 11:25:58 crc kubenswrapper[4824]: I0121 11:25:58.292555 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/76e33763-1a41-4640-8e4a-492ab92009e8-operator-scripts\") pod \"76e33763-1a41-4640-8e4a-492ab92009e8\" (UID: \"76e33763-1a41-4640-8e4a-492ab92009e8\") " Jan 21 11:25:58 crc kubenswrapper[4824]: I0121 11:25:58.292655 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2p2xd\" (UniqueName: \"kubernetes.io/projected/76e33763-1a41-4640-8e4a-492ab92009e8-kube-api-access-2p2xd\") pod \"76e33763-1a41-4640-8e4a-492ab92009e8\" (UID: \"76e33763-1a41-4640-8e4a-492ab92009e8\") " Jan 21 11:25:58 crc kubenswrapper[4824]: I0121 11:25:58.293391 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/76e33763-1a41-4640-8e4a-492ab92009e8-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "76e33763-1a41-4640-8e4a-492ab92009e8" (UID: "76e33763-1a41-4640-8e4a-492ab92009e8"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:25:58 crc kubenswrapper[4824]: I0121 11:25:58.296130 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/76e33763-1a41-4640-8e4a-492ab92009e8-kube-api-access-2p2xd" (OuterVolumeSpecName: "kube-api-access-2p2xd") pod "76e33763-1a41-4640-8e4a-492ab92009e8" (UID: "76e33763-1a41-4640-8e4a-492ab92009e8"). InnerVolumeSpecName "kube-api-access-2p2xd". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:25:58 crc kubenswrapper[4824]: I0121 11:25:58.394122 4824 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/76e33763-1a41-4640-8e4a-492ab92009e8-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 21 11:25:58 crc kubenswrapper[4824]: I0121 11:25:58.394148 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2p2xd\" (UniqueName: \"kubernetes.io/projected/76e33763-1a41-4640-8e4a-492ab92009e8-kube-api-access-2p2xd\") on node \"crc\" DevicePath \"\"" Jan 21 11:25:58 crc kubenswrapper[4824]: I0121 11:25:58.874700 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-5s6wj"] Jan 21 11:25:58 crc kubenswrapper[4824]: I0121 11:25:58.967370 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"156fec50-b486-4e84-a7bf-b40491a863c7","Type":"ContainerStarted","Data":"902d1a3e9a8d54f58be516ba97809a95404a4ebe849ca1cb1fe5805c23862e32"} Jan 21 11:25:58 crc kubenswrapper[4824]: I0121 11:25:58.968081 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Jan 21 11:25:58 crc kubenswrapper[4824]: I0121 11:25:58.969629 4824 generic.go:334] "Generic (PLEG): container finished" podID="1c75f0f9-53e1-4a64-9211-559ef51c1b53" containerID="dc92b9cbcf516998ce8a3e5f19a7012166529b9eee969255e60da7b567e11c76" exitCode=0 Jan 21 11:25:58 crc kubenswrapper[4824]: I0121 11:25:58.969684 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b795q" event={"ID":"1c75f0f9-53e1-4a64-9211-559ef51c1b53","Type":"ContainerDied","Data":"dc92b9cbcf516998ce8a3e5f19a7012166529b9eee969255e60da7b567e11c76"} Jan 21 11:25:58 crc kubenswrapper[4824]: I0121 11:25:58.971276 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-m6rpv" event={"ID":"76e33763-1a41-4640-8e4a-492ab92009e8","Type":"ContainerDied","Data":"78f8c0c168ea13f1ae2d548e6e02a4143f0dde836f9d6fb9f2378253323b9a7a"} Jan 21 11:25:58 crc kubenswrapper[4824]: I0121 11:25:58.971293 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="78f8c0c168ea13f1ae2d548e6e02a4143f0dde836f9d6fb9f2378253323b9a7a" Jan 21 11:25:58 crc kubenswrapper[4824]: I0121 11:25:58.971330 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-m6rpv" Jan 21 11:25:58 crc kubenswrapper[4824]: I0121 11:25:58.984489 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-5s6wj" podUID="6426f379-b3d5-4589-b5e1-659aed932410" containerName="registry-server" containerID="cri-o://828858ab38aa4d0e3c2ec1fe8be4d8444b2d96a181e3ed3dd9412225aeb16e0e" gracePeriod=2 Jan 21 11:25:58 crc kubenswrapper[4824]: I0121 11:25:58.984737 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"84e74d7b-18a0-4a3f-8680-6246ac538a6e","Type":"ContainerStarted","Data":"407422f3cadd164c7680cc8b0bf1b21f0ebb5d9d1b6255a64420fbd1143666e7"} Jan 21 11:25:58 crc kubenswrapper[4824]: I0121 11:25:58.985519 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Jan 21 11:25:58 crc kubenswrapper[4824]: I0121 11:25:58.986804 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=43.4561252 podStartE2EDuration="48.98679663s" podCreationTimestamp="2026-01-21 11:25:10 +0000 UTC" firstStartedPulling="2026-01-21 11:25:19.035726855 +0000 UTC m=+861.328756147" lastFinishedPulling="2026-01-21 11:25:24.566398285 +0000 UTC m=+866.859427577" observedRunningTime="2026-01-21 11:25:58.985856247 +0000 UTC m=+901.278885540" watchObservedRunningTime="2026-01-21 11:25:58.98679663 +0000 UTC m=+901.279825922" Jan 21 11:25:59 crc kubenswrapper[4824]: I0121 11:25:59.011445 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=43.567915502 podStartE2EDuration="49.011429051s" podCreationTimestamp="2026-01-21 11:25:10 +0000 UTC" firstStartedPulling="2026-01-21 11:25:19.153926917 +0000 UTC m=+861.446956210" lastFinishedPulling="2026-01-21 11:25:24.597440467 +0000 UTC m=+866.890469759" observedRunningTime="2026-01-21 11:25:59.006029493 +0000 UTC m=+901.299058785" watchObservedRunningTime="2026-01-21 11:25:59.011429051 +0000 UTC m=+901.304458343" Jan 21 11:25:59 crc kubenswrapper[4824]: I0121 11:25:59.382541 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5s6wj" Jan 21 11:25:59 crc kubenswrapper[4824]: I0121 11:25:59.509785 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6426f379-b3d5-4589-b5e1-659aed932410-catalog-content\") pod \"6426f379-b3d5-4589-b5e1-659aed932410\" (UID: \"6426f379-b3d5-4589-b5e1-659aed932410\") " Jan 21 11:25:59 crc kubenswrapper[4824]: I0121 11:25:59.509990 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6426f379-b3d5-4589-b5e1-659aed932410-utilities\") pod \"6426f379-b3d5-4589-b5e1-659aed932410\" (UID: \"6426f379-b3d5-4589-b5e1-659aed932410\") " Jan 21 11:25:59 crc kubenswrapper[4824]: I0121 11:25:59.510036 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rqdmn\" (UniqueName: \"kubernetes.io/projected/6426f379-b3d5-4589-b5e1-659aed932410-kube-api-access-rqdmn\") pod \"6426f379-b3d5-4589-b5e1-659aed932410\" (UID: \"6426f379-b3d5-4589-b5e1-659aed932410\") " Jan 21 11:25:59 crc kubenswrapper[4824]: I0121 11:25:59.510574 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6426f379-b3d5-4589-b5e1-659aed932410-utilities" (OuterVolumeSpecName: "utilities") pod "6426f379-b3d5-4589-b5e1-659aed932410" (UID: "6426f379-b3d5-4589-b5e1-659aed932410"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:25:59 crc kubenswrapper[4824]: I0121 11:25:59.515280 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6426f379-b3d5-4589-b5e1-659aed932410-kube-api-access-rqdmn" (OuterVolumeSpecName: "kube-api-access-rqdmn") pod "6426f379-b3d5-4589-b5e1-659aed932410" (UID: "6426f379-b3d5-4589-b5e1-659aed932410"). InnerVolumeSpecName "kube-api-access-rqdmn". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:25:59 crc kubenswrapper[4824]: I0121 11:25:59.612103 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6426f379-b3d5-4589-b5e1-659aed932410-utilities\") on node \"crc\" DevicePath \"\"" Jan 21 11:25:59 crc kubenswrapper[4824]: I0121 11:25:59.612125 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rqdmn\" (UniqueName: \"kubernetes.io/projected/6426f379-b3d5-4589-b5e1-659aed932410-kube-api-access-rqdmn\") on node \"crc\" DevicePath \"\"" Jan 21 11:25:59 crc kubenswrapper[4824]: I0121 11:25:59.695813 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6426f379-b3d5-4589-b5e1-659aed932410-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6426f379-b3d5-4589-b5e1-659aed932410" (UID: "6426f379-b3d5-4589-b5e1-659aed932410"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:25:59 crc kubenswrapper[4824]: I0121 11:25:59.713331 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6426f379-b3d5-4589-b5e1-659aed932410-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 21 11:25:59 crc kubenswrapper[4824]: I0121 11:25:59.991645 4824 generic.go:334] "Generic (PLEG): container finished" podID="1c75f0f9-53e1-4a64-9211-559ef51c1b53" containerID="af83a2d8b2c7751e366bf2114a7547afc9bf4e4dc745f7e937384620bcdb2e71" exitCode=0 Jan 21 11:25:59 crc kubenswrapper[4824]: I0121 11:25:59.991723 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b795q" event={"ID":"1c75f0f9-53e1-4a64-9211-559ef51c1b53","Type":"ContainerDied","Data":"af83a2d8b2c7751e366bf2114a7547afc9bf4e4dc745f7e937384620bcdb2e71"} Jan 21 11:25:59 crc kubenswrapper[4824]: I0121 11:25:59.993793 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"f980e493-ca74-47f1-8f6f-aca9c9011bf5","Type":"ContainerStarted","Data":"d29ad988a899f06204abf43bd5ea374c7b88de2de25e37348d02940d02652131"} Jan 21 11:25:59 crc kubenswrapper[4824]: I0121 11:25:59.993996 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"f980e493-ca74-47f1-8f6f-aca9c9011bf5","Type":"ContainerStarted","Data":"b804a5ae6f06ff30e4552c437c35633b5fa1b97571607261f6b664323dc9ffc2"} Jan 21 11:25:59 crc kubenswrapper[4824]: I0121 11:25:59.994009 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"f980e493-ca74-47f1-8f6f-aca9c9011bf5","Type":"ContainerStarted","Data":"857de0e92541a2e1927ec6b1d7b50def1b5a78ae48f2a6eb7c779e2fc4c4b45e"} Jan 21 11:25:59 crc kubenswrapper[4824]: I0121 11:25:59.994016 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"f980e493-ca74-47f1-8f6f-aca9c9011bf5","Type":"ContainerStarted","Data":"aaccd44ac3f4aff508890cdda80001df1cc8ec7e9277bfe539d8bef3c06d1be3"} Jan 21 11:25:59 crc kubenswrapper[4824]: I0121 11:25:59.995805 4824 generic.go:334] "Generic (PLEG): container finished" podID="6426f379-b3d5-4589-b5e1-659aed932410" containerID="828858ab38aa4d0e3c2ec1fe8be4d8444b2d96a181e3ed3dd9412225aeb16e0e" exitCode=0 Jan 21 11:25:59 crc kubenswrapper[4824]: I0121 11:25:59.996021 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5s6wj" event={"ID":"6426f379-b3d5-4589-b5e1-659aed932410","Type":"ContainerDied","Data":"828858ab38aa4d0e3c2ec1fe8be4d8444b2d96a181e3ed3dd9412225aeb16e0e"} Jan 21 11:25:59 crc kubenswrapper[4824]: I0121 11:25:59.996045 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5s6wj" event={"ID":"6426f379-b3d5-4589-b5e1-659aed932410","Type":"ContainerDied","Data":"bc78d731286fa1edc4820267f7da4e9a25da4496ce988c15be5ce2476d00d97c"} Jan 21 11:25:59 crc kubenswrapper[4824]: I0121 11:25:59.996064 4824 scope.go:117] "RemoveContainer" containerID="828858ab38aa4d0e3c2ec1fe8be4d8444b2d96a181e3ed3dd9412225aeb16e0e" Jan 21 11:25:59 crc kubenswrapper[4824]: I0121 11:25:59.996079 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5s6wj" Jan 21 11:26:00 crc kubenswrapper[4824]: I0121 11:26:00.011514 4824 scope.go:117] "RemoveContainer" containerID="44fe649276096a6e22d13f5809a21cfaa7a7d7fb3e877b6985edd5b4017b012c" Jan 21 11:26:00 crc kubenswrapper[4824]: I0121 11:26:00.029306 4824 scope.go:117] "RemoveContainer" containerID="7d27efc53a3e07270c497cc3025ae83bbe01ff00a96faf06c726ec8704aea5d3" Jan 21 11:26:00 crc kubenswrapper[4824]: I0121 11:26:00.030940 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-5s6wj"] Jan 21 11:26:00 crc kubenswrapper[4824]: I0121 11:26:00.036948 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-5s6wj"] Jan 21 11:26:00 crc kubenswrapper[4824]: I0121 11:26:00.053488 4824 scope.go:117] "RemoveContainer" containerID="828858ab38aa4d0e3c2ec1fe8be4d8444b2d96a181e3ed3dd9412225aeb16e0e" Jan 21 11:26:00 crc kubenswrapper[4824]: E0121 11:26:00.055689 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"828858ab38aa4d0e3c2ec1fe8be4d8444b2d96a181e3ed3dd9412225aeb16e0e\": container with ID starting with 828858ab38aa4d0e3c2ec1fe8be4d8444b2d96a181e3ed3dd9412225aeb16e0e not found: ID does not exist" containerID="828858ab38aa4d0e3c2ec1fe8be4d8444b2d96a181e3ed3dd9412225aeb16e0e" Jan 21 11:26:00 crc kubenswrapper[4824]: I0121 11:26:00.055719 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"828858ab38aa4d0e3c2ec1fe8be4d8444b2d96a181e3ed3dd9412225aeb16e0e"} err="failed to get container status \"828858ab38aa4d0e3c2ec1fe8be4d8444b2d96a181e3ed3dd9412225aeb16e0e\": rpc error: code = NotFound desc = could not find container \"828858ab38aa4d0e3c2ec1fe8be4d8444b2d96a181e3ed3dd9412225aeb16e0e\": container with ID starting with 828858ab38aa4d0e3c2ec1fe8be4d8444b2d96a181e3ed3dd9412225aeb16e0e not found: ID does not exist" Jan 21 11:26:00 crc kubenswrapper[4824]: I0121 11:26:00.055736 4824 scope.go:117] "RemoveContainer" containerID="44fe649276096a6e22d13f5809a21cfaa7a7d7fb3e877b6985edd5b4017b012c" Jan 21 11:26:00 crc kubenswrapper[4824]: E0121 11:26:00.056093 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"44fe649276096a6e22d13f5809a21cfaa7a7d7fb3e877b6985edd5b4017b012c\": container with ID starting with 44fe649276096a6e22d13f5809a21cfaa7a7d7fb3e877b6985edd5b4017b012c not found: ID does not exist" containerID="44fe649276096a6e22d13f5809a21cfaa7a7d7fb3e877b6985edd5b4017b012c" Jan 21 11:26:00 crc kubenswrapper[4824]: I0121 11:26:00.056127 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"44fe649276096a6e22d13f5809a21cfaa7a7d7fb3e877b6985edd5b4017b012c"} err="failed to get container status \"44fe649276096a6e22d13f5809a21cfaa7a7d7fb3e877b6985edd5b4017b012c\": rpc error: code = NotFound desc = could not find container \"44fe649276096a6e22d13f5809a21cfaa7a7d7fb3e877b6985edd5b4017b012c\": container with ID starting with 44fe649276096a6e22d13f5809a21cfaa7a7d7fb3e877b6985edd5b4017b012c not found: ID does not exist" Jan 21 11:26:00 crc kubenswrapper[4824]: I0121 11:26:00.056149 4824 scope.go:117] "RemoveContainer" containerID="7d27efc53a3e07270c497cc3025ae83bbe01ff00a96faf06c726ec8704aea5d3" Jan 21 11:26:00 crc kubenswrapper[4824]: E0121 11:26:00.056450 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7d27efc53a3e07270c497cc3025ae83bbe01ff00a96faf06c726ec8704aea5d3\": container with ID starting with 7d27efc53a3e07270c497cc3025ae83bbe01ff00a96faf06c726ec8704aea5d3 not found: ID does not exist" containerID="7d27efc53a3e07270c497cc3025ae83bbe01ff00a96faf06c726ec8704aea5d3" Jan 21 11:26:00 crc kubenswrapper[4824]: I0121 11:26:00.056475 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7d27efc53a3e07270c497cc3025ae83bbe01ff00a96faf06c726ec8704aea5d3"} err="failed to get container status \"7d27efc53a3e07270c497cc3025ae83bbe01ff00a96faf06c726ec8704aea5d3\": rpc error: code = NotFound desc = could not find container \"7d27efc53a3e07270c497cc3025ae83bbe01ff00a96faf06c726ec8704aea5d3\": container with ID starting with 7d27efc53a3e07270c497cc3025ae83bbe01ff00a96faf06c726ec8704aea5d3 not found: ID does not exist" Jan 21 11:26:00 crc kubenswrapper[4824]: I0121 11:26:00.062658 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6426f379-b3d5-4589-b5e1-659aed932410" path="/var/lib/kubelet/pods/6426f379-b3d5-4589-b5e1-659aed932410/volumes" Jan 21 11:26:00 crc kubenswrapper[4824]: I0121 11:26:00.137654 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-htwd6" podUID="511f2b6d-a08d-49f8-b393-ab222219d4a7" containerName="ovn-controller" probeResult="failure" output=< Jan 21 11:26:00 crc kubenswrapper[4824]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Jan 21 11:26:00 crc kubenswrapper[4824]: > Jan 21 11:26:00 crc kubenswrapper[4824]: I0121 11:26:00.242468 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-gr9gj" Jan 21 11:26:00 crc kubenswrapper[4824]: I0121 11:26:00.257166 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-gr9gj" Jan 21 11:26:00 crc kubenswrapper[4824]: I0121 11:26:00.448274 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-htwd6-config-jbc69"] Jan 21 11:26:00 crc kubenswrapper[4824]: E0121 11:26:00.448576 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6426f379-b3d5-4589-b5e1-659aed932410" containerName="registry-server" Jan 21 11:26:00 crc kubenswrapper[4824]: I0121 11:26:00.448592 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="6426f379-b3d5-4589-b5e1-659aed932410" containerName="registry-server" Jan 21 11:26:00 crc kubenswrapper[4824]: E0121 11:26:00.448604 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6426f379-b3d5-4589-b5e1-659aed932410" containerName="extract-content" Jan 21 11:26:00 crc kubenswrapper[4824]: I0121 11:26:00.448610 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="6426f379-b3d5-4589-b5e1-659aed932410" containerName="extract-content" Jan 21 11:26:00 crc kubenswrapper[4824]: E0121 11:26:00.448619 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76e33763-1a41-4640-8e4a-492ab92009e8" containerName="mariadb-account-create-update" Jan 21 11:26:00 crc kubenswrapper[4824]: I0121 11:26:00.448624 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="76e33763-1a41-4640-8e4a-492ab92009e8" containerName="mariadb-account-create-update" Jan 21 11:26:00 crc kubenswrapper[4824]: E0121 11:26:00.448642 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6426f379-b3d5-4589-b5e1-659aed932410" containerName="extract-utilities" Jan 21 11:26:00 crc kubenswrapper[4824]: I0121 11:26:00.448648 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="6426f379-b3d5-4589-b5e1-659aed932410" containerName="extract-utilities" Jan 21 11:26:00 crc kubenswrapper[4824]: I0121 11:26:00.448806 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="76e33763-1a41-4640-8e4a-492ab92009e8" containerName="mariadb-account-create-update" Jan 21 11:26:00 crc kubenswrapper[4824]: I0121 11:26:00.448826 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="6426f379-b3d5-4589-b5e1-659aed932410" containerName="registry-server" Jan 21 11:26:00 crc kubenswrapper[4824]: I0121 11:26:00.449314 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-htwd6-config-jbc69" Jan 21 11:26:00 crc kubenswrapper[4824]: I0121 11:26:00.456026 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Jan 21 11:26:00 crc kubenswrapper[4824]: I0121 11:26:00.469028 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-htwd6-config-jbc69"] Jan 21 11:26:00 crc kubenswrapper[4824]: I0121 11:26:00.533784 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/6d3982e6-513d-454f-a08a-34b32779b559-var-log-ovn\") pod \"ovn-controller-htwd6-config-jbc69\" (UID: \"6d3982e6-513d-454f-a08a-34b32779b559\") " pod="openstack/ovn-controller-htwd6-config-jbc69" Jan 21 11:26:00 crc kubenswrapper[4824]: I0121 11:26:00.533818 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dmxtd\" (UniqueName: \"kubernetes.io/projected/6d3982e6-513d-454f-a08a-34b32779b559-kube-api-access-dmxtd\") pod \"ovn-controller-htwd6-config-jbc69\" (UID: \"6d3982e6-513d-454f-a08a-34b32779b559\") " pod="openstack/ovn-controller-htwd6-config-jbc69" Jan 21 11:26:00 crc kubenswrapper[4824]: I0121 11:26:00.533870 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6d3982e6-513d-454f-a08a-34b32779b559-scripts\") pod \"ovn-controller-htwd6-config-jbc69\" (UID: \"6d3982e6-513d-454f-a08a-34b32779b559\") " pod="openstack/ovn-controller-htwd6-config-jbc69" Jan 21 11:26:00 crc kubenswrapper[4824]: I0121 11:26:00.533998 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/6d3982e6-513d-454f-a08a-34b32779b559-additional-scripts\") pod \"ovn-controller-htwd6-config-jbc69\" (UID: \"6d3982e6-513d-454f-a08a-34b32779b559\") " pod="openstack/ovn-controller-htwd6-config-jbc69" Jan 21 11:26:00 crc kubenswrapper[4824]: I0121 11:26:00.534022 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/6d3982e6-513d-454f-a08a-34b32779b559-var-run\") pod \"ovn-controller-htwd6-config-jbc69\" (UID: \"6d3982e6-513d-454f-a08a-34b32779b559\") " pod="openstack/ovn-controller-htwd6-config-jbc69" Jan 21 11:26:00 crc kubenswrapper[4824]: I0121 11:26:00.534053 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/6d3982e6-513d-454f-a08a-34b32779b559-var-run-ovn\") pod \"ovn-controller-htwd6-config-jbc69\" (UID: \"6d3982e6-513d-454f-a08a-34b32779b559\") " pod="openstack/ovn-controller-htwd6-config-jbc69" Jan 21 11:26:00 crc kubenswrapper[4824]: I0121 11:26:00.635267 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/6d3982e6-513d-454f-a08a-34b32779b559-var-log-ovn\") pod \"ovn-controller-htwd6-config-jbc69\" (UID: \"6d3982e6-513d-454f-a08a-34b32779b559\") " pod="openstack/ovn-controller-htwd6-config-jbc69" Jan 21 11:26:00 crc kubenswrapper[4824]: I0121 11:26:00.635451 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dmxtd\" (UniqueName: \"kubernetes.io/projected/6d3982e6-513d-454f-a08a-34b32779b559-kube-api-access-dmxtd\") pod \"ovn-controller-htwd6-config-jbc69\" (UID: \"6d3982e6-513d-454f-a08a-34b32779b559\") " pod="openstack/ovn-controller-htwd6-config-jbc69" Jan 21 11:26:00 crc kubenswrapper[4824]: I0121 11:26:00.635547 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/6d3982e6-513d-454f-a08a-34b32779b559-var-log-ovn\") pod \"ovn-controller-htwd6-config-jbc69\" (UID: \"6d3982e6-513d-454f-a08a-34b32779b559\") " pod="openstack/ovn-controller-htwd6-config-jbc69" Jan 21 11:26:00 crc kubenswrapper[4824]: I0121 11:26:00.635650 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6d3982e6-513d-454f-a08a-34b32779b559-scripts\") pod \"ovn-controller-htwd6-config-jbc69\" (UID: \"6d3982e6-513d-454f-a08a-34b32779b559\") " pod="openstack/ovn-controller-htwd6-config-jbc69" Jan 21 11:26:00 crc kubenswrapper[4824]: I0121 11:26:00.635862 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/6d3982e6-513d-454f-a08a-34b32779b559-additional-scripts\") pod \"ovn-controller-htwd6-config-jbc69\" (UID: \"6d3982e6-513d-454f-a08a-34b32779b559\") " pod="openstack/ovn-controller-htwd6-config-jbc69" Jan 21 11:26:00 crc kubenswrapper[4824]: I0121 11:26:00.635942 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/6d3982e6-513d-454f-a08a-34b32779b559-var-run\") pod \"ovn-controller-htwd6-config-jbc69\" (UID: \"6d3982e6-513d-454f-a08a-34b32779b559\") " pod="openstack/ovn-controller-htwd6-config-jbc69" Jan 21 11:26:00 crc kubenswrapper[4824]: I0121 11:26:00.636068 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/6d3982e6-513d-454f-a08a-34b32779b559-var-run\") pod \"ovn-controller-htwd6-config-jbc69\" (UID: \"6d3982e6-513d-454f-a08a-34b32779b559\") " pod="openstack/ovn-controller-htwd6-config-jbc69" Jan 21 11:26:00 crc kubenswrapper[4824]: I0121 11:26:00.636087 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/6d3982e6-513d-454f-a08a-34b32779b559-var-run-ovn\") pod \"ovn-controller-htwd6-config-jbc69\" (UID: \"6d3982e6-513d-454f-a08a-34b32779b559\") " pod="openstack/ovn-controller-htwd6-config-jbc69" Jan 21 11:26:00 crc kubenswrapper[4824]: I0121 11:26:00.636253 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/6d3982e6-513d-454f-a08a-34b32779b559-var-run-ovn\") pod \"ovn-controller-htwd6-config-jbc69\" (UID: \"6d3982e6-513d-454f-a08a-34b32779b559\") " pod="openstack/ovn-controller-htwd6-config-jbc69" Jan 21 11:26:00 crc kubenswrapper[4824]: I0121 11:26:00.636430 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/6d3982e6-513d-454f-a08a-34b32779b559-additional-scripts\") pod \"ovn-controller-htwd6-config-jbc69\" (UID: \"6d3982e6-513d-454f-a08a-34b32779b559\") " pod="openstack/ovn-controller-htwd6-config-jbc69" Jan 21 11:26:00 crc kubenswrapper[4824]: I0121 11:26:00.637474 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6d3982e6-513d-454f-a08a-34b32779b559-scripts\") pod \"ovn-controller-htwd6-config-jbc69\" (UID: \"6d3982e6-513d-454f-a08a-34b32779b559\") " pod="openstack/ovn-controller-htwd6-config-jbc69" Jan 21 11:26:00 crc kubenswrapper[4824]: I0121 11:26:00.658183 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dmxtd\" (UniqueName: \"kubernetes.io/projected/6d3982e6-513d-454f-a08a-34b32779b559-kube-api-access-dmxtd\") pod \"ovn-controller-htwd6-config-jbc69\" (UID: \"6d3982e6-513d-454f-a08a-34b32779b559\") " pod="openstack/ovn-controller-htwd6-config-jbc69" Jan 21 11:26:00 crc kubenswrapper[4824]: I0121 11:26:00.773205 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-htwd6-config-jbc69" Jan 21 11:26:01 crc kubenswrapper[4824]: I0121 11:26:01.006021 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"f980e493-ca74-47f1-8f6f-aca9c9011bf5","Type":"ContainerStarted","Data":"dc6ddaa09fb1ec22ec741c78662d75b99df4b6da66737456c8dccf858b7522d7"} Jan 21 11:26:01 crc kubenswrapper[4824]: I0121 11:26:01.006398 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"f980e493-ca74-47f1-8f6f-aca9c9011bf5","Type":"ContainerStarted","Data":"4bda5ab0e896dbb529ef5290ea2a3c7e17af838da67d703a146d1016845d42a6"} Jan 21 11:26:01 crc kubenswrapper[4824]: I0121 11:26:01.011202 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b795q" event={"ID":"1c75f0f9-53e1-4a64-9211-559ef51c1b53","Type":"ContainerStarted","Data":"9e1d278d517cd8394b73f2d45c4430f8e3523f06927bcb011e76233b05081e45"} Jan 21 11:26:01 crc kubenswrapper[4824]: I0121 11:26:01.028754 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-b795q" podStartSLOduration=3.321225141 podStartE2EDuration="5.028740899s" podCreationTimestamp="2026-01-21 11:25:56 +0000 UTC" firstStartedPulling="2026-01-21 11:25:58.970461649 +0000 UTC m=+901.263490941" lastFinishedPulling="2026-01-21 11:26:00.677977407 +0000 UTC m=+902.971006699" observedRunningTime="2026-01-21 11:26:01.024055567 +0000 UTC m=+903.317084870" watchObservedRunningTime="2026-01-21 11:26:01.028740899 +0000 UTC m=+903.321770181" Jan 21 11:26:01 crc kubenswrapper[4824]: I0121 11:26:01.149091 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-htwd6-config-jbc69"] Jan 21 11:26:01 crc kubenswrapper[4824]: W0121 11:26:01.159577 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6d3982e6_513d_454f_a08a_34b32779b559.slice/crio-f83aa520815a13fae32f75f56d080344355c254e33f32202648b20326ba13712 WatchSource:0}: Error finding container f83aa520815a13fae32f75f56d080344355c254e33f32202648b20326ba13712: Status 404 returned error can't find the container with id f83aa520815a13fae32f75f56d080344355c254e33f32202648b20326ba13712 Jan 21 11:26:02 crc kubenswrapper[4824]: I0121 11:26:02.021934 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"f980e493-ca74-47f1-8f6f-aca9c9011bf5","Type":"ContainerStarted","Data":"e964792da3bebb08ca73e92297f79f1b50c4c3b30e75a5c6fc4c9f54326727d1"} Jan 21 11:26:02 crc kubenswrapper[4824]: I0121 11:26:02.022264 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"f980e493-ca74-47f1-8f6f-aca9c9011bf5","Type":"ContainerStarted","Data":"a2eb36fe6d265a82cf21727e4989a133a77f37bcc8b16fed1124388f79b0884b"} Jan 21 11:26:02 crc kubenswrapper[4824]: I0121 11:26:02.023784 4824 generic.go:334] "Generic (PLEG): container finished" podID="6d3982e6-513d-454f-a08a-34b32779b559" containerID="7802d3564cb8577bdb2e0cb7acab4d7e0721818cb3c2ece34e851778aff1034e" exitCode=0 Jan 21 11:26:02 crc kubenswrapper[4824]: I0121 11:26:02.023827 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-htwd6-config-jbc69" event={"ID":"6d3982e6-513d-454f-a08a-34b32779b559","Type":"ContainerDied","Data":"7802d3564cb8577bdb2e0cb7acab4d7e0721818cb3c2ece34e851778aff1034e"} Jan 21 11:26:02 crc kubenswrapper[4824]: I0121 11:26:02.023843 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-htwd6-config-jbc69" event={"ID":"6d3982e6-513d-454f-a08a-34b32779b559","Type":"ContainerStarted","Data":"f83aa520815a13fae32f75f56d080344355c254e33f32202648b20326ba13712"} Jan 21 11:26:02 crc kubenswrapper[4824]: I0121 11:26:02.025421 4824 generic.go:334] "Generic (PLEG): container finished" podID="ee7a30d4-dcec-4719-a629-dfaf588d4169" containerID="2b4aa0578cb659bd1fe3cb0351e682b347beae954c60ccf215edaf9c2da1abd4" exitCode=0 Jan 21 11:26:02 crc kubenswrapper[4824]: I0121 11:26:02.026137 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-sx9lb" event={"ID":"ee7a30d4-dcec-4719-a629-dfaf588d4169","Type":"ContainerDied","Data":"2b4aa0578cb659bd1fe3cb0351e682b347beae954c60ccf215edaf9c2da1abd4"} Jan 21 11:26:03 crc kubenswrapper[4824]: I0121 11:26:03.034193 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"f980e493-ca74-47f1-8f6f-aca9c9011bf5","Type":"ContainerStarted","Data":"3455e0eeedb7b32a4c23c322afa378a52b1af730fe411dd14be91d06eaa924c6"} Jan 21 11:26:03 crc kubenswrapper[4824]: I0121 11:26:03.034398 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"f980e493-ca74-47f1-8f6f-aca9c9011bf5","Type":"ContainerStarted","Data":"c626637d2ac5daec740897c7b85b3e7ac1dcc9b2bf37eef13e76aae43abc2677"} Jan 21 11:26:03 crc kubenswrapper[4824]: I0121 11:26:03.034410 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"f980e493-ca74-47f1-8f6f-aca9c9011bf5","Type":"ContainerStarted","Data":"b954af21a9cbab6f49f0ca38e4aa8205d0f06bb458256451df49256caf0cec2e"} Jan 21 11:26:03 crc kubenswrapper[4824]: I0121 11:26:03.286286 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-htwd6-config-jbc69" Jan 21 11:26:03 crc kubenswrapper[4824]: I0121 11:26:03.373101 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/6d3982e6-513d-454f-a08a-34b32779b559-var-log-ovn\") pod \"6d3982e6-513d-454f-a08a-34b32779b559\" (UID: \"6d3982e6-513d-454f-a08a-34b32779b559\") " Jan 21 11:26:03 crc kubenswrapper[4824]: I0121 11:26:03.373145 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/6d3982e6-513d-454f-a08a-34b32779b559-additional-scripts\") pod \"6d3982e6-513d-454f-a08a-34b32779b559\" (UID: \"6d3982e6-513d-454f-a08a-34b32779b559\") " Jan 21 11:26:03 crc kubenswrapper[4824]: I0121 11:26:03.373186 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dmxtd\" (UniqueName: \"kubernetes.io/projected/6d3982e6-513d-454f-a08a-34b32779b559-kube-api-access-dmxtd\") pod \"6d3982e6-513d-454f-a08a-34b32779b559\" (UID: \"6d3982e6-513d-454f-a08a-34b32779b559\") " Jan 21 11:26:03 crc kubenswrapper[4824]: I0121 11:26:03.373214 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6d3982e6-513d-454f-a08a-34b32779b559-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "6d3982e6-513d-454f-a08a-34b32779b559" (UID: "6d3982e6-513d-454f-a08a-34b32779b559"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 21 11:26:03 crc kubenswrapper[4824]: I0121 11:26:03.373280 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6d3982e6-513d-454f-a08a-34b32779b559-scripts\") pod \"6d3982e6-513d-454f-a08a-34b32779b559\" (UID: \"6d3982e6-513d-454f-a08a-34b32779b559\") " Jan 21 11:26:03 crc kubenswrapper[4824]: I0121 11:26:03.373343 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/6d3982e6-513d-454f-a08a-34b32779b559-var-run-ovn\") pod \"6d3982e6-513d-454f-a08a-34b32779b559\" (UID: \"6d3982e6-513d-454f-a08a-34b32779b559\") " Jan 21 11:26:03 crc kubenswrapper[4824]: I0121 11:26:03.373364 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/6d3982e6-513d-454f-a08a-34b32779b559-var-run\") pod \"6d3982e6-513d-454f-a08a-34b32779b559\" (UID: \"6d3982e6-513d-454f-a08a-34b32779b559\") " Jan 21 11:26:03 crc kubenswrapper[4824]: I0121 11:26:03.373455 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6d3982e6-513d-454f-a08a-34b32779b559-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "6d3982e6-513d-454f-a08a-34b32779b559" (UID: "6d3982e6-513d-454f-a08a-34b32779b559"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 21 11:26:03 crc kubenswrapper[4824]: I0121 11:26:03.373572 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6d3982e6-513d-454f-a08a-34b32779b559-var-run" (OuterVolumeSpecName: "var-run") pod "6d3982e6-513d-454f-a08a-34b32779b559" (UID: "6d3982e6-513d-454f-a08a-34b32779b559"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 21 11:26:03 crc kubenswrapper[4824]: I0121 11:26:03.373767 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6d3982e6-513d-454f-a08a-34b32779b559-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "6d3982e6-513d-454f-a08a-34b32779b559" (UID: "6d3982e6-513d-454f-a08a-34b32779b559"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:26:03 crc kubenswrapper[4824]: I0121 11:26:03.373935 4824 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/6d3982e6-513d-454f-a08a-34b32779b559-var-run-ovn\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:03 crc kubenswrapper[4824]: I0121 11:26:03.373953 4824 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/6d3982e6-513d-454f-a08a-34b32779b559-var-run\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:03 crc kubenswrapper[4824]: I0121 11:26:03.373973 4824 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/6d3982e6-513d-454f-a08a-34b32779b559-var-log-ovn\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:03 crc kubenswrapper[4824]: I0121 11:26:03.373983 4824 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/6d3982e6-513d-454f-a08a-34b32779b559-additional-scripts\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:03 crc kubenswrapper[4824]: I0121 11:26:03.374079 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6d3982e6-513d-454f-a08a-34b32779b559-scripts" (OuterVolumeSpecName: "scripts") pod "6d3982e6-513d-454f-a08a-34b32779b559" (UID: "6d3982e6-513d-454f-a08a-34b32779b559"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:26:03 crc kubenswrapper[4824]: I0121 11:26:03.377917 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6d3982e6-513d-454f-a08a-34b32779b559-kube-api-access-dmxtd" (OuterVolumeSpecName: "kube-api-access-dmxtd") pod "6d3982e6-513d-454f-a08a-34b32779b559" (UID: "6d3982e6-513d-454f-a08a-34b32779b559"). InnerVolumeSpecName "kube-api-access-dmxtd". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:26:03 crc kubenswrapper[4824]: I0121 11:26:03.386830 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-sx9lb" Jan 21 11:26:03 crc kubenswrapper[4824]: I0121 11:26:03.474666 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee7a30d4-dcec-4719-a629-dfaf588d4169-combined-ca-bundle\") pod \"ee7a30d4-dcec-4719-a629-dfaf588d4169\" (UID: \"ee7a30d4-dcec-4719-a629-dfaf588d4169\") " Jan 21 11:26:03 crc kubenswrapper[4824]: I0121 11:26:03.474723 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ee7a30d4-dcec-4719-a629-dfaf588d4169-db-sync-config-data\") pod \"ee7a30d4-dcec-4719-a629-dfaf588d4169\" (UID: \"ee7a30d4-dcec-4719-a629-dfaf588d4169\") " Jan 21 11:26:03 crc kubenswrapper[4824]: I0121 11:26:03.474807 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cz8ld\" (UniqueName: \"kubernetes.io/projected/ee7a30d4-dcec-4719-a629-dfaf588d4169-kube-api-access-cz8ld\") pod \"ee7a30d4-dcec-4719-a629-dfaf588d4169\" (UID: \"ee7a30d4-dcec-4719-a629-dfaf588d4169\") " Jan 21 11:26:03 crc kubenswrapper[4824]: I0121 11:26:03.474874 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ee7a30d4-dcec-4719-a629-dfaf588d4169-config-data\") pod \"ee7a30d4-dcec-4719-a629-dfaf588d4169\" (UID: \"ee7a30d4-dcec-4719-a629-dfaf588d4169\") " Jan 21 11:26:03 crc kubenswrapper[4824]: I0121 11:26:03.475234 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dmxtd\" (UniqueName: \"kubernetes.io/projected/6d3982e6-513d-454f-a08a-34b32779b559-kube-api-access-dmxtd\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:03 crc kubenswrapper[4824]: I0121 11:26:03.475253 4824 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6d3982e6-513d-454f-a08a-34b32779b559-scripts\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:03 crc kubenswrapper[4824]: I0121 11:26:03.477983 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ee7a30d4-dcec-4719-a629-dfaf588d4169-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "ee7a30d4-dcec-4719-a629-dfaf588d4169" (UID: "ee7a30d4-dcec-4719-a629-dfaf588d4169"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:26:03 crc kubenswrapper[4824]: I0121 11:26:03.478249 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ee7a30d4-dcec-4719-a629-dfaf588d4169-kube-api-access-cz8ld" (OuterVolumeSpecName: "kube-api-access-cz8ld") pod "ee7a30d4-dcec-4719-a629-dfaf588d4169" (UID: "ee7a30d4-dcec-4719-a629-dfaf588d4169"). InnerVolumeSpecName "kube-api-access-cz8ld". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:26:03 crc kubenswrapper[4824]: I0121 11:26:03.490604 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ee7a30d4-dcec-4719-a629-dfaf588d4169-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ee7a30d4-dcec-4719-a629-dfaf588d4169" (UID: "ee7a30d4-dcec-4719-a629-dfaf588d4169"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:26:03 crc kubenswrapper[4824]: I0121 11:26:03.502760 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ee7a30d4-dcec-4719-a629-dfaf588d4169-config-data" (OuterVolumeSpecName: "config-data") pod "ee7a30d4-dcec-4719-a629-dfaf588d4169" (UID: "ee7a30d4-dcec-4719-a629-dfaf588d4169"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:26:03 crc kubenswrapper[4824]: I0121 11:26:03.576292 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cz8ld\" (UniqueName: \"kubernetes.io/projected/ee7a30d4-dcec-4719-a629-dfaf588d4169-kube-api-access-cz8ld\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:03 crc kubenswrapper[4824]: I0121 11:26:03.576318 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ee7a30d4-dcec-4719-a629-dfaf588d4169-config-data\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:03 crc kubenswrapper[4824]: I0121 11:26:03.576329 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee7a30d4-dcec-4719-a629-dfaf588d4169-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:03 crc kubenswrapper[4824]: I0121 11:26:03.576337 4824 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ee7a30d4-dcec-4719-a629-dfaf588d4169-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:04 crc kubenswrapper[4824]: I0121 11:26:04.048248 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-sx9lb" Jan 21 11:26:04 crc kubenswrapper[4824]: I0121 11:26:04.048242 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-sx9lb" event={"ID":"ee7a30d4-dcec-4719-a629-dfaf588d4169","Type":"ContainerDied","Data":"0094c2d2871ffe548bcd39b394fb34d9654678fd14fa5a472f33e3e6e434588c"} Jan 21 11:26:04 crc kubenswrapper[4824]: I0121 11:26:04.048713 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0094c2d2871ffe548bcd39b394fb34d9654678fd14fa5a472f33e3e6e434588c" Jan 21 11:26:04 crc kubenswrapper[4824]: I0121 11:26:04.051074 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-htwd6-config-jbc69" Jan 21 11:26:04 crc kubenswrapper[4824]: I0121 11:26:04.060524 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-htwd6-config-jbc69" event={"ID":"6d3982e6-513d-454f-a08a-34b32779b559","Type":"ContainerDied","Data":"f83aa520815a13fae32f75f56d080344355c254e33f32202648b20326ba13712"} Jan 21 11:26:04 crc kubenswrapper[4824]: I0121 11:26:04.060558 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f83aa520815a13fae32f75f56d080344355c254e33f32202648b20326ba13712" Jan 21 11:26:04 crc kubenswrapper[4824]: I0121 11:26:04.384680 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5b946c75cc-6mzz4"] Jan 21 11:26:04 crc kubenswrapper[4824]: E0121 11:26:04.384967 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee7a30d4-dcec-4719-a629-dfaf588d4169" containerName="glance-db-sync" Jan 21 11:26:04 crc kubenswrapper[4824]: I0121 11:26:04.384995 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee7a30d4-dcec-4719-a629-dfaf588d4169" containerName="glance-db-sync" Jan 21 11:26:04 crc kubenswrapper[4824]: E0121 11:26:04.385015 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d3982e6-513d-454f-a08a-34b32779b559" containerName="ovn-config" Jan 21 11:26:04 crc kubenswrapper[4824]: I0121 11:26:04.385020 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d3982e6-513d-454f-a08a-34b32779b559" containerName="ovn-config" Jan 21 11:26:04 crc kubenswrapper[4824]: I0121 11:26:04.385164 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee7a30d4-dcec-4719-a629-dfaf588d4169" containerName="glance-db-sync" Jan 21 11:26:04 crc kubenswrapper[4824]: I0121 11:26:04.385177 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="6d3982e6-513d-454f-a08a-34b32779b559" containerName="ovn-config" Jan 21 11:26:04 crc kubenswrapper[4824]: I0121 11:26:04.385851 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b946c75cc-6mzz4" Jan 21 11:26:04 crc kubenswrapper[4824]: I0121 11:26:04.402086 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5b946c75cc-6mzz4"] Jan 21 11:26:04 crc kubenswrapper[4824]: I0121 11:26:04.454610 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-htwd6-config-jbc69"] Jan 21 11:26:04 crc kubenswrapper[4824]: I0121 11:26:04.463637 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-htwd6-config-jbc69"] Jan 21 11:26:04 crc kubenswrapper[4824]: I0121 11:26:04.487065 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3e491cbf-b795-4735-a356-c30d79d2fefa-config\") pod \"dnsmasq-dns-5b946c75cc-6mzz4\" (UID: \"3e491cbf-b795-4735-a356-c30d79d2fefa\") " pod="openstack/dnsmasq-dns-5b946c75cc-6mzz4" Jan 21 11:26:04 crc kubenswrapper[4824]: I0121 11:26:04.487146 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3e491cbf-b795-4735-a356-c30d79d2fefa-ovsdbserver-nb\") pod \"dnsmasq-dns-5b946c75cc-6mzz4\" (UID: \"3e491cbf-b795-4735-a356-c30d79d2fefa\") " pod="openstack/dnsmasq-dns-5b946c75cc-6mzz4" Jan 21 11:26:04 crc kubenswrapper[4824]: I0121 11:26:04.487224 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fpg89\" (UniqueName: \"kubernetes.io/projected/3e491cbf-b795-4735-a356-c30d79d2fefa-kube-api-access-fpg89\") pod \"dnsmasq-dns-5b946c75cc-6mzz4\" (UID: \"3e491cbf-b795-4735-a356-c30d79d2fefa\") " pod="openstack/dnsmasq-dns-5b946c75cc-6mzz4" Jan 21 11:26:04 crc kubenswrapper[4824]: I0121 11:26:04.487265 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3e491cbf-b795-4735-a356-c30d79d2fefa-ovsdbserver-sb\") pod \"dnsmasq-dns-5b946c75cc-6mzz4\" (UID: \"3e491cbf-b795-4735-a356-c30d79d2fefa\") " pod="openstack/dnsmasq-dns-5b946c75cc-6mzz4" Jan 21 11:26:04 crc kubenswrapper[4824]: I0121 11:26:04.487320 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3e491cbf-b795-4735-a356-c30d79d2fefa-dns-svc\") pod \"dnsmasq-dns-5b946c75cc-6mzz4\" (UID: \"3e491cbf-b795-4735-a356-c30d79d2fefa\") " pod="openstack/dnsmasq-dns-5b946c75cc-6mzz4" Jan 21 11:26:04 crc kubenswrapper[4824]: I0121 11:26:04.588995 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fpg89\" (UniqueName: \"kubernetes.io/projected/3e491cbf-b795-4735-a356-c30d79d2fefa-kube-api-access-fpg89\") pod \"dnsmasq-dns-5b946c75cc-6mzz4\" (UID: \"3e491cbf-b795-4735-a356-c30d79d2fefa\") " pod="openstack/dnsmasq-dns-5b946c75cc-6mzz4" Jan 21 11:26:04 crc kubenswrapper[4824]: I0121 11:26:04.589058 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3e491cbf-b795-4735-a356-c30d79d2fefa-ovsdbserver-sb\") pod \"dnsmasq-dns-5b946c75cc-6mzz4\" (UID: \"3e491cbf-b795-4735-a356-c30d79d2fefa\") " pod="openstack/dnsmasq-dns-5b946c75cc-6mzz4" Jan 21 11:26:04 crc kubenswrapper[4824]: I0121 11:26:04.589113 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3e491cbf-b795-4735-a356-c30d79d2fefa-dns-svc\") pod \"dnsmasq-dns-5b946c75cc-6mzz4\" (UID: \"3e491cbf-b795-4735-a356-c30d79d2fefa\") " pod="openstack/dnsmasq-dns-5b946c75cc-6mzz4" Jan 21 11:26:04 crc kubenswrapper[4824]: I0121 11:26:04.589156 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3e491cbf-b795-4735-a356-c30d79d2fefa-config\") pod \"dnsmasq-dns-5b946c75cc-6mzz4\" (UID: \"3e491cbf-b795-4735-a356-c30d79d2fefa\") " pod="openstack/dnsmasq-dns-5b946c75cc-6mzz4" Jan 21 11:26:04 crc kubenswrapper[4824]: I0121 11:26:04.589215 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3e491cbf-b795-4735-a356-c30d79d2fefa-ovsdbserver-nb\") pod \"dnsmasq-dns-5b946c75cc-6mzz4\" (UID: \"3e491cbf-b795-4735-a356-c30d79d2fefa\") " pod="openstack/dnsmasq-dns-5b946c75cc-6mzz4" Jan 21 11:26:04 crc kubenswrapper[4824]: I0121 11:26:04.589987 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3e491cbf-b795-4735-a356-c30d79d2fefa-ovsdbserver-sb\") pod \"dnsmasq-dns-5b946c75cc-6mzz4\" (UID: \"3e491cbf-b795-4735-a356-c30d79d2fefa\") " pod="openstack/dnsmasq-dns-5b946c75cc-6mzz4" Jan 21 11:26:04 crc kubenswrapper[4824]: I0121 11:26:04.590009 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3e491cbf-b795-4735-a356-c30d79d2fefa-dns-svc\") pod \"dnsmasq-dns-5b946c75cc-6mzz4\" (UID: \"3e491cbf-b795-4735-a356-c30d79d2fefa\") " pod="openstack/dnsmasq-dns-5b946c75cc-6mzz4" Jan 21 11:26:04 crc kubenswrapper[4824]: I0121 11:26:04.590094 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3e491cbf-b795-4735-a356-c30d79d2fefa-config\") pod \"dnsmasq-dns-5b946c75cc-6mzz4\" (UID: \"3e491cbf-b795-4735-a356-c30d79d2fefa\") " pod="openstack/dnsmasq-dns-5b946c75cc-6mzz4" Jan 21 11:26:04 crc kubenswrapper[4824]: I0121 11:26:04.590102 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3e491cbf-b795-4735-a356-c30d79d2fefa-ovsdbserver-nb\") pod \"dnsmasq-dns-5b946c75cc-6mzz4\" (UID: \"3e491cbf-b795-4735-a356-c30d79d2fefa\") " pod="openstack/dnsmasq-dns-5b946c75cc-6mzz4" Jan 21 11:26:04 crc kubenswrapper[4824]: I0121 11:26:04.605356 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fpg89\" (UniqueName: \"kubernetes.io/projected/3e491cbf-b795-4735-a356-c30d79d2fefa-kube-api-access-fpg89\") pod \"dnsmasq-dns-5b946c75cc-6mzz4\" (UID: \"3e491cbf-b795-4735-a356-c30d79d2fefa\") " pod="openstack/dnsmasq-dns-5b946c75cc-6mzz4" Jan 21 11:26:04 crc kubenswrapper[4824]: I0121 11:26:04.701243 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b946c75cc-6mzz4" Jan 21 11:26:05 crc kubenswrapper[4824]: I0121 11:26:05.061500 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"f980e493-ca74-47f1-8f6f-aca9c9011bf5","Type":"ContainerStarted","Data":"19a1cf9e015e25e337fcd599f29b0b6c4f318463d25fc0a8e8b038d88277f73f"} Jan 21 11:26:05 crc kubenswrapper[4824]: I0121 11:26:05.061711 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"f980e493-ca74-47f1-8f6f-aca9c9011bf5","Type":"ContainerStarted","Data":"8f762056829510d59deced7738a3ca721ea70f910b73bcb4d792931d7b694216"} Jan 21 11:26:05 crc kubenswrapper[4824]: I0121 11:26:05.061722 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"f980e493-ca74-47f1-8f6f-aca9c9011bf5","Type":"ContainerStarted","Data":"ebfa28688dafa5bd49a43d8c94f03fab1ec7dfa1926981e4bcf2dade062d2ccf"} Jan 21 11:26:05 crc kubenswrapper[4824]: I0121 11:26:05.061729 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"f980e493-ca74-47f1-8f6f-aca9c9011bf5","Type":"ContainerStarted","Data":"6c4e2b6ccc2b96fd74c1d55a4f0ff2eb7064ee3cd7ca9e89713240c4c149439d"} Jan 21 11:26:05 crc kubenswrapper[4824]: I0121 11:26:05.079514 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5b946c75cc-6mzz4"] Jan 21 11:26:05 crc kubenswrapper[4824]: I0121 11:26:05.087361 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-storage-0" podStartSLOduration=23.215689231 podStartE2EDuration="29.087346411s" podCreationTimestamp="2026-01-21 11:25:36 +0000 UTC" firstStartedPulling="2026-01-21 11:25:56.493180239 +0000 UTC m=+898.786209531" lastFinishedPulling="2026-01-21 11:26:02.364837419 +0000 UTC m=+904.657866711" observedRunningTime="2026-01-21 11:26:05.083921436 +0000 UTC m=+907.376950727" watchObservedRunningTime="2026-01-21 11:26:05.087346411 +0000 UTC m=+907.380375703" Jan 21 11:26:05 crc kubenswrapper[4824]: I0121 11:26:05.136778 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-htwd6" Jan 21 11:26:05 crc kubenswrapper[4824]: I0121 11:26:05.366966 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5b946c75cc-6mzz4"] Jan 21 11:26:05 crc kubenswrapper[4824]: I0121 11:26:05.400993 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-sj79w"] Jan 21 11:26:05 crc kubenswrapper[4824]: I0121 11:26:05.402027 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74f6bcbc87-sj79w" Jan 21 11:26:05 crc kubenswrapper[4824]: I0121 11:26:05.404851 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-0" Jan 21 11:26:05 crc kubenswrapper[4824]: I0121 11:26:05.419089 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-sj79w"] Jan 21 11:26:05 crc kubenswrapper[4824]: I0121 11:26:05.504341 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c4a81903-5e1f-4d9c-bacf-60639069db18-ovsdbserver-sb\") pod \"dnsmasq-dns-74f6bcbc87-sj79w\" (UID: \"c4a81903-5e1f-4d9c-bacf-60639069db18\") " pod="openstack/dnsmasq-dns-74f6bcbc87-sj79w" Jan 21 11:26:05 crc kubenswrapper[4824]: I0121 11:26:05.504537 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c4a81903-5e1f-4d9c-bacf-60639069db18-dns-swift-storage-0\") pod \"dnsmasq-dns-74f6bcbc87-sj79w\" (UID: \"c4a81903-5e1f-4d9c-bacf-60639069db18\") " pod="openstack/dnsmasq-dns-74f6bcbc87-sj79w" Jan 21 11:26:05 crc kubenswrapper[4824]: I0121 11:26:05.504597 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c4a81903-5e1f-4d9c-bacf-60639069db18-ovsdbserver-nb\") pod \"dnsmasq-dns-74f6bcbc87-sj79w\" (UID: \"c4a81903-5e1f-4d9c-bacf-60639069db18\") " pod="openstack/dnsmasq-dns-74f6bcbc87-sj79w" Jan 21 11:26:05 crc kubenswrapper[4824]: I0121 11:26:05.504639 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c4a81903-5e1f-4d9c-bacf-60639069db18-dns-svc\") pod \"dnsmasq-dns-74f6bcbc87-sj79w\" (UID: \"c4a81903-5e1f-4d9c-bacf-60639069db18\") " pod="openstack/dnsmasq-dns-74f6bcbc87-sj79w" Jan 21 11:26:05 crc kubenswrapper[4824]: I0121 11:26:05.504781 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9xgrw\" (UniqueName: \"kubernetes.io/projected/c4a81903-5e1f-4d9c-bacf-60639069db18-kube-api-access-9xgrw\") pod \"dnsmasq-dns-74f6bcbc87-sj79w\" (UID: \"c4a81903-5e1f-4d9c-bacf-60639069db18\") " pod="openstack/dnsmasq-dns-74f6bcbc87-sj79w" Jan 21 11:26:05 crc kubenswrapper[4824]: I0121 11:26:05.504820 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c4a81903-5e1f-4d9c-bacf-60639069db18-config\") pod \"dnsmasq-dns-74f6bcbc87-sj79w\" (UID: \"c4a81903-5e1f-4d9c-bacf-60639069db18\") " pod="openstack/dnsmasq-dns-74f6bcbc87-sj79w" Jan 21 11:26:05 crc kubenswrapper[4824]: I0121 11:26:05.606204 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c4a81903-5e1f-4d9c-bacf-60639069db18-dns-swift-storage-0\") pod \"dnsmasq-dns-74f6bcbc87-sj79w\" (UID: \"c4a81903-5e1f-4d9c-bacf-60639069db18\") " pod="openstack/dnsmasq-dns-74f6bcbc87-sj79w" Jan 21 11:26:05 crc kubenswrapper[4824]: I0121 11:26:05.606255 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c4a81903-5e1f-4d9c-bacf-60639069db18-ovsdbserver-nb\") pod \"dnsmasq-dns-74f6bcbc87-sj79w\" (UID: \"c4a81903-5e1f-4d9c-bacf-60639069db18\") " pod="openstack/dnsmasq-dns-74f6bcbc87-sj79w" Jan 21 11:26:05 crc kubenswrapper[4824]: I0121 11:26:05.606283 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c4a81903-5e1f-4d9c-bacf-60639069db18-dns-svc\") pod \"dnsmasq-dns-74f6bcbc87-sj79w\" (UID: \"c4a81903-5e1f-4d9c-bacf-60639069db18\") " pod="openstack/dnsmasq-dns-74f6bcbc87-sj79w" Jan 21 11:26:05 crc kubenswrapper[4824]: I0121 11:26:05.606341 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9xgrw\" (UniqueName: \"kubernetes.io/projected/c4a81903-5e1f-4d9c-bacf-60639069db18-kube-api-access-9xgrw\") pod \"dnsmasq-dns-74f6bcbc87-sj79w\" (UID: \"c4a81903-5e1f-4d9c-bacf-60639069db18\") " pod="openstack/dnsmasq-dns-74f6bcbc87-sj79w" Jan 21 11:26:05 crc kubenswrapper[4824]: I0121 11:26:05.606370 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c4a81903-5e1f-4d9c-bacf-60639069db18-config\") pod \"dnsmasq-dns-74f6bcbc87-sj79w\" (UID: \"c4a81903-5e1f-4d9c-bacf-60639069db18\") " pod="openstack/dnsmasq-dns-74f6bcbc87-sj79w" Jan 21 11:26:05 crc kubenswrapper[4824]: I0121 11:26:05.606423 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c4a81903-5e1f-4d9c-bacf-60639069db18-ovsdbserver-sb\") pod \"dnsmasq-dns-74f6bcbc87-sj79w\" (UID: \"c4a81903-5e1f-4d9c-bacf-60639069db18\") " pod="openstack/dnsmasq-dns-74f6bcbc87-sj79w" Jan 21 11:26:05 crc kubenswrapper[4824]: I0121 11:26:05.607191 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c4a81903-5e1f-4d9c-bacf-60639069db18-ovsdbserver-sb\") pod \"dnsmasq-dns-74f6bcbc87-sj79w\" (UID: \"c4a81903-5e1f-4d9c-bacf-60639069db18\") " pod="openstack/dnsmasq-dns-74f6bcbc87-sj79w" Jan 21 11:26:05 crc kubenswrapper[4824]: I0121 11:26:05.607190 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c4a81903-5e1f-4d9c-bacf-60639069db18-dns-svc\") pod \"dnsmasq-dns-74f6bcbc87-sj79w\" (UID: \"c4a81903-5e1f-4d9c-bacf-60639069db18\") " pod="openstack/dnsmasq-dns-74f6bcbc87-sj79w" Jan 21 11:26:05 crc kubenswrapper[4824]: I0121 11:26:05.607234 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c4a81903-5e1f-4d9c-bacf-60639069db18-config\") pod \"dnsmasq-dns-74f6bcbc87-sj79w\" (UID: \"c4a81903-5e1f-4d9c-bacf-60639069db18\") " pod="openstack/dnsmasq-dns-74f6bcbc87-sj79w" Jan 21 11:26:05 crc kubenswrapper[4824]: I0121 11:26:05.607338 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c4a81903-5e1f-4d9c-bacf-60639069db18-ovsdbserver-nb\") pod \"dnsmasq-dns-74f6bcbc87-sj79w\" (UID: \"c4a81903-5e1f-4d9c-bacf-60639069db18\") " pod="openstack/dnsmasq-dns-74f6bcbc87-sj79w" Jan 21 11:26:05 crc kubenswrapper[4824]: I0121 11:26:05.607902 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c4a81903-5e1f-4d9c-bacf-60639069db18-dns-swift-storage-0\") pod \"dnsmasq-dns-74f6bcbc87-sj79w\" (UID: \"c4a81903-5e1f-4d9c-bacf-60639069db18\") " pod="openstack/dnsmasq-dns-74f6bcbc87-sj79w" Jan 21 11:26:05 crc kubenswrapper[4824]: I0121 11:26:05.620170 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9xgrw\" (UniqueName: \"kubernetes.io/projected/c4a81903-5e1f-4d9c-bacf-60639069db18-kube-api-access-9xgrw\") pod \"dnsmasq-dns-74f6bcbc87-sj79w\" (UID: \"c4a81903-5e1f-4d9c-bacf-60639069db18\") " pod="openstack/dnsmasq-dns-74f6bcbc87-sj79w" Jan 21 11:26:05 crc kubenswrapper[4824]: I0121 11:26:05.717566 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74f6bcbc87-sj79w" Jan 21 11:26:06 crc kubenswrapper[4824]: I0121 11:26:06.056633 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6d3982e6-513d-454f-a08a-34b32779b559" path="/var/lib/kubelet/pods/6d3982e6-513d-454f-a08a-34b32779b559/volumes" Jan 21 11:26:06 crc kubenswrapper[4824]: I0121 11:26:06.068267 4824 generic.go:334] "Generic (PLEG): container finished" podID="3e491cbf-b795-4735-a356-c30d79d2fefa" containerID="a758213be7540d83b553495253277f3c7a2c31c59873892123ffe6b48a3b966f" exitCode=0 Jan 21 11:26:06 crc kubenswrapper[4824]: I0121 11:26:06.068362 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b946c75cc-6mzz4" event={"ID":"3e491cbf-b795-4735-a356-c30d79d2fefa","Type":"ContainerDied","Data":"a758213be7540d83b553495253277f3c7a2c31c59873892123ffe6b48a3b966f"} Jan 21 11:26:06 crc kubenswrapper[4824]: I0121 11:26:06.068401 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b946c75cc-6mzz4" event={"ID":"3e491cbf-b795-4735-a356-c30d79d2fefa","Type":"ContainerStarted","Data":"c2adb67835a6dab5ee88f8c74c3273f9d6eb5ecef5ab9da3b16f2dada3d6a2a1"} Jan 21 11:26:06 crc kubenswrapper[4824]: I0121 11:26:06.115876 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-sj79w"] Jan 21 11:26:06 crc kubenswrapper[4824]: I0121 11:26:06.830501 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-b795q" Jan 21 11:26:06 crc kubenswrapper[4824]: I0121 11:26:06.830701 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-b795q" Jan 21 11:26:06 crc kubenswrapper[4824]: I0121 11:26:06.863239 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-b795q" Jan 21 11:26:07 crc kubenswrapper[4824]: I0121 11:26:07.074564 4824 generic.go:334] "Generic (PLEG): container finished" podID="c4a81903-5e1f-4d9c-bacf-60639069db18" containerID="0aacefa5ca6057f616f2afc320e43a0e95b25311f579f1cc69ade834fb63b2f0" exitCode=0 Jan 21 11:26:07 crc kubenswrapper[4824]: I0121 11:26:07.074629 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74f6bcbc87-sj79w" event={"ID":"c4a81903-5e1f-4d9c-bacf-60639069db18","Type":"ContainerDied","Data":"0aacefa5ca6057f616f2afc320e43a0e95b25311f579f1cc69ade834fb63b2f0"} Jan 21 11:26:07 crc kubenswrapper[4824]: I0121 11:26:07.074654 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74f6bcbc87-sj79w" event={"ID":"c4a81903-5e1f-4d9c-bacf-60639069db18","Type":"ContainerStarted","Data":"19d64cd123fe9ef3646e70cf5808fefa9acba69592569cd21640e9399cbca5d9"} Jan 21 11:26:07 crc kubenswrapper[4824]: I0121 11:26:07.076133 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b946c75cc-6mzz4" event={"ID":"3e491cbf-b795-4735-a356-c30d79d2fefa","Type":"ContainerStarted","Data":"481485ca50c37a441b7a8a265c33cb37a4b1614ddb93a2f74113ad330f10dc6f"} Jan 21 11:26:07 crc kubenswrapper[4824]: I0121 11:26:07.076210 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5b946c75cc-6mzz4" podUID="3e491cbf-b795-4735-a356-c30d79d2fefa" containerName="dnsmasq-dns" containerID="cri-o://481485ca50c37a441b7a8a265c33cb37a4b1614ddb93a2f74113ad330f10dc6f" gracePeriod=10 Jan 21 11:26:07 crc kubenswrapper[4824]: I0121 11:26:07.076707 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5b946c75cc-6mzz4" Jan 21 11:26:07 crc kubenswrapper[4824]: I0121 11:26:07.102921 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5b946c75cc-6mzz4" podStartSLOduration=3.102904885 podStartE2EDuration="3.102904885s" podCreationTimestamp="2026-01-21 11:26:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:26:07.101426197 +0000 UTC m=+909.394455489" watchObservedRunningTime="2026-01-21 11:26:07.102904885 +0000 UTC m=+909.395934176" Jan 21 11:26:07 crc kubenswrapper[4824]: I0121 11:26:07.110598 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-b795q" Jan 21 11:26:07 crc kubenswrapper[4824]: I0121 11:26:07.158504 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-b795q"] Jan 21 11:26:07 crc kubenswrapper[4824]: I0121 11:26:07.428407 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b946c75cc-6mzz4" Jan 21 11:26:07 crc kubenswrapper[4824]: I0121 11:26:07.534510 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3e491cbf-b795-4735-a356-c30d79d2fefa-config\") pod \"3e491cbf-b795-4735-a356-c30d79d2fefa\" (UID: \"3e491cbf-b795-4735-a356-c30d79d2fefa\") " Jan 21 11:26:07 crc kubenswrapper[4824]: I0121 11:26:07.534616 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3e491cbf-b795-4735-a356-c30d79d2fefa-ovsdbserver-nb\") pod \"3e491cbf-b795-4735-a356-c30d79d2fefa\" (UID: \"3e491cbf-b795-4735-a356-c30d79d2fefa\") " Jan 21 11:26:07 crc kubenswrapper[4824]: I0121 11:26:07.534646 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3e491cbf-b795-4735-a356-c30d79d2fefa-dns-svc\") pod \"3e491cbf-b795-4735-a356-c30d79d2fefa\" (UID: \"3e491cbf-b795-4735-a356-c30d79d2fefa\") " Jan 21 11:26:07 crc kubenswrapper[4824]: I0121 11:26:07.534688 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3e491cbf-b795-4735-a356-c30d79d2fefa-ovsdbserver-sb\") pod \"3e491cbf-b795-4735-a356-c30d79d2fefa\" (UID: \"3e491cbf-b795-4735-a356-c30d79d2fefa\") " Jan 21 11:26:07 crc kubenswrapper[4824]: I0121 11:26:07.534780 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fpg89\" (UniqueName: \"kubernetes.io/projected/3e491cbf-b795-4735-a356-c30d79d2fefa-kube-api-access-fpg89\") pod \"3e491cbf-b795-4735-a356-c30d79d2fefa\" (UID: \"3e491cbf-b795-4735-a356-c30d79d2fefa\") " Jan 21 11:26:07 crc kubenswrapper[4824]: I0121 11:26:07.538042 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3e491cbf-b795-4735-a356-c30d79d2fefa-kube-api-access-fpg89" (OuterVolumeSpecName: "kube-api-access-fpg89") pod "3e491cbf-b795-4735-a356-c30d79d2fefa" (UID: "3e491cbf-b795-4735-a356-c30d79d2fefa"). InnerVolumeSpecName "kube-api-access-fpg89". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:26:07 crc kubenswrapper[4824]: I0121 11:26:07.563485 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3e491cbf-b795-4735-a356-c30d79d2fefa-config" (OuterVolumeSpecName: "config") pod "3e491cbf-b795-4735-a356-c30d79d2fefa" (UID: "3e491cbf-b795-4735-a356-c30d79d2fefa"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:26:07 crc kubenswrapper[4824]: I0121 11:26:07.563804 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3e491cbf-b795-4735-a356-c30d79d2fefa-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "3e491cbf-b795-4735-a356-c30d79d2fefa" (UID: "3e491cbf-b795-4735-a356-c30d79d2fefa"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:26:07 crc kubenswrapper[4824]: I0121 11:26:07.568488 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3e491cbf-b795-4735-a356-c30d79d2fefa-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "3e491cbf-b795-4735-a356-c30d79d2fefa" (UID: "3e491cbf-b795-4735-a356-c30d79d2fefa"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:26:07 crc kubenswrapper[4824]: I0121 11:26:07.569377 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3e491cbf-b795-4735-a356-c30d79d2fefa-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "3e491cbf-b795-4735-a356-c30d79d2fefa" (UID: "3e491cbf-b795-4735-a356-c30d79d2fefa"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:26:07 crc kubenswrapper[4824]: I0121 11:26:07.636795 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3e491cbf-b795-4735-a356-c30d79d2fefa-config\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:07 crc kubenswrapper[4824]: I0121 11:26:07.637036 4824 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3e491cbf-b795-4735-a356-c30d79d2fefa-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:07 crc kubenswrapper[4824]: I0121 11:26:07.637116 4824 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3e491cbf-b795-4735-a356-c30d79d2fefa-dns-svc\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:07 crc kubenswrapper[4824]: I0121 11:26:07.637178 4824 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3e491cbf-b795-4735-a356-c30d79d2fefa-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:07 crc kubenswrapper[4824]: I0121 11:26:07.637240 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fpg89\" (UniqueName: \"kubernetes.io/projected/3e491cbf-b795-4735-a356-c30d79d2fefa-kube-api-access-fpg89\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:08 crc kubenswrapper[4824]: I0121 11:26:08.083725 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74f6bcbc87-sj79w" event={"ID":"c4a81903-5e1f-4d9c-bacf-60639069db18","Type":"ContainerStarted","Data":"862050906c3faeb783c3cf1a55874b97964bf3e73dcab5972c6e49f1ea08c530"} Jan 21 11:26:08 crc kubenswrapper[4824]: I0121 11:26:08.083787 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-74f6bcbc87-sj79w" Jan 21 11:26:08 crc kubenswrapper[4824]: I0121 11:26:08.086124 4824 generic.go:334] "Generic (PLEG): container finished" podID="3e491cbf-b795-4735-a356-c30d79d2fefa" containerID="481485ca50c37a441b7a8a265c33cb37a4b1614ddb93a2f74113ad330f10dc6f" exitCode=0 Jan 21 11:26:08 crc kubenswrapper[4824]: I0121 11:26:08.086179 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b946c75cc-6mzz4" Jan 21 11:26:08 crc kubenswrapper[4824]: I0121 11:26:08.086213 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b946c75cc-6mzz4" event={"ID":"3e491cbf-b795-4735-a356-c30d79d2fefa","Type":"ContainerDied","Data":"481485ca50c37a441b7a8a265c33cb37a4b1614ddb93a2f74113ad330f10dc6f"} Jan 21 11:26:08 crc kubenswrapper[4824]: I0121 11:26:08.086237 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b946c75cc-6mzz4" event={"ID":"3e491cbf-b795-4735-a356-c30d79d2fefa","Type":"ContainerDied","Data":"c2adb67835a6dab5ee88f8c74c3273f9d6eb5ecef5ab9da3b16f2dada3d6a2a1"} Jan 21 11:26:08 crc kubenswrapper[4824]: I0121 11:26:08.086253 4824 scope.go:117] "RemoveContainer" containerID="481485ca50c37a441b7a8a265c33cb37a4b1614ddb93a2f74113ad330f10dc6f" Jan 21 11:26:08 crc kubenswrapper[4824]: I0121 11:26:08.101217 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-74f6bcbc87-sj79w" podStartSLOduration=3.101204052 podStartE2EDuration="3.101204052s" podCreationTimestamp="2026-01-21 11:26:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:26:08.099172212 +0000 UTC m=+910.392201505" watchObservedRunningTime="2026-01-21 11:26:08.101204052 +0000 UTC m=+910.394233344" Jan 21 11:26:08 crc kubenswrapper[4824]: I0121 11:26:08.106224 4824 scope.go:117] "RemoveContainer" containerID="a758213be7540d83b553495253277f3c7a2c31c59873892123ffe6b48a3b966f" Jan 21 11:26:08 crc kubenswrapper[4824]: I0121 11:26:08.122526 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5b946c75cc-6mzz4"] Jan 21 11:26:08 crc kubenswrapper[4824]: I0121 11:26:08.126647 4824 scope.go:117] "RemoveContainer" containerID="481485ca50c37a441b7a8a265c33cb37a4b1614ddb93a2f74113ad330f10dc6f" Jan 21 11:26:08 crc kubenswrapper[4824]: E0121 11:26:08.127062 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"481485ca50c37a441b7a8a265c33cb37a4b1614ddb93a2f74113ad330f10dc6f\": container with ID starting with 481485ca50c37a441b7a8a265c33cb37a4b1614ddb93a2f74113ad330f10dc6f not found: ID does not exist" containerID="481485ca50c37a441b7a8a265c33cb37a4b1614ddb93a2f74113ad330f10dc6f" Jan 21 11:26:08 crc kubenswrapper[4824]: I0121 11:26:08.127092 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"481485ca50c37a441b7a8a265c33cb37a4b1614ddb93a2f74113ad330f10dc6f"} err="failed to get container status \"481485ca50c37a441b7a8a265c33cb37a4b1614ddb93a2f74113ad330f10dc6f\": rpc error: code = NotFound desc = could not find container \"481485ca50c37a441b7a8a265c33cb37a4b1614ddb93a2f74113ad330f10dc6f\": container with ID starting with 481485ca50c37a441b7a8a265c33cb37a4b1614ddb93a2f74113ad330f10dc6f not found: ID does not exist" Jan 21 11:26:08 crc kubenswrapper[4824]: I0121 11:26:08.127112 4824 scope.go:117] "RemoveContainer" containerID="a758213be7540d83b553495253277f3c7a2c31c59873892123ffe6b48a3b966f" Jan 21 11:26:08 crc kubenswrapper[4824]: E0121 11:26:08.127343 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a758213be7540d83b553495253277f3c7a2c31c59873892123ffe6b48a3b966f\": container with ID starting with a758213be7540d83b553495253277f3c7a2c31c59873892123ffe6b48a3b966f not found: ID does not exist" containerID="a758213be7540d83b553495253277f3c7a2c31c59873892123ffe6b48a3b966f" Jan 21 11:26:08 crc kubenswrapper[4824]: I0121 11:26:08.127361 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a758213be7540d83b553495253277f3c7a2c31c59873892123ffe6b48a3b966f"} err="failed to get container status \"a758213be7540d83b553495253277f3c7a2c31c59873892123ffe6b48a3b966f\": rpc error: code = NotFound desc = could not find container \"a758213be7540d83b553495253277f3c7a2c31c59873892123ffe6b48a3b966f\": container with ID starting with a758213be7540d83b553495253277f3c7a2c31c59873892123ffe6b48a3b966f not found: ID does not exist" Jan 21 11:26:08 crc kubenswrapper[4824]: I0121 11:26:08.128100 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5b946c75cc-6mzz4"] Jan 21 11:26:09 crc kubenswrapper[4824]: I0121 11:26:09.093170 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-b795q" podUID="1c75f0f9-53e1-4a64-9211-559ef51c1b53" containerName="registry-server" containerID="cri-o://9e1d278d517cd8394b73f2d45c4430f8e3523f06927bcb011e76233b05081e45" gracePeriod=2 Jan 21 11:26:09 crc kubenswrapper[4824]: I0121 11:26:09.470513 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-b795q" Jan 21 11:26:09 crc kubenswrapper[4824]: I0121 11:26:09.566771 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1c75f0f9-53e1-4a64-9211-559ef51c1b53-utilities\") pod \"1c75f0f9-53e1-4a64-9211-559ef51c1b53\" (UID: \"1c75f0f9-53e1-4a64-9211-559ef51c1b53\") " Jan 21 11:26:09 crc kubenswrapper[4824]: I0121 11:26:09.566834 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1c75f0f9-53e1-4a64-9211-559ef51c1b53-catalog-content\") pod \"1c75f0f9-53e1-4a64-9211-559ef51c1b53\" (UID: \"1c75f0f9-53e1-4a64-9211-559ef51c1b53\") " Jan 21 11:26:09 crc kubenswrapper[4824]: I0121 11:26:09.566913 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hd85x\" (UniqueName: \"kubernetes.io/projected/1c75f0f9-53e1-4a64-9211-559ef51c1b53-kube-api-access-hd85x\") pod \"1c75f0f9-53e1-4a64-9211-559ef51c1b53\" (UID: \"1c75f0f9-53e1-4a64-9211-559ef51c1b53\") " Jan 21 11:26:09 crc kubenswrapper[4824]: I0121 11:26:09.567443 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1c75f0f9-53e1-4a64-9211-559ef51c1b53-utilities" (OuterVolumeSpecName: "utilities") pod "1c75f0f9-53e1-4a64-9211-559ef51c1b53" (UID: "1c75f0f9-53e1-4a64-9211-559ef51c1b53"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:26:09 crc kubenswrapper[4824]: I0121 11:26:09.571399 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1c75f0f9-53e1-4a64-9211-559ef51c1b53-kube-api-access-hd85x" (OuterVolumeSpecName: "kube-api-access-hd85x") pod "1c75f0f9-53e1-4a64-9211-559ef51c1b53" (UID: "1c75f0f9-53e1-4a64-9211-559ef51c1b53"). InnerVolumeSpecName "kube-api-access-hd85x". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:26:09 crc kubenswrapper[4824]: I0121 11:26:09.582568 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1c75f0f9-53e1-4a64-9211-559ef51c1b53-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1c75f0f9-53e1-4a64-9211-559ef51c1b53" (UID: "1c75f0f9-53e1-4a64-9211-559ef51c1b53"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:26:09 crc kubenswrapper[4824]: I0121 11:26:09.668741 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1c75f0f9-53e1-4a64-9211-559ef51c1b53-utilities\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:09 crc kubenswrapper[4824]: I0121 11:26:09.668775 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1c75f0f9-53e1-4a64-9211-559ef51c1b53-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:09 crc kubenswrapper[4824]: I0121 11:26:09.668786 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hd85x\" (UniqueName: \"kubernetes.io/projected/1c75f0f9-53e1-4a64-9211-559ef51c1b53-kube-api-access-hd85x\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:10 crc kubenswrapper[4824]: I0121 11:26:10.056572 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3e491cbf-b795-4735-a356-c30d79d2fefa" path="/var/lib/kubelet/pods/3e491cbf-b795-4735-a356-c30d79d2fefa/volumes" Jan 21 11:26:10 crc kubenswrapper[4824]: I0121 11:26:10.100338 4824 generic.go:334] "Generic (PLEG): container finished" podID="1c75f0f9-53e1-4a64-9211-559ef51c1b53" containerID="9e1d278d517cd8394b73f2d45c4430f8e3523f06927bcb011e76233b05081e45" exitCode=0 Jan 21 11:26:10 crc kubenswrapper[4824]: I0121 11:26:10.100373 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b795q" event={"ID":"1c75f0f9-53e1-4a64-9211-559ef51c1b53","Type":"ContainerDied","Data":"9e1d278d517cd8394b73f2d45c4430f8e3523f06927bcb011e76233b05081e45"} Jan 21 11:26:10 crc kubenswrapper[4824]: I0121 11:26:10.100383 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-b795q" Jan 21 11:26:10 crc kubenswrapper[4824]: I0121 11:26:10.100396 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b795q" event={"ID":"1c75f0f9-53e1-4a64-9211-559ef51c1b53","Type":"ContainerDied","Data":"799120a9a7a20b359b616c2030579c280c96e8a87a304704ae580c5acba60e25"} Jan 21 11:26:10 crc kubenswrapper[4824]: I0121 11:26:10.100411 4824 scope.go:117] "RemoveContainer" containerID="9e1d278d517cd8394b73f2d45c4430f8e3523f06927bcb011e76233b05081e45" Jan 21 11:26:10 crc kubenswrapper[4824]: I0121 11:26:10.115825 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-b795q"] Jan 21 11:26:10 crc kubenswrapper[4824]: I0121 11:26:10.117094 4824 scope.go:117] "RemoveContainer" containerID="af83a2d8b2c7751e366bf2114a7547afc9bf4e4dc745f7e937384620bcdb2e71" Jan 21 11:26:10 crc kubenswrapper[4824]: I0121 11:26:10.120579 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-b795q"] Jan 21 11:26:10 crc kubenswrapper[4824]: I0121 11:26:10.133688 4824 scope.go:117] "RemoveContainer" containerID="dc92b9cbcf516998ce8a3e5f19a7012166529b9eee969255e60da7b567e11c76" Jan 21 11:26:10 crc kubenswrapper[4824]: I0121 11:26:10.155401 4824 scope.go:117] "RemoveContainer" containerID="9e1d278d517cd8394b73f2d45c4430f8e3523f06927bcb011e76233b05081e45" Jan 21 11:26:10 crc kubenswrapper[4824]: E0121 11:26:10.155821 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9e1d278d517cd8394b73f2d45c4430f8e3523f06927bcb011e76233b05081e45\": container with ID starting with 9e1d278d517cd8394b73f2d45c4430f8e3523f06927bcb011e76233b05081e45 not found: ID does not exist" containerID="9e1d278d517cd8394b73f2d45c4430f8e3523f06927bcb011e76233b05081e45" Jan 21 11:26:10 crc kubenswrapper[4824]: I0121 11:26:10.155857 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9e1d278d517cd8394b73f2d45c4430f8e3523f06927bcb011e76233b05081e45"} err="failed to get container status \"9e1d278d517cd8394b73f2d45c4430f8e3523f06927bcb011e76233b05081e45\": rpc error: code = NotFound desc = could not find container \"9e1d278d517cd8394b73f2d45c4430f8e3523f06927bcb011e76233b05081e45\": container with ID starting with 9e1d278d517cd8394b73f2d45c4430f8e3523f06927bcb011e76233b05081e45 not found: ID does not exist" Jan 21 11:26:10 crc kubenswrapper[4824]: I0121 11:26:10.155880 4824 scope.go:117] "RemoveContainer" containerID="af83a2d8b2c7751e366bf2114a7547afc9bf4e4dc745f7e937384620bcdb2e71" Jan 21 11:26:10 crc kubenswrapper[4824]: E0121 11:26:10.156222 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"af83a2d8b2c7751e366bf2114a7547afc9bf4e4dc745f7e937384620bcdb2e71\": container with ID starting with af83a2d8b2c7751e366bf2114a7547afc9bf4e4dc745f7e937384620bcdb2e71 not found: ID does not exist" containerID="af83a2d8b2c7751e366bf2114a7547afc9bf4e4dc745f7e937384620bcdb2e71" Jan 21 11:26:10 crc kubenswrapper[4824]: I0121 11:26:10.156262 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"af83a2d8b2c7751e366bf2114a7547afc9bf4e4dc745f7e937384620bcdb2e71"} err="failed to get container status \"af83a2d8b2c7751e366bf2114a7547afc9bf4e4dc745f7e937384620bcdb2e71\": rpc error: code = NotFound desc = could not find container \"af83a2d8b2c7751e366bf2114a7547afc9bf4e4dc745f7e937384620bcdb2e71\": container with ID starting with af83a2d8b2c7751e366bf2114a7547afc9bf4e4dc745f7e937384620bcdb2e71 not found: ID does not exist" Jan 21 11:26:10 crc kubenswrapper[4824]: I0121 11:26:10.156289 4824 scope.go:117] "RemoveContainer" containerID="dc92b9cbcf516998ce8a3e5f19a7012166529b9eee969255e60da7b567e11c76" Jan 21 11:26:10 crc kubenswrapper[4824]: E0121 11:26:10.156566 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dc92b9cbcf516998ce8a3e5f19a7012166529b9eee969255e60da7b567e11c76\": container with ID starting with dc92b9cbcf516998ce8a3e5f19a7012166529b9eee969255e60da7b567e11c76 not found: ID does not exist" containerID="dc92b9cbcf516998ce8a3e5f19a7012166529b9eee969255e60da7b567e11c76" Jan 21 11:26:10 crc kubenswrapper[4824]: I0121 11:26:10.156593 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dc92b9cbcf516998ce8a3e5f19a7012166529b9eee969255e60da7b567e11c76"} err="failed to get container status \"dc92b9cbcf516998ce8a3e5f19a7012166529b9eee969255e60da7b567e11c76\": rpc error: code = NotFound desc = could not find container \"dc92b9cbcf516998ce8a3e5f19a7012166529b9eee969255e60da7b567e11c76\": container with ID starting with dc92b9cbcf516998ce8a3e5f19a7012166529b9eee969255e60da7b567e11c76 not found: ID does not exist" Jan 21 11:26:11 crc kubenswrapper[4824]: I0121 11:26:11.376139 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Jan 21 11:26:11 crc kubenswrapper[4824]: I0121 11:26:11.633114 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Jan 21 11:26:11 crc kubenswrapper[4824]: I0121 11:26:11.812292 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-db-create-kk8kb"] Jan 21 11:26:11 crc kubenswrapper[4824]: E0121 11:26:11.812557 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1c75f0f9-53e1-4a64-9211-559ef51c1b53" containerName="extract-utilities" Jan 21 11:26:11 crc kubenswrapper[4824]: I0121 11:26:11.812574 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="1c75f0f9-53e1-4a64-9211-559ef51c1b53" containerName="extract-utilities" Jan 21 11:26:11 crc kubenswrapper[4824]: E0121 11:26:11.812584 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1c75f0f9-53e1-4a64-9211-559ef51c1b53" containerName="registry-server" Jan 21 11:26:11 crc kubenswrapper[4824]: I0121 11:26:11.812591 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="1c75f0f9-53e1-4a64-9211-559ef51c1b53" containerName="registry-server" Jan 21 11:26:11 crc kubenswrapper[4824]: E0121 11:26:11.812610 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1c75f0f9-53e1-4a64-9211-559ef51c1b53" containerName="extract-content" Jan 21 11:26:11 crc kubenswrapper[4824]: I0121 11:26:11.812615 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="1c75f0f9-53e1-4a64-9211-559ef51c1b53" containerName="extract-content" Jan 21 11:26:11 crc kubenswrapper[4824]: E0121 11:26:11.812626 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e491cbf-b795-4735-a356-c30d79d2fefa" containerName="init" Jan 21 11:26:11 crc kubenswrapper[4824]: I0121 11:26:11.812631 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e491cbf-b795-4735-a356-c30d79d2fefa" containerName="init" Jan 21 11:26:11 crc kubenswrapper[4824]: E0121 11:26:11.812640 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e491cbf-b795-4735-a356-c30d79d2fefa" containerName="dnsmasq-dns" Jan 21 11:26:11 crc kubenswrapper[4824]: I0121 11:26:11.812645 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e491cbf-b795-4735-a356-c30d79d2fefa" containerName="dnsmasq-dns" Jan 21 11:26:11 crc kubenswrapper[4824]: I0121 11:26:11.812809 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="1c75f0f9-53e1-4a64-9211-559ef51c1b53" containerName="registry-server" Jan 21 11:26:11 crc kubenswrapper[4824]: I0121 11:26:11.812821 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="3e491cbf-b795-4735-a356-c30d79d2fefa" containerName="dnsmasq-dns" Jan 21 11:26:11 crc kubenswrapper[4824]: I0121 11:26:11.813267 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-kk8kb" Jan 21 11:26:11 crc kubenswrapper[4824]: I0121 11:26:11.823549 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-2619-account-create-update-9csh2"] Jan 21 11:26:11 crc kubenswrapper[4824]: I0121 11:26:11.824376 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-2619-account-create-update-9csh2" Jan 21 11:26:11 crc kubenswrapper[4824]: I0121 11:26:11.827914 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-db-secret" Jan 21 11:26:11 crc kubenswrapper[4824]: I0121 11:26:11.835662 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-create-kk8kb"] Jan 21 11:26:11 crc kubenswrapper[4824]: I0121 11:26:11.840730 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-2619-account-create-update-9csh2"] Jan 21 11:26:11 crc kubenswrapper[4824]: I0121 11:26:11.902420 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fe6e7565-793b-4555-b6d5-758f87a5b9c7-operator-scripts\") pod \"heat-2619-account-create-update-9csh2\" (UID: \"fe6e7565-793b-4555-b6d5-758f87a5b9c7\") " pod="openstack/heat-2619-account-create-update-9csh2" Jan 21 11:26:11 crc kubenswrapper[4824]: I0121 11:26:11.902571 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q7phq\" (UniqueName: \"kubernetes.io/projected/8600373d-f466-43d4-92fc-1fa938f6e91b-kube-api-access-q7phq\") pod \"heat-db-create-kk8kb\" (UID: \"8600373d-f466-43d4-92fc-1fa938f6e91b\") " pod="openstack/heat-db-create-kk8kb" Jan 21 11:26:11 crc kubenswrapper[4824]: I0121 11:26:11.902663 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w4p9x\" (UniqueName: \"kubernetes.io/projected/fe6e7565-793b-4555-b6d5-758f87a5b9c7-kube-api-access-w4p9x\") pod \"heat-2619-account-create-update-9csh2\" (UID: \"fe6e7565-793b-4555-b6d5-758f87a5b9c7\") " pod="openstack/heat-2619-account-create-update-9csh2" Jan 21 11:26:11 crc kubenswrapper[4824]: I0121 11:26:11.902742 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8600373d-f466-43d4-92fc-1fa938f6e91b-operator-scripts\") pod \"heat-db-create-kk8kb\" (UID: \"8600373d-f466-43d4-92fc-1fa938f6e91b\") " pod="openstack/heat-db-create-kk8kb" Jan 21 11:26:11 crc kubenswrapper[4824]: I0121 11:26:11.944943 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-84f4b"] Jan 21 11:26:11 crc kubenswrapper[4824]: I0121 11:26:11.945779 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-84f4b" Jan 21 11:26:11 crc kubenswrapper[4824]: I0121 11:26:11.949190 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Jan 21 11:26:11 crc kubenswrapper[4824]: I0121 11:26:11.950121 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Jan 21 11:26:11 crc kubenswrapper[4824]: I0121 11:26:11.950258 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-qxhzv" Jan 21 11:26:11 crc kubenswrapper[4824]: I0121 11:26:11.950274 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Jan 21 11:26:11 crc kubenswrapper[4824]: I0121 11:26:11.986805 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-84f4b"] Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.003885 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q7phq\" (UniqueName: \"kubernetes.io/projected/8600373d-f466-43d4-92fc-1fa938f6e91b-kube-api-access-q7phq\") pod \"heat-db-create-kk8kb\" (UID: \"8600373d-f466-43d4-92fc-1fa938f6e91b\") " pod="openstack/heat-db-create-kk8kb" Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.003930 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w4p9x\" (UniqueName: \"kubernetes.io/projected/fe6e7565-793b-4555-b6d5-758f87a5b9c7-kube-api-access-w4p9x\") pod \"heat-2619-account-create-update-9csh2\" (UID: \"fe6e7565-793b-4555-b6d5-758f87a5b9c7\") " pod="openstack/heat-2619-account-create-update-9csh2" Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.003973 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8600373d-f466-43d4-92fc-1fa938f6e91b-operator-scripts\") pod \"heat-db-create-kk8kb\" (UID: \"8600373d-f466-43d4-92fc-1fa938f6e91b\") " pod="openstack/heat-db-create-kk8kb" Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.004094 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5240d35b-cbf1-472d-91da-debf418dd208-combined-ca-bundle\") pod \"keystone-db-sync-84f4b\" (UID: \"5240d35b-cbf1-472d-91da-debf418dd208\") " pod="openstack/keystone-db-sync-84f4b" Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.004190 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f4nqn\" (UniqueName: \"kubernetes.io/projected/5240d35b-cbf1-472d-91da-debf418dd208-kube-api-access-f4nqn\") pod \"keystone-db-sync-84f4b\" (UID: \"5240d35b-cbf1-472d-91da-debf418dd208\") " pod="openstack/keystone-db-sync-84f4b" Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.004212 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5240d35b-cbf1-472d-91da-debf418dd208-config-data\") pod \"keystone-db-sync-84f4b\" (UID: \"5240d35b-cbf1-472d-91da-debf418dd208\") " pod="openstack/keystone-db-sync-84f4b" Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.004230 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fe6e7565-793b-4555-b6d5-758f87a5b9c7-operator-scripts\") pod \"heat-2619-account-create-update-9csh2\" (UID: \"fe6e7565-793b-4555-b6d5-758f87a5b9c7\") " pod="openstack/heat-2619-account-create-update-9csh2" Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.004706 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8600373d-f466-43d4-92fc-1fa938f6e91b-operator-scripts\") pod \"heat-db-create-kk8kb\" (UID: \"8600373d-f466-43d4-92fc-1fa938f6e91b\") " pod="openstack/heat-db-create-kk8kb" Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.004796 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fe6e7565-793b-4555-b6d5-758f87a5b9c7-operator-scripts\") pod \"heat-2619-account-create-update-9csh2\" (UID: \"fe6e7565-793b-4555-b6d5-758f87a5b9c7\") " pod="openstack/heat-2619-account-create-update-9csh2" Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.008997 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-twgjw"] Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.009915 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-twgjw" Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.018858 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-twgjw"] Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.018902 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-3c9f-account-create-update-rtprr"] Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.019887 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-3c9f-account-create-update-rtprr" Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.024300 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w4p9x\" (UniqueName: \"kubernetes.io/projected/fe6e7565-793b-4555-b6d5-758f87a5b9c7-kube-api-access-w4p9x\") pod \"heat-2619-account-create-update-9csh2\" (UID: \"fe6e7565-793b-4555-b6d5-758f87a5b9c7\") " pod="openstack/heat-2619-account-create-update-9csh2" Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.025826 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-3c9f-account-create-update-rtprr"] Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.033340 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q7phq\" (UniqueName: \"kubernetes.io/projected/8600373d-f466-43d4-92fc-1fa938f6e91b-kube-api-access-q7phq\") pod \"heat-db-create-kk8kb\" (UID: \"8600373d-f466-43d4-92fc-1fa938f6e91b\") " pod="openstack/heat-db-create-kk8kb" Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.035068 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.056496 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1c75f0f9-53e1-4a64-9211-559ef51c1b53" path="/var/lib/kubelet/pods/1c75f0f9-53e1-4a64-9211-559ef51c1b53/volumes" Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.092231 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-ngtnj"] Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.093373 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-ngtnj" Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.097834 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-ngtnj"] Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.111860 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5c0ddc66-7696-4816-8e96-340aeb39d242-operator-scripts\") pod \"barbican-3c9f-account-create-update-rtprr\" (UID: \"5c0ddc66-7696-4816-8e96-340aeb39d242\") " pod="openstack/barbican-3c9f-account-create-update-rtprr" Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.112027 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lbmg7\" (UniqueName: \"kubernetes.io/projected/5c0ddc66-7696-4816-8e96-340aeb39d242-kube-api-access-lbmg7\") pod \"barbican-3c9f-account-create-update-rtprr\" (UID: \"5c0ddc66-7696-4816-8e96-340aeb39d242\") " pod="openstack/barbican-3c9f-account-create-update-rtprr" Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.112127 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6876j\" (UniqueName: \"kubernetes.io/projected/7b747df8-e61f-4689-a08c-0dc6688f1a0e-kube-api-access-6876j\") pod \"cinder-db-create-twgjw\" (UID: \"7b747df8-e61f-4689-a08c-0dc6688f1a0e\") " pod="openstack/cinder-db-create-twgjw" Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.112162 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5240d35b-cbf1-472d-91da-debf418dd208-combined-ca-bundle\") pod \"keystone-db-sync-84f4b\" (UID: \"5240d35b-cbf1-472d-91da-debf418dd208\") " pod="openstack/keystone-db-sync-84f4b" Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.112217 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7b747df8-e61f-4689-a08c-0dc6688f1a0e-operator-scripts\") pod \"cinder-db-create-twgjw\" (UID: \"7b747df8-e61f-4689-a08c-0dc6688f1a0e\") " pod="openstack/cinder-db-create-twgjw" Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.112307 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f4nqn\" (UniqueName: \"kubernetes.io/projected/5240d35b-cbf1-472d-91da-debf418dd208-kube-api-access-f4nqn\") pod \"keystone-db-sync-84f4b\" (UID: \"5240d35b-cbf1-472d-91da-debf418dd208\") " pod="openstack/keystone-db-sync-84f4b" Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.112336 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5240d35b-cbf1-472d-91da-debf418dd208-config-data\") pod \"keystone-db-sync-84f4b\" (UID: \"5240d35b-cbf1-472d-91da-debf418dd208\") " pod="openstack/keystone-db-sync-84f4b" Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.117272 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5240d35b-cbf1-472d-91da-debf418dd208-combined-ca-bundle\") pod \"keystone-db-sync-84f4b\" (UID: \"5240d35b-cbf1-472d-91da-debf418dd208\") " pod="openstack/keystone-db-sync-84f4b" Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.117554 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5240d35b-cbf1-472d-91da-debf418dd208-config-data\") pod \"keystone-db-sync-84f4b\" (UID: \"5240d35b-cbf1-472d-91da-debf418dd208\") " pod="openstack/keystone-db-sync-84f4b" Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.126269 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f4nqn\" (UniqueName: \"kubernetes.io/projected/5240d35b-cbf1-472d-91da-debf418dd208-kube-api-access-f4nqn\") pod \"keystone-db-sync-84f4b\" (UID: \"5240d35b-cbf1-472d-91da-debf418dd208\") " pod="openstack/keystone-db-sync-84f4b" Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.134486 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-kk8kb" Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.140859 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-2619-account-create-update-9csh2" Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.199732 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-60f6-account-create-update-pjh6n"] Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.200767 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-60f6-account-create-update-pjh6n" Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.202913 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.207060 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-nskxc"] Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.211198 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-nskxc" Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.215725 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-60f6-account-create-update-pjh6n"] Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.216617 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6876j\" (UniqueName: \"kubernetes.io/projected/7b747df8-e61f-4689-a08c-0dc6688f1a0e-kube-api-access-6876j\") pod \"cinder-db-create-twgjw\" (UID: \"7b747df8-e61f-4689-a08c-0dc6688f1a0e\") " pod="openstack/cinder-db-create-twgjw" Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.216686 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5nfhq\" (UniqueName: \"kubernetes.io/projected/9be37b35-7ad5-434a-8d16-c2f9f1661821-kube-api-access-5nfhq\") pod \"barbican-db-create-ngtnj\" (UID: \"9be37b35-7ad5-434a-8d16-c2f9f1661821\") " pod="openstack/barbican-db-create-ngtnj" Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.216722 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7b747df8-e61f-4689-a08c-0dc6688f1a0e-operator-scripts\") pod \"cinder-db-create-twgjw\" (UID: \"7b747df8-e61f-4689-a08c-0dc6688f1a0e\") " pod="openstack/cinder-db-create-twgjw" Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.216752 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9be37b35-7ad5-434a-8d16-c2f9f1661821-operator-scripts\") pod \"barbican-db-create-ngtnj\" (UID: \"9be37b35-7ad5-434a-8d16-c2f9f1661821\") " pod="openstack/barbican-db-create-ngtnj" Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.216803 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5c0ddc66-7696-4816-8e96-340aeb39d242-operator-scripts\") pod \"barbican-3c9f-account-create-update-rtprr\" (UID: \"5c0ddc66-7696-4816-8e96-340aeb39d242\") " pod="openstack/barbican-3c9f-account-create-update-rtprr" Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.216857 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lbmg7\" (UniqueName: \"kubernetes.io/projected/5c0ddc66-7696-4816-8e96-340aeb39d242-kube-api-access-lbmg7\") pod \"barbican-3c9f-account-create-update-rtprr\" (UID: \"5c0ddc66-7696-4816-8e96-340aeb39d242\") " pod="openstack/barbican-3c9f-account-create-update-rtprr" Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.217832 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7b747df8-e61f-4689-a08c-0dc6688f1a0e-operator-scripts\") pod \"cinder-db-create-twgjw\" (UID: \"7b747df8-e61f-4689-a08c-0dc6688f1a0e\") " pod="openstack/cinder-db-create-twgjw" Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.219408 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5c0ddc66-7696-4816-8e96-340aeb39d242-operator-scripts\") pod \"barbican-3c9f-account-create-update-rtprr\" (UID: \"5c0ddc66-7696-4816-8e96-340aeb39d242\") " pod="openstack/barbican-3c9f-account-create-update-rtprr" Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.245022 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6876j\" (UniqueName: \"kubernetes.io/projected/7b747df8-e61f-4689-a08c-0dc6688f1a0e-kube-api-access-6876j\") pod \"cinder-db-create-twgjw\" (UID: \"7b747df8-e61f-4689-a08c-0dc6688f1a0e\") " pod="openstack/cinder-db-create-twgjw" Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.246442 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lbmg7\" (UniqueName: \"kubernetes.io/projected/5c0ddc66-7696-4816-8e96-340aeb39d242-kube-api-access-lbmg7\") pod \"barbican-3c9f-account-create-update-rtprr\" (UID: \"5c0ddc66-7696-4816-8e96-340aeb39d242\") " pod="openstack/barbican-3c9f-account-create-update-rtprr" Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.260502 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-nskxc"] Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.266240 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-84f4b" Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.314222 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-0662-account-create-update-j7bhp"] Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.315444 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-0662-account-create-update-j7bhp" Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.317209 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.317732 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9be37b35-7ad5-434a-8d16-c2f9f1661821-operator-scripts\") pod \"barbican-db-create-ngtnj\" (UID: \"9be37b35-7ad5-434a-8d16-c2f9f1661821\") " pod="openstack/barbican-db-create-ngtnj" Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.317831 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pq7qd\" (UniqueName: \"kubernetes.io/projected/93c8ce3d-a4f0-4cc7-8592-61653eb09d3c-kube-api-access-pq7qd\") pod \"cinder-60f6-account-create-update-pjh6n\" (UID: \"93c8ce3d-a4f0-4cc7-8592-61653eb09d3c\") " pod="openstack/cinder-60f6-account-create-update-pjh6n" Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.317861 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pqhbb\" (UniqueName: \"kubernetes.io/projected/a10a8f17-c27b-4544-8f9d-a6afd9991a8e-kube-api-access-pqhbb\") pod \"neutron-db-create-nskxc\" (UID: \"a10a8f17-c27b-4544-8f9d-a6afd9991a8e\") " pod="openstack/neutron-db-create-nskxc" Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.317986 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/93c8ce3d-a4f0-4cc7-8592-61653eb09d3c-operator-scripts\") pod \"cinder-60f6-account-create-update-pjh6n\" (UID: \"93c8ce3d-a4f0-4cc7-8592-61653eb09d3c\") " pod="openstack/cinder-60f6-account-create-update-pjh6n" Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.318021 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a10a8f17-c27b-4544-8f9d-a6afd9991a8e-operator-scripts\") pod \"neutron-db-create-nskxc\" (UID: \"a10a8f17-c27b-4544-8f9d-a6afd9991a8e\") " pod="openstack/neutron-db-create-nskxc" Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.318050 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5nfhq\" (UniqueName: \"kubernetes.io/projected/9be37b35-7ad5-434a-8d16-c2f9f1661821-kube-api-access-5nfhq\") pod \"barbican-db-create-ngtnj\" (UID: \"9be37b35-7ad5-434a-8d16-c2f9f1661821\") " pod="openstack/barbican-db-create-ngtnj" Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.325057 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9be37b35-7ad5-434a-8d16-c2f9f1661821-operator-scripts\") pod \"barbican-db-create-ngtnj\" (UID: \"9be37b35-7ad5-434a-8d16-c2f9f1661821\") " pod="openstack/barbican-db-create-ngtnj" Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.329633 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-0662-account-create-update-j7bhp"] Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.346690 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5nfhq\" (UniqueName: \"kubernetes.io/projected/9be37b35-7ad5-434a-8d16-c2f9f1661821-kube-api-access-5nfhq\") pod \"barbican-db-create-ngtnj\" (UID: \"9be37b35-7ad5-434a-8d16-c2f9f1661821\") " pod="openstack/barbican-db-create-ngtnj" Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.380341 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-twgjw" Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.385489 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-3c9f-account-create-update-rtprr" Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.412833 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-ngtnj" Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.420620 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wvj2q\" (UniqueName: \"kubernetes.io/projected/b4773e7a-a50c-42e6-bb27-a25b0055b8e3-kube-api-access-wvj2q\") pod \"neutron-0662-account-create-update-j7bhp\" (UID: \"b4773e7a-a50c-42e6-bb27-a25b0055b8e3\") " pod="openstack/neutron-0662-account-create-update-j7bhp" Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.420717 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pq7qd\" (UniqueName: \"kubernetes.io/projected/93c8ce3d-a4f0-4cc7-8592-61653eb09d3c-kube-api-access-pq7qd\") pod \"cinder-60f6-account-create-update-pjh6n\" (UID: \"93c8ce3d-a4f0-4cc7-8592-61653eb09d3c\") " pod="openstack/cinder-60f6-account-create-update-pjh6n" Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.420749 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pqhbb\" (UniqueName: \"kubernetes.io/projected/a10a8f17-c27b-4544-8f9d-a6afd9991a8e-kube-api-access-pqhbb\") pod \"neutron-db-create-nskxc\" (UID: \"a10a8f17-c27b-4544-8f9d-a6afd9991a8e\") " pod="openstack/neutron-db-create-nskxc" Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.420773 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b4773e7a-a50c-42e6-bb27-a25b0055b8e3-operator-scripts\") pod \"neutron-0662-account-create-update-j7bhp\" (UID: \"b4773e7a-a50c-42e6-bb27-a25b0055b8e3\") " pod="openstack/neutron-0662-account-create-update-j7bhp" Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.420876 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/93c8ce3d-a4f0-4cc7-8592-61653eb09d3c-operator-scripts\") pod \"cinder-60f6-account-create-update-pjh6n\" (UID: \"93c8ce3d-a4f0-4cc7-8592-61653eb09d3c\") " pod="openstack/cinder-60f6-account-create-update-pjh6n" Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.420918 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a10a8f17-c27b-4544-8f9d-a6afd9991a8e-operator-scripts\") pod \"neutron-db-create-nskxc\" (UID: \"a10a8f17-c27b-4544-8f9d-a6afd9991a8e\") " pod="openstack/neutron-db-create-nskxc" Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.421488 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/93c8ce3d-a4f0-4cc7-8592-61653eb09d3c-operator-scripts\") pod \"cinder-60f6-account-create-update-pjh6n\" (UID: \"93c8ce3d-a4f0-4cc7-8592-61653eb09d3c\") " pod="openstack/cinder-60f6-account-create-update-pjh6n" Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.421496 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a10a8f17-c27b-4544-8f9d-a6afd9991a8e-operator-scripts\") pod \"neutron-db-create-nskxc\" (UID: \"a10a8f17-c27b-4544-8f9d-a6afd9991a8e\") " pod="openstack/neutron-db-create-nskxc" Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.435515 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pqhbb\" (UniqueName: \"kubernetes.io/projected/a10a8f17-c27b-4544-8f9d-a6afd9991a8e-kube-api-access-pqhbb\") pod \"neutron-db-create-nskxc\" (UID: \"a10a8f17-c27b-4544-8f9d-a6afd9991a8e\") " pod="openstack/neutron-db-create-nskxc" Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.436950 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pq7qd\" (UniqueName: \"kubernetes.io/projected/93c8ce3d-a4f0-4cc7-8592-61653eb09d3c-kube-api-access-pq7qd\") pod \"cinder-60f6-account-create-update-pjh6n\" (UID: \"93c8ce3d-a4f0-4cc7-8592-61653eb09d3c\") " pod="openstack/cinder-60f6-account-create-update-pjh6n" Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.516744 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-60f6-account-create-update-pjh6n" Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.521865 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wvj2q\" (UniqueName: \"kubernetes.io/projected/b4773e7a-a50c-42e6-bb27-a25b0055b8e3-kube-api-access-wvj2q\") pod \"neutron-0662-account-create-update-j7bhp\" (UID: \"b4773e7a-a50c-42e6-bb27-a25b0055b8e3\") " pod="openstack/neutron-0662-account-create-update-j7bhp" Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.521980 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b4773e7a-a50c-42e6-bb27-a25b0055b8e3-operator-scripts\") pod \"neutron-0662-account-create-update-j7bhp\" (UID: \"b4773e7a-a50c-42e6-bb27-a25b0055b8e3\") " pod="openstack/neutron-0662-account-create-update-j7bhp" Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.522588 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b4773e7a-a50c-42e6-bb27-a25b0055b8e3-operator-scripts\") pod \"neutron-0662-account-create-update-j7bhp\" (UID: \"b4773e7a-a50c-42e6-bb27-a25b0055b8e3\") " pod="openstack/neutron-0662-account-create-update-j7bhp" Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.534472 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wvj2q\" (UniqueName: \"kubernetes.io/projected/b4773e7a-a50c-42e6-bb27-a25b0055b8e3-kube-api-access-wvj2q\") pod \"neutron-0662-account-create-update-j7bhp\" (UID: \"b4773e7a-a50c-42e6-bb27-a25b0055b8e3\") " pod="openstack/neutron-0662-account-create-update-j7bhp" Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.539459 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-nskxc" Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.601502 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-create-kk8kb"] Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.637035 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-0662-account-create-update-j7bhp" Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.687867 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-2619-account-create-update-9csh2"] Jan 21 11:26:12 crc kubenswrapper[4824]: I0121 11:26:12.982480 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-84f4b"] Jan 21 11:26:12 crc kubenswrapper[4824]: W0121 11:26:12.986526 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5240d35b_cbf1_472d_91da_debf418dd208.slice/crio-9132d198700ff29f709588ab7e526f22887b3f414c249d3511b9eb2a27c47581 WatchSource:0}: Error finding container 9132d198700ff29f709588ab7e526f22887b3f414c249d3511b9eb2a27c47581: Status 404 returned error can't find the container with id 9132d198700ff29f709588ab7e526f22887b3f414c249d3511b9eb2a27c47581 Jan 21 11:26:13 crc kubenswrapper[4824]: E0121 11:26:13.005760 4824 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 192.168.26.120:34444->192.168.26.120:32865: write tcp 192.168.26.120:34444->192.168.26.120:32865: write: broken pipe Jan 21 11:26:13 crc kubenswrapper[4824]: I0121 11:26:13.135203 4824 generic.go:334] "Generic (PLEG): container finished" podID="8600373d-f466-43d4-92fc-1fa938f6e91b" containerID="81f52e55093614da3aed12f82750111917b47ee8401eb58ccc0da58488e3019f" exitCode=0 Jan 21 11:26:13 crc kubenswrapper[4824]: I0121 11:26:13.135419 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-kk8kb" event={"ID":"8600373d-f466-43d4-92fc-1fa938f6e91b","Type":"ContainerDied","Data":"81f52e55093614da3aed12f82750111917b47ee8401eb58ccc0da58488e3019f"} Jan 21 11:26:13 crc kubenswrapper[4824]: I0121 11:26:13.135443 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-kk8kb" event={"ID":"8600373d-f466-43d4-92fc-1fa938f6e91b","Type":"ContainerStarted","Data":"e76ccb54eba21f586b5ec6a78b2a5550037daa9208d52611b08df26c0e3ac56b"} Jan 21 11:26:13 crc kubenswrapper[4824]: I0121 11:26:13.135601 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-nskxc"] Jan 21 11:26:13 crc kubenswrapper[4824]: I0121 11:26:13.142010 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-3c9f-account-create-update-rtprr"] Jan 21 11:26:13 crc kubenswrapper[4824]: I0121 11:26:13.143146 4824 generic.go:334] "Generic (PLEG): container finished" podID="fe6e7565-793b-4555-b6d5-758f87a5b9c7" containerID="f658b544252ce5134cf2b3d03b008e5a8cf0c0a33b02b6a5554a5288bb79bbfe" exitCode=0 Jan 21 11:26:13 crc kubenswrapper[4824]: I0121 11:26:13.143203 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-2619-account-create-update-9csh2" event={"ID":"fe6e7565-793b-4555-b6d5-758f87a5b9c7","Type":"ContainerDied","Data":"f658b544252ce5134cf2b3d03b008e5a8cf0c0a33b02b6a5554a5288bb79bbfe"} Jan 21 11:26:13 crc kubenswrapper[4824]: I0121 11:26:13.143225 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-2619-account-create-update-9csh2" event={"ID":"fe6e7565-793b-4555-b6d5-758f87a5b9c7","Type":"ContainerStarted","Data":"8526f4c3098ff09c46e3c8957568e817f203fa1293f2cd973998f5c043bb31d2"} Jan 21 11:26:13 crc kubenswrapper[4824]: I0121 11:26:13.145689 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-84f4b" event={"ID":"5240d35b-cbf1-472d-91da-debf418dd208","Type":"ContainerStarted","Data":"9132d198700ff29f709588ab7e526f22887b3f414c249d3511b9eb2a27c47581"} Jan 21 11:26:13 crc kubenswrapper[4824]: I0121 11:26:13.146934 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-60f6-account-create-update-pjh6n"] Jan 21 11:26:13 crc kubenswrapper[4824]: I0121 11:26:13.222149 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-twgjw"] Jan 21 11:26:13 crc kubenswrapper[4824]: I0121 11:26:13.333620 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-ngtnj"] Jan 21 11:26:13 crc kubenswrapper[4824]: W0121 11:26:13.339827 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9be37b35_7ad5_434a_8d16_c2f9f1661821.slice/crio-e6fc3836b7afabeab4a7371d16c4b247a2670b8dfe835939c4f3868093185822 WatchSource:0}: Error finding container e6fc3836b7afabeab4a7371d16c4b247a2670b8dfe835939c4f3868093185822: Status 404 returned error can't find the container with id e6fc3836b7afabeab4a7371d16c4b247a2670b8dfe835939c4f3868093185822 Jan 21 11:26:13 crc kubenswrapper[4824]: I0121 11:26:13.350501 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-0662-account-create-update-j7bhp"] Jan 21 11:26:13 crc kubenswrapper[4824]: W0121 11:26:13.384667 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb4773e7a_a50c_42e6_bb27_a25b0055b8e3.slice/crio-6d872ef3f227d5894d64148f328773527cef85531ed41d2b5f9fdbbdf412e8bf WatchSource:0}: Error finding container 6d872ef3f227d5894d64148f328773527cef85531ed41d2b5f9fdbbdf412e8bf: Status 404 returned error can't find the container with id 6d872ef3f227d5894d64148f328773527cef85531ed41d2b5f9fdbbdf412e8bf Jan 21 11:26:14 crc kubenswrapper[4824]: I0121 11:26:14.167476 4824 generic.go:334] "Generic (PLEG): container finished" podID="b4773e7a-a50c-42e6-bb27-a25b0055b8e3" containerID="cc788777a506818eef849722d9795fcf44c54eb36cb8cfbc69470609fce80bd0" exitCode=0 Jan 21 11:26:14 crc kubenswrapper[4824]: I0121 11:26:14.167725 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-0662-account-create-update-j7bhp" event={"ID":"b4773e7a-a50c-42e6-bb27-a25b0055b8e3","Type":"ContainerDied","Data":"cc788777a506818eef849722d9795fcf44c54eb36cb8cfbc69470609fce80bd0"} Jan 21 11:26:14 crc kubenswrapper[4824]: I0121 11:26:14.167748 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-0662-account-create-update-j7bhp" event={"ID":"b4773e7a-a50c-42e6-bb27-a25b0055b8e3","Type":"ContainerStarted","Data":"6d872ef3f227d5894d64148f328773527cef85531ed41d2b5f9fdbbdf412e8bf"} Jan 21 11:26:14 crc kubenswrapper[4824]: I0121 11:26:14.171745 4824 generic.go:334] "Generic (PLEG): container finished" podID="a10a8f17-c27b-4544-8f9d-a6afd9991a8e" containerID="b6f52bcc83bf92c2ef07afe3fa11fe445d957e311fc35ba4f1767a43394658f9" exitCode=0 Jan 21 11:26:14 crc kubenswrapper[4824]: I0121 11:26:14.171807 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-nskxc" event={"ID":"a10a8f17-c27b-4544-8f9d-a6afd9991a8e","Type":"ContainerDied","Data":"b6f52bcc83bf92c2ef07afe3fa11fe445d957e311fc35ba4f1767a43394658f9"} Jan 21 11:26:14 crc kubenswrapper[4824]: I0121 11:26:14.171831 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-nskxc" event={"ID":"a10a8f17-c27b-4544-8f9d-a6afd9991a8e","Type":"ContainerStarted","Data":"d2e64d054f528de025088752cbba2a64296b9807a6337f5fdd775173aefc9828"} Jan 21 11:26:14 crc kubenswrapper[4824]: I0121 11:26:14.178982 4824 generic.go:334] "Generic (PLEG): container finished" podID="93c8ce3d-a4f0-4cc7-8592-61653eb09d3c" containerID="45d2f067a75588f2f74c579f26d8f9af2b6a872de4eb9d340b34e9fee4726547" exitCode=0 Jan 21 11:26:14 crc kubenswrapper[4824]: I0121 11:26:14.179101 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-60f6-account-create-update-pjh6n" event={"ID":"93c8ce3d-a4f0-4cc7-8592-61653eb09d3c","Type":"ContainerDied","Data":"45d2f067a75588f2f74c579f26d8f9af2b6a872de4eb9d340b34e9fee4726547"} Jan 21 11:26:14 crc kubenswrapper[4824]: I0121 11:26:14.179163 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-60f6-account-create-update-pjh6n" event={"ID":"93c8ce3d-a4f0-4cc7-8592-61653eb09d3c","Type":"ContainerStarted","Data":"a60e0d515d078f64392d5d1ca5e1034d522ce6ad7bf98f9a82fa5e036ed775d2"} Jan 21 11:26:14 crc kubenswrapper[4824]: I0121 11:26:14.184371 4824 generic.go:334] "Generic (PLEG): container finished" podID="5c0ddc66-7696-4816-8e96-340aeb39d242" containerID="2176bc084f2009e4d942790b8c3474f286b639f648b347b28da26c0a5a3ccec2" exitCode=0 Jan 21 11:26:14 crc kubenswrapper[4824]: I0121 11:26:14.184430 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-3c9f-account-create-update-rtprr" event={"ID":"5c0ddc66-7696-4816-8e96-340aeb39d242","Type":"ContainerDied","Data":"2176bc084f2009e4d942790b8c3474f286b639f648b347b28da26c0a5a3ccec2"} Jan 21 11:26:14 crc kubenswrapper[4824]: I0121 11:26:14.184453 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-3c9f-account-create-update-rtprr" event={"ID":"5c0ddc66-7696-4816-8e96-340aeb39d242","Type":"ContainerStarted","Data":"e9e93c5745e7fe74ef1e643164e9a0a4e165267deeb109faf4df81c36f336691"} Jan 21 11:26:14 crc kubenswrapper[4824]: I0121 11:26:14.194506 4824 generic.go:334] "Generic (PLEG): container finished" podID="9be37b35-7ad5-434a-8d16-c2f9f1661821" containerID="c1a92fb4372889a2ab5383adea30e43a1fa74caae3039082b2ef5482212d2758" exitCode=0 Jan 21 11:26:14 crc kubenswrapper[4824]: I0121 11:26:14.194557 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-ngtnj" event={"ID":"9be37b35-7ad5-434a-8d16-c2f9f1661821","Type":"ContainerDied","Data":"c1a92fb4372889a2ab5383adea30e43a1fa74caae3039082b2ef5482212d2758"} Jan 21 11:26:14 crc kubenswrapper[4824]: I0121 11:26:14.194575 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-ngtnj" event={"ID":"9be37b35-7ad5-434a-8d16-c2f9f1661821","Type":"ContainerStarted","Data":"e6fc3836b7afabeab4a7371d16c4b247a2670b8dfe835939c4f3868093185822"} Jan 21 11:26:14 crc kubenswrapper[4824]: I0121 11:26:14.209338 4824 generic.go:334] "Generic (PLEG): container finished" podID="7b747df8-e61f-4689-a08c-0dc6688f1a0e" containerID="e12a5e35428ef122a88a06ec7747afe8e712c9a7807b731859f6ec595579f719" exitCode=0 Jan 21 11:26:14 crc kubenswrapper[4824]: I0121 11:26:14.209393 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-twgjw" event={"ID":"7b747df8-e61f-4689-a08c-0dc6688f1a0e","Type":"ContainerDied","Data":"e12a5e35428ef122a88a06ec7747afe8e712c9a7807b731859f6ec595579f719"} Jan 21 11:26:14 crc kubenswrapper[4824]: I0121 11:26:14.209425 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-twgjw" event={"ID":"7b747df8-e61f-4689-a08c-0dc6688f1a0e","Type":"ContainerStarted","Data":"5cbfd6e560404567a0bbc9ae88f4245a4c606e0f65ba74f1fdb8d8f25d487579"} Jan 21 11:26:14 crc kubenswrapper[4824]: I0121 11:26:14.665751 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-kk8kb" Jan 21 11:26:14 crc kubenswrapper[4824]: I0121 11:26:14.737998 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-2619-account-create-update-9csh2" Jan 21 11:26:14 crc kubenswrapper[4824]: I0121 11:26:14.753843 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8600373d-f466-43d4-92fc-1fa938f6e91b-operator-scripts\") pod \"8600373d-f466-43d4-92fc-1fa938f6e91b\" (UID: \"8600373d-f466-43d4-92fc-1fa938f6e91b\") " Jan 21 11:26:14 crc kubenswrapper[4824]: I0121 11:26:14.753986 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q7phq\" (UniqueName: \"kubernetes.io/projected/8600373d-f466-43d4-92fc-1fa938f6e91b-kube-api-access-q7phq\") pod \"8600373d-f466-43d4-92fc-1fa938f6e91b\" (UID: \"8600373d-f466-43d4-92fc-1fa938f6e91b\") " Jan 21 11:26:14 crc kubenswrapper[4824]: I0121 11:26:14.754501 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8600373d-f466-43d4-92fc-1fa938f6e91b-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "8600373d-f466-43d4-92fc-1fa938f6e91b" (UID: "8600373d-f466-43d4-92fc-1fa938f6e91b"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:26:14 crc kubenswrapper[4824]: I0121 11:26:14.758463 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8600373d-f466-43d4-92fc-1fa938f6e91b-kube-api-access-q7phq" (OuterVolumeSpecName: "kube-api-access-q7phq") pod "8600373d-f466-43d4-92fc-1fa938f6e91b" (UID: "8600373d-f466-43d4-92fc-1fa938f6e91b"). InnerVolumeSpecName "kube-api-access-q7phq". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:26:14 crc kubenswrapper[4824]: I0121 11:26:14.855124 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4p9x\" (UniqueName: \"kubernetes.io/projected/fe6e7565-793b-4555-b6d5-758f87a5b9c7-kube-api-access-w4p9x\") pod \"fe6e7565-793b-4555-b6d5-758f87a5b9c7\" (UID: \"fe6e7565-793b-4555-b6d5-758f87a5b9c7\") " Jan 21 11:26:14 crc kubenswrapper[4824]: I0121 11:26:14.855176 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fe6e7565-793b-4555-b6d5-758f87a5b9c7-operator-scripts\") pod \"fe6e7565-793b-4555-b6d5-758f87a5b9c7\" (UID: \"fe6e7565-793b-4555-b6d5-758f87a5b9c7\") " Jan 21 11:26:14 crc kubenswrapper[4824]: I0121 11:26:14.855672 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fe6e7565-793b-4555-b6d5-758f87a5b9c7-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "fe6e7565-793b-4555-b6d5-758f87a5b9c7" (UID: "fe6e7565-793b-4555-b6d5-758f87a5b9c7"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:26:14 crc kubenswrapper[4824]: I0121 11:26:14.856190 4824 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fe6e7565-793b-4555-b6d5-758f87a5b9c7-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:14 crc kubenswrapper[4824]: I0121 11:26:14.856217 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q7phq\" (UniqueName: \"kubernetes.io/projected/8600373d-f466-43d4-92fc-1fa938f6e91b-kube-api-access-q7phq\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:14 crc kubenswrapper[4824]: I0121 11:26:14.856228 4824 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8600373d-f466-43d4-92fc-1fa938f6e91b-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:14 crc kubenswrapper[4824]: I0121 11:26:14.857655 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fe6e7565-793b-4555-b6d5-758f87a5b9c7-kube-api-access-w4p9x" (OuterVolumeSpecName: "kube-api-access-w4p9x") pod "fe6e7565-793b-4555-b6d5-758f87a5b9c7" (UID: "fe6e7565-793b-4555-b6d5-758f87a5b9c7"). InnerVolumeSpecName "kube-api-access-w4p9x". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:26:14 crc kubenswrapper[4824]: I0121 11:26:14.959380 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4p9x\" (UniqueName: \"kubernetes.io/projected/fe6e7565-793b-4555-b6d5-758f87a5b9c7-kube-api-access-w4p9x\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:15 crc kubenswrapper[4824]: I0121 11:26:15.218612 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-kk8kb" Jan 21 11:26:15 crc kubenswrapper[4824]: I0121 11:26:15.218618 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-kk8kb" event={"ID":"8600373d-f466-43d4-92fc-1fa938f6e91b","Type":"ContainerDied","Data":"e76ccb54eba21f586b5ec6a78b2a5550037daa9208d52611b08df26c0e3ac56b"} Jan 21 11:26:15 crc kubenswrapper[4824]: I0121 11:26:15.218714 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e76ccb54eba21f586b5ec6a78b2a5550037daa9208d52611b08df26c0e3ac56b" Jan 21 11:26:15 crc kubenswrapper[4824]: I0121 11:26:15.220180 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-2619-account-create-update-9csh2" event={"ID":"fe6e7565-793b-4555-b6d5-758f87a5b9c7","Type":"ContainerDied","Data":"8526f4c3098ff09c46e3c8957568e817f203fa1293f2cd973998f5c043bb31d2"} Jan 21 11:26:15 crc kubenswrapper[4824]: I0121 11:26:15.220212 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8526f4c3098ff09c46e3c8957568e817f203fa1293f2cd973998f5c043bb31d2" Jan 21 11:26:15 crc kubenswrapper[4824]: I0121 11:26:15.220375 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-2619-account-create-update-9csh2" Jan 21 11:26:15 crc kubenswrapper[4824]: I0121 11:26:15.720140 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-74f6bcbc87-sj79w" Jan 21 11:26:15 crc kubenswrapper[4824]: I0121 11:26:15.759419 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-698758b865-qzfrf"] Jan 21 11:26:15 crc kubenswrapper[4824]: I0121 11:26:15.759750 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-698758b865-qzfrf" podUID="e94778b1-d36a-47e9-b201-9b7cd1cdb640" containerName="dnsmasq-dns" containerID="cri-o://8a0b04dc02e5d8c871ec744b980efd9b4e195f4024e4ef8187ce38c2f8284265" gracePeriod=10 Jan 21 11:26:16 crc kubenswrapper[4824]: I0121 11:26:16.073664 4824 patch_prober.go:28] interesting pod/machine-config-daemon-zqjg9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 21 11:26:16 crc kubenswrapper[4824]: I0121 11:26:16.073889 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 21 11:26:16 crc kubenswrapper[4824]: I0121 11:26:16.227770 4824 generic.go:334] "Generic (PLEG): container finished" podID="e94778b1-d36a-47e9-b201-9b7cd1cdb640" containerID="8a0b04dc02e5d8c871ec744b980efd9b4e195f4024e4ef8187ce38c2f8284265" exitCode=0 Jan 21 11:26:16 crc kubenswrapper[4824]: I0121 11:26:16.227812 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-qzfrf" event={"ID":"e94778b1-d36a-47e9-b201-9b7cd1cdb640","Type":"ContainerDied","Data":"8a0b04dc02e5d8c871ec744b980efd9b4e195f4024e4ef8187ce38c2f8284265"} Jan 21 11:26:16 crc kubenswrapper[4824]: I0121 11:26:16.695786 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-698758b865-qzfrf" podUID="e94778b1-d36a-47e9-b201-9b7cd1cdb640" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.112:5353: connect: connection refused" Jan 21 11:26:17 crc kubenswrapper[4824]: I0121 11:26:17.404801 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-nskxc" Jan 21 11:26:17 crc kubenswrapper[4824]: I0121 11:26:17.420303 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-0662-account-create-update-j7bhp" Jan 21 11:26:17 crc kubenswrapper[4824]: I0121 11:26:17.423307 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-60f6-account-create-update-pjh6n" Jan 21 11:26:17 crc kubenswrapper[4824]: I0121 11:26:17.442194 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-3c9f-account-create-update-rtprr" Jan 21 11:26:17 crc kubenswrapper[4824]: I0121 11:26:17.449329 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-ngtnj" Jan 21 11:26:17 crc kubenswrapper[4824]: I0121 11:26:17.450146 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-twgjw" Jan 21 11:26:17 crc kubenswrapper[4824]: I0121 11:26:17.504642 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6876j\" (UniqueName: \"kubernetes.io/projected/7b747df8-e61f-4689-a08c-0dc6688f1a0e-kube-api-access-6876j\") pod \"7b747df8-e61f-4689-a08c-0dc6688f1a0e\" (UID: \"7b747df8-e61f-4689-a08c-0dc6688f1a0e\") " Jan 21 11:26:17 crc kubenswrapper[4824]: I0121 11:26:17.504675 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pq7qd\" (UniqueName: \"kubernetes.io/projected/93c8ce3d-a4f0-4cc7-8592-61653eb09d3c-kube-api-access-pq7qd\") pod \"93c8ce3d-a4f0-4cc7-8592-61653eb09d3c\" (UID: \"93c8ce3d-a4f0-4cc7-8592-61653eb09d3c\") " Jan 21 11:26:17 crc kubenswrapper[4824]: I0121 11:26:17.504716 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a10a8f17-c27b-4544-8f9d-a6afd9991a8e-operator-scripts\") pod \"a10a8f17-c27b-4544-8f9d-a6afd9991a8e\" (UID: \"a10a8f17-c27b-4544-8f9d-a6afd9991a8e\") " Jan 21 11:26:17 crc kubenswrapper[4824]: I0121 11:26:17.504745 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pqhbb\" (UniqueName: \"kubernetes.io/projected/a10a8f17-c27b-4544-8f9d-a6afd9991a8e-kube-api-access-pqhbb\") pod \"a10a8f17-c27b-4544-8f9d-a6afd9991a8e\" (UID: \"a10a8f17-c27b-4544-8f9d-a6afd9991a8e\") " Jan 21 11:26:17 crc kubenswrapper[4824]: I0121 11:26:17.504795 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b4773e7a-a50c-42e6-bb27-a25b0055b8e3-operator-scripts\") pod \"b4773e7a-a50c-42e6-bb27-a25b0055b8e3\" (UID: \"b4773e7a-a50c-42e6-bb27-a25b0055b8e3\") " Jan 21 11:26:17 crc kubenswrapper[4824]: I0121 11:26:17.504818 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5c0ddc66-7696-4816-8e96-340aeb39d242-operator-scripts\") pod \"5c0ddc66-7696-4816-8e96-340aeb39d242\" (UID: \"5c0ddc66-7696-4816-8e96-340aeb39d242\") " Jan 21 11:26:17 crc kubenswrapper[4824]: I0121 11:26:17.504844 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5nfhq\" (UniqueName: \"kubernetes.io/projected/9be37b35-7ad5-434a-8d16-c2f9f1661821-kube-api-access-5nfhq\") pod \"9be37b35-7ad5-434a-8d16-c2f9f1661821\" (UID: \"9be37b35-7ad5-434a-8d16-c2f9f1661821\") " Jan 21 11:26:17 crc kubenswrapper[4824]: I0121 11:26:17.504875 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7b747df8-e61f-4689-a08c-0dc6688f1a0e-operator-scripts\") pod \"7b747df8-e61f-4689-a08c-0dc6688f1a0e\" (UID: \"7b747df8-e61f-4689-a08c-0dc6688f1a0e\") " Jan 21 11:26:17 crc kubenswrapper[4824]: I0121 11:26:17.504890 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/93c8ce3d-a4f0-4cc7-8592-61653eb09d3c-operator-scripts\") pod \"93c8ce3d-a4f0-4cc7-8592-61653eb09d3c\" (UID: \"93c8ce3d-a4f0-4cc7-8592-61653eb09d3c\") " Jan 21 11:26:17 crc kubenswrapper[4824]: I0121 11:26:17.504911 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lbmg7\" (UniqueName: \"kubernetes.io/projected/5c0ddc66-7696-4816-8e96-340aeb39d242-kube-api-access-lbmg7\") pod \"5c0ddc66-7696-4816-8e96-340aeb39d242\" (UID: \"5c0ddc66-7696-4816-8e96-340aeb39d242\") " Jan 21 11:26:17 crc kubenswrapper[4824]: I0121 11:26:17.504942 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9be37b35-7ad5-434a-8d16-c2f9f1661821-operator-scripts\") pod \"9be37b35-7ad5-434a-8d16-c2f9f1661821\" (UID: \"9be37b35-7ad5-434a-8d16-c2f9f1661821\") " Jan 21 11:26:17 crc kubenswrapper[4824]: I0121 11:26:17.505007 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wvj2q\" (UniqueName: \"kubernetes.io/projected/b4773e7a-a50c-42e6-bb27-a25b0055b8e3-kube-api-access-wvj2q\") pod \"b4773e7a-a50c-42e6-bb27-a25b0055b8e3\" (UID: \"b4773e7a-a50c-42e6-bb27-a25b0055b8e3\") " Jan 21 11:26:17 crc kubenswrapper[4824]: I0121 11:26:17.506773 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/93c8ce3d-a4f0-4cc7-8592-61653eb09d3c-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "93c8ce3d-a4f0-4cc7-8592-61653eb09d3c" (UID: "93c8ce3d-a4f0-4cc7-8592-61653eb09d3c"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:26:17 crc kubenswrapper[4824]: I0121 11:26:17.506782 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a10a8f17-c27b-4544-8f9d-a6afd9991a8e-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "a10a8f17-c27b-4544-8f9d-a6afd9991a8e" (UID: "a10a8f17-c27b-4544-8f9d-a6afd9991a8e"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:26:17 crc kubenswrapper[4824]: I0121 11:26:17.507240 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b4773e7a-a50c-42e6-bb27-a25b0055b8e3-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b4773e7a-a50c-42e6-bb27-a25b0055b8e3" (UID: "b4773e7a-a50c-42e6-bb27-a25b0055b8e3"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:26:17 crc kubenswrapper[4824]: I0121 11:26:17.507286 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5c0ddc66-7696-4816-8e96-340aeb39d242-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "5c0ddc66-7696-4816-8e96-340aeb39d242" (UID: "5c0ddc66-7696-4816-8e96-340aeb39d242"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:26:17 crc kubenswrapper[4824]: I0121 11:26:17.507490 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9be37b35-7ad5-434a-8d16-c2f9f1661821-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "9be37b35-7ad5-434a-8d16-c2f9f1661821" (UID: "9be37b35-7ad5-434a-8d16-c2f9f1661821"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:26:17 crc kubenswrapper[4824]: I0121 11:26:17.508175 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7b747df8-e61f-4689-a08c-0dc6688f1a0e-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "7b747df8-e61f-4689-a08c-0dc6688f1a0e" (UID: "7b747df8-e61f-4689-a08c-0dc6688f1a0e"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:26:17 crc kubenswrapper[4824]: I0121 11:26:17.508449 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9be37b35-7ad5-434a-8d16-c2f9f1661821-kube-api-access-5nfhq" (OuterVolumeSpecName: "kube-api-access-5nfhq") pod "9be37b35-7ad5-434a-8d16-c2f9f1661821" (UID: "9be37b35-7ad5-434a-8d16-c2f9f1661821"). InnerVolumeSpecName "kube-api-access-5nfhq". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:26:17 crc kubenswrapper[4824]: I0121 11:26:17.508983 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/93c8ce3d-a4f0-4cc7-8592-61653eb09d3c-kube-api-access-pq7qd" (OuterVolumeSpecName: "kube-api-access-pq7qd") pod "93c8ce3d-a4f0-4cc7-8592-61653eb09d3c" (UID: "93c8ce3d-a4f0-4cc7-8592-61653eb09d3c"). InnerVolumeSpecName "kube-api-access-pq7qd". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:26:17 crc kubenswrapper[4824]: I0121 11:26:17.515495 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7b747df8-e61f-4689-a08c-0dc6688f1a0e-kube-api-access-6876j" (OuterVolumeSpecName: "kube-api-access-6876j") pod "7b747df8-e61f-4689-a08c-0dc6688f1a0e" (UID: "7b747df8-e61f-4689-a08c-0dc6688f1a0e"). InnerVolumeSpecName "kube-api-access-6876j". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:26:17 crc kubenswrapper[4824]: I0121 11:26:17.523204 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a10a8f17-c27b-4544-8f9d-a6afd9991a8e-kube-api-access-pqhbb" (OuterVolumeSpecName: "kube-api-access-pqhbb") pod "a10a8f17-c27b-4544-8f9d-a6afd9991a8e" (UID: "a10a8f17-c27b-4544-8f9d-a6afd9991a8e"). InnerVolumeSpecName "kube-api-access-pqhbb". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:26:17 crc kubenswrapper[4824]: I0121 11:26:17.523257 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5c0ddc66-7696-4816-8e96-340aeb39d242-kube-api-access-lbmg7" (OuterVolumeSpecName: "kube-api-access-lbmg7") pod "5c0ddc66-7696-4816-8e96-340aeb39d242" (UID: "5c0ddc66-7696-4816-8e96-340aeb39d242"). InnerVolumeSpecName "kube-api-access-lbmg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:26:17 crc kubenswrapper[4824]: I0121 11:26:17.526664 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b4773e7a-a50c-42e6-bb27-a25b0055b8e3-kube-api-access-wvj2q" (OuterVolumeSpecName: "kube-api-access-wvj2q") pod "b4773e7a-a50c-42e6-bb27-a25b0055b8e3" (UID: "b4773e7a-a50c-42e6-bb27-a25b0055b8e3"). InnerVolumeSpecName "kube-api-access-wvj2q". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:26:17 crc kubenswrapper[4824]: I0121 11:26:17.607544 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6876j\" (UniqueName: \"kubernetes.io/projected/7b747df8-e61f-4689-a08c-0dc6688f1a0e-kube-api-access-6876j\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:17 crc kubenswrapper[4824]: I0121 11:26:17.607805 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pq7qd\" (UniqueName: \"kubernetes.io/projected/93c8ce3d-a4f0-4cc7-8592-61653eb09d3c-kube-api-access-pq7qd\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:17 crc kubenswrapper[4824]: I0121 11:26:17.607815 4824 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a10a8f17-c27b-4544-8f9d-a6afd9991a8e-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:17 crc kubenswrapper[4824]: I0121 11:26:17.607824 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pqhbb\" (UniqueName: \"kubernetes.io/projected/a10a8f17-c27b-4544-8f9d-a6afd9991a8e-kube-api-access-pqhbb\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:17 crc kubenswrapper[4824]: I0121 11:26:17.607832 4824 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b4773e7a-a50c-42e6-bb27-a25b0055b8e3-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:17 crc kubenswrapper[4824]: I0121 11:26:17.607841 4824 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5c0ddc66-7696-4816-8e96-340aeb39d242-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:17 crc kubenswrapper[4824]: I0121 11:26:17.607848 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5nfhq\" (UniqueName: \"kubernetes.io/projected/9be37b35-7ad5-434a-8d16-c2f9f1661821-kube-api-access-5nfhq\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:17 crc kubenswrapper[4824]: I0121 11:26:17.607858 4824 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7b747df8-e61f-4689-a08c-0dc6688f1a0e-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:17 crc kubenswrapper[4824]: I0121 11:26:17.607865 4824 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/93c8ce3d-a4f0-4cc7-8592-61653eb09d3c-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:17 crc kubenswrapper[4824]: I0121 11:26:17.607873 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lbmg7\" (UniqueName: \"kubernetes.io/projected/5c0ddc66-7696-4816-8e96-340aeb39d242-kube-api-access-lbmg7\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:17 crc kubenswrapper[4824]: I0121 11:26:17.607881 4824 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9be37b35-7ad5-434a-8d16-c2f9f1661821-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:17 crc kubenswrapper[4824]: I0121 11:26:17.607890 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wvj2q\" (UniqueName: \"kubernetes.io/projected/b4773e7a-a50c-42e6-bb27-a25b0055b8e3-kube-api-access-wvj2q\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:17 crc kubenswrapper[4824]: I0121 11:26:17.705343 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-qzfrf" Jan 21 11:26:17 crc kubenswrapper[4824]: I0121 11:26:17.810624 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e94778b1-d36a-47e9-b201-9b7cd1cdb640-ovsdbserver-nb\") pod \"e94778b1-d36a-47e9-b201-9b7cd1cdb640\" (UID: \"e94778b1-d36a-47e9-b201-9b7cd1cdb640\") " Jan 21 11:26:17 crc kubenswrapper[4824]: I0121 11:26:17.810704 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e94778b1-d36a-47e9-b201-9b7cd1cdb640-ovsdbserver-sb\") pod \"e94778b1-d36a-47e9-b201-9b7cd1cdb640\" (UID: \"e94778b1-d36a-47e9-b201-9b7cd1cdb640\") " Jan 21 11:26:17 crc kubenswrapper[4824]: I0121 11:26:17.810755 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e94778b1-d36a-47e9-b201-9b7cd1cdb640-dns-svc\") pod \"e94778b1-d36a-47e9-b201-9b7cd1cdb640\" (UID: \"e94778b1-d36a-47e9-b201-9b7cd1cdb640\") " Jan 21 11:26:17 crc kubenswrapper[4824]: I0121 11:26:17.810786 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e94778b1-d36a-47e9-b201-9b7cd1cdb640-config\") pod \"e94778b1-d36a-47e9-b201-9b7cd1cdb640\" (UID: \"e94778b1-d36a-47e9-b201-9b7cd1cdb640\") " Jan 21 11:26:17 crc kubenswrapper[4824]: I0121 11:26:17.810848 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q9jhb\" (UniqueName: \"kubernetes.io/projected/e94778b1-d36a-47e9-b201-9b7cd1cdb640-kube-api-access-q9jhb\") pod \"e94778b1-d36a-47e9-b201-9b7cd1cdb640\" (UID: \"e94778b1-d36a-47e9-b201-9b7cd1cdb640\") " Jan 21 11:26:17 crc kubenswrapper[4824]: I0121 11:26:17.814663 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e94778b1-d36a-47e9-b201-9b7cd1cdb640-kube-api-access-q9jhb" (OuterVolumeSpecName: "kube-api-access-q9jhb") pod "e94778b1-d36a-47e9-b201-9b7cd1cdb640" (UID: "e94778b1-d36a-47e9-b201-9b7cd1cdb640"). InnerVolumeSpecName "kube-api-access-q9jhb". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:26:17 crc kubenswrapper[4824]: I0121 11:26:17.837174 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e94778b1-d36a-47e9-b201-9b7cd1cdb640-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "e94778b1-d36a-47e9-b201-9b7cd1cdb640" (UID: "e94778b1-d36a-47e9-b201-9b7cd1cdb640"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:26:17 crc kubenswrapper[4824]: I0121 11:26:17.839031 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e94778b1-d36a-47e9-b201-9b7cd1cdb640-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "e94778b1-d36a-47e9-b201-9b7cd1cdb640" (UID: "e94778b1-d36a-47e9-b201-9b7cd1cdb640"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:26:17 crc kubenswrapper[4824]: I0121 11:26:17.839373 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e94778b1-d36a-47e9-b201-9b7cd1cdb640-config" (OuterVolumeSpecName: "config") pod "e94778b1-d36a-47e9-b201-9b7cd1cdb640" (UID: "e94778b1-d36a-47e9-b201-9b7cd1cdb640"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:26:17 crc kubenswrapper[4824]: I0121 11:26:17.839976 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e94778b1-d36a-47e9-b201-9b7cd1cdb640-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "e94778b1-d36a-47e9-b201-9b7cd1cdb640" (UID: "e94778b1-d36a-47e9-b201-9b7cd1cdb640"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:26:17 crc kubenswrapper[4824]: I0121 11:26:17.913719 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q9jhb\" (UniqueName: \"kubernetes.io/projected/e94778b1-d36a-47e9-b201-9b7cd1cdb640-kube-api-access-q9jhb\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:17 crc kubenswrapper[4824]: I0121 11:26:17.913741 4824 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e94778b1-d36a-47e9-b201-9b7cd1cdb640-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:17 crc kubenswrapper[4824]: I0121 11:26:17.913750 4824 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e94778b1-d36a-47e9-b201-9b7cd1cdb640-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:17 crc kubenswrapper[4824]: I0121 11:26:17.913759 4824 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e94778b1-d36a-47e9-b201-9b7cd1cdb640-dns-svc\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:17 crc kubenswrapper[4824]: I0121 11:26:17.913767 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e94778b1-d36a-47e9-b201-9b7cd1cdb640-config\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:18 crc kubenswrapper[4824]: I0121 11:26:18.242182 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-3c9f-account-create-update-rtprr" event={"ID":"5c0ddc66-7696-4816-8e96-340aeb39d242","Type":"ContainerDied","Data":"e9e93c5745e7fe74ef1e643164e9a0a4e165267deeb109faf4df81c36f336691"} Jan 21 11:26:18 crc kubenswrapper[4824]: I0121 11:26:18.242223 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e9e93c5745e7fe74ef1e643164e9a0a4e165267deeb109faf4df81c36f336691" Jan 21 11:26:18 crc kubenswrapper[4824]: I0121 11:26:18.242290 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-3c9f-account-create-update-rtprr" Jan 21 11:26:18 crc kubenswrapper[4824]: I0121 11:26:18.243866 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-ngtnj" event={"ID":"9be37b35-7ad5-434a-8d16-c2f9f1661821","Type":"ContainerDied","Data":"e6fc3836b7afabeab4a7371d16c4b247a2670b8dfe835939c4f3868093185822"} Jan 21 11:26:18 crc kubenswrapper[4824]: I0121 11:26:18.243888 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e6fc3836b7afabeab4a7371d16c4b247a2670b8dfe835939c4f3868093185822" Jan 21 11:26:18 crc kubenswrapper[4824]: I0121 11:26:18.243936 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-ngtnj" Jan 21 11:26:18 crc kubenswrapper[4824]: I0121 11:26:18.245148 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-twgjw" event={"ID":"7b747df8-e61f-4689-a08c-0dc6688f1a0e","Type":"ContainerDied","Data":"5cbfd6e560404567a0bbc9ae88f4245a4c606e0f65ba74f1fdb8d8f25d487579"} Jan 21 11:26:18 crc kubenswrapper[4824]: I0121 11:26:18.245172 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5cbfd6e560404567a0bbc9ae88f4245a4c606e0f65ba74f1fdb8d8f25d487579" Jan 21 11:26:18 crc kubenswrapper[4824]: I0121 11:26:18.245213 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-twgjw" Jan 21 11:26:18 crc kubenswrapper[4824]: I0121 11:26:18.246546 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-qzfrf" event={"ID":"e94778b1-d36a-47e9-b201-9b7cd1cdb640","Type":"ContainerDied","Data":"48fb0631b23867cac173ec7a9df005de8edd1cb0df987f8dceb763f2434dd5b7"} Jan 21 11:26:18 crc kubenswrapper[4824]: I0121 11:26:18.246575 4824 scope.go:117] "RemoveContainer" containerID="8a0b04dc02e5d8c871ec744b980efd9b4e195f4024e4ef8187ce38c2f8284265" Jan 21 11:26:18 crc kubenswrapper[4824]: I0121 11:26:18.246671 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-qzfrf" Jan 21 11:26:18 crc kubenswrapper[4824]: I0121 11:26:18.248845 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-84f4b" event={"ID":"5240d35b-cbf1-472d-91da-debf418dd208","Type":"ContainerStarted","Data":"3d07104d28aceb5ad31acec840c18d559f5c98c4390845b748361ff21f72d204"} Jan 21 11:26:18 crc kubenswrapper[4824]: I0121 11:26:18.253663 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-0662-account-create-update-j7bhp" event={"ID":"b4773e7a-a50c-42e6-bb27-a25b0055b8e3","Type":"ContainerDied","Data":"6d872ef3f227d5894d64148f328773527cef85531ed41d2b5f9fdbbdf412e8bf"} Jan 21 11:26:18 crc kubenswrapper[4824]: I0121 11:26:18.253695 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6d872ef3f227d5894d64148f328773527cef85531ed41d2b5f9fdbbdf412e8bf" Jan 21 11:26:18 crc kubenswrapper[4824]: I0121 11:26:18.254718 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-0662-account-create-update-j7bhp" Jan 21 11:26:18 crc kubenswrapper[4824]: I0121 11:26:18.261068 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-nskxc" event={"ID":"a10a8f17-c27b-4544-8f9d-a6afd9991a8e","Type":"ContainerDied","Data":"d2e64d054f528de025088752cbba2a64296b9807a6337f5fdd775173aefc9828"} Jan 21 11:26:18 crc kubenswrapper[4824]: I0121 11:26:18.261097 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d2e64d054f528de025088752cbba2a64296b9807a6337f5fdd775173aefc9828" Jan 21 11:26:18 crc kubenswrapper[4824]: I0121 11:26:18.261076 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-nskxc" Jan 21 11:26:18 crc kubenswrapper[4824]: I0121 11:26:18.265575 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-60f6-account-create-update-pjh6n" event={"ID":"93c8ce3d-a4f0-4cc7-8592-61653eb09d3c","Type":"ContainerDied","Data":"a60e0d515d078f64392d5d1ca5e1034d522ce6ad7bf98f9a82fa5e036ed775d2"} Jan 21 11:26:18 crc kubenswrapper[4824]: I0121 11:26:18.265609 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a60e0d515d078f64392d5d1ca5e1034d522ce6ad7bf98f9a82fa5e036ed775d2" Jan 21 11:26:18 crc kubenswrapper[4824]: I0121 11:26:18.265618 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-60f6-account-create-update-pjh6n" Jan 21 11:26:18 crc kubenswrapper[4824]: I0121 11:26:18.271392 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-84f4b" podStartSLOduration=2.9813106400000002 podStartE2EDuration="7.271381381s" podCreationTimestamp="2026-01-21 11:26:11 +0000 UTC" firstStartedPulling="2026-01-21 11:26:12.989679987 +0000 UTC m=+915.282709279" lastFinishedPulling="2026-01-21 11:26:17.279750729 +0000 UTC m=+919.572780020" observedRunningTime="2026-01-21 11:26:18.265777851 +0000 UTC m=+920.558807143" watchObservedRunningTime="2026-01-21 11:26:18.271381381 +0000 UTC m=+920.564410673" Jan 21 11:26:18 crc kubenswrapper[4824]: I0121 11:26:18.296160 4824 scope.go:117] "RemoveContainer" containerID="a517f23f781efca10654bdff86d5e2efc32071ea09d5af5829095603d962c629" Jan 21 11:26:18 crc kubenswrapper[4824]: I0121 11:26:18.342664 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-698758b865-qzfrf"] Jan 21 11:26:18 crc kubenswrapper[4824]: I0121 11:26:18.347720 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-698758b865-qzfrf"] Jan 21 11:26:19 crc kubenswrapper[4824]: I0121 11:26:19.273749 4824 generic.go:334] "Generic (PLEG): container finished" podID="5240d35b-cbf1-472d-91da-debf418dd208" containerID="3d07104d28aceb5ad31acec840c18d559f5c98c4390845b748361ff21f72d204" exitCode=0 Jan 21 11:26:19 crc kubenswrapper[4824]: I0121 11:26:19.273784 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-84f4b" event={"ID":"5240d35b-cbf1-472d-91da-debf418dd208","Type":"ContainerDied","Data":"3d07104d28aceb5ad31acec840c18d559f5c98c4390845b748361ff21f72d204"} Jan 21 11:26:20 crc kubenswrapper[4824]: I0121 11:26:20.056175 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e94778b1-d36a-47e9-b201-9b7cd1cdb640" path="/var/lib/kubelet/pods/e94778b1-d36a-47e9-b201-9b7cd1cdb640/volumes" Jan 21 11:26:20 crc kubenswrapper[4824]: I0121 11:26:20.553663 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-84f4b" Jan 21 11:26:20 crc kubenswrapper[4824]: I0121 11:26:20.652827 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5240d35b-cbf1-472d-91da-debf418dd208-config-data\") pod \"5240d35b-cbf1-472d-91da-debf418dd208\" (UID: \"5240d35b-cbf1-472d-91da-debf418dd208\") " Jan 21 11:26:20 crc kubenswrapper[4824]: I0121 11:26:20.652902 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f4nqn\" (UniqueName: \"kubernetes.io/projected/5240d35b-cbf1-472d-91da-debf418dd208-kube-api-access-f4nqn\") pod \"5240d35b-cbf1-472d-91da-debf418dd208\" (UID: \"5240d35b-cbf1-472d-91da-debf418dd208\") " Jan 21 11:26:20 crc kubenswrapper[4824]: I0121 11:26:20.652954 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5240d35b-cbf1-472d-91da-debf418dd208-combined-ca-bundle\") pod \"5240d35b-cbf1-472d-91da-debf418dd208\" (UID: \"5240d35b-cbf1-472d-91da-debf418dd208\") " Jan 21 11:26:20 crc kubenswrapper[4824]: I0121 11:26:20.658900 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5240d35b-cbf1-472d-91da-debf418dd208-kube-api-access-f4nqn" (OuterVolumeSpecName: "kube-api-access-f4nqn") pod "5240d35b-cbf1-472d-91da-debf418dd208" (UID: "5240d35b-cbf1-472d-91da-debf418dd208"). InnerVolumeSpecName "kube-api-access-f4nqn". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:26:20 crc kubenswrapper[4824]: I0121 11:26:20.671162 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5240d35b-cbf1-472d-91da-debf418dd208-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5240d35b-cbf1-472d-91da-debf418dd208" (UID: "5240d35b-cbf1-472d-91da-debf418dd208"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:26:20 crc kubenswrapper[4824]: I0121 11:26:20.683663 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5240d35b-cbf1-472d-91da-debf418dd208-config-data" (OuterVolumeSpecName: "config-data") pod "5240d35b-cbf1-472d-91da-debf418dd208" (UID: "5240d35b-cbf1-472d-91da-debf418dd208"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:26:20 crc kubenswrapper[4824]: I0121 11:26:20.754443 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5240d35b-cbf1-472d-91da-debf418dd208-config-data\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:20 crc kubenswrapper[4824]: I0121 11:26:20.754468 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f4nqn\" (UniqueName: \"kubernetes.io/projected/5240d35b-cbf1-472d-91da-debf418dd208-kube-api-access-f4nqn\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:20 crc kubenswrapper[4824]: I0121 11:26:20.754477 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5240d35b-cbf1-472d-91da-debf418dd208-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.279995 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-kb6p9"] Jan 21 11:26:21 crc kubenswrapper[4824]: E0121 11:26:21.280638 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e94778b1-d36a-47e9-b201-9b7cd1cdb640" containerName="dnsmasq-dns" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.280656 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="e94778b1-d36a-47e9-b201-9b7cd1cdb640" containerName="dnsmasq-dns" Jan 21 11:26:21 crc kubenswrapper[4824]: E0121 11:26:21.280698 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93c8ce3d-a4f0-4cc7-8592-61653eb09d3c" containerName="mariadb-account-create-update" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.280705 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="93c8ce3d-a4f0-4cc7-8592-61653eb09d3c" containerName="mariadb-account-create-update" Jan 21 11:26:21 crc kubenswrapper[4824]: E0121 11:26:21.280714 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c0ddc66-7696-4816-8e96-340aeb39d242" containerName="mariadb-account-create-update" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.280719 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c0ddc66-7696-4816-8e96-340aeb39d242" containerName="mariadb-account-create-update" Jan 21 11:26:21 crc kubenswrapper[4824]: E0121 11:26:21.280730 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a10a8f17-c27b-4544-8f9d-a6afd9991a8e" containerName="mariadb-database-create" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.280736 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="a10a8f17-c27b-4544-8f9d-a6afd9991a8e" containerName="mariadb-database-create" Jan 21 11:26:21 crc kubenswrapper[4824]: E0121 11:26:21.280763 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8600373d-f466-43d4-92fc-1fa938f6e91b" containerName="mariadb-database-create" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.280769 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="8600373d-f466-43d4-92fc-1fa938f6e91b" containerName="mariadb-database-create" Jan 21 11:26:21 crc kubenswrapper[4824]: E0121 11:26:21.280775 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b747df8-e61f-4689-a08c-0dc6688f1a0e" containerName="mariadb-database-create" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.280782 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b747df8-e61f-4689-a08c-0dc6688f1a0e" containerName="mariadb-database-create" Jan 21 11:26:21 crc kubenswrapper[4824]: E0121 11:26:21.280794 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5240d35b-cbf1-472d-91da-debf418dd208" containerName="keystone-db-sync" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.280800 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="5240d35b-cbf1-472d-91da-debf418dd208" containerName="keystone-db-sync" Jan 21 11:26:21 crc kubenswrapper[4824]: E0121 11:26:21.280818 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b4773e7a-a50c-42e6-bb27-a25b0055b8e3" containerName="mariadb-account-create-update" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.280840 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="b4773e7a-a50c-42e6-bb27-a25b0055b8e3" containerName="mariadb-account-create-update" Jan 21 11:26:21 crc kubenswrapper[4824]: E0121 11:26:21.280847 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e94778b1-d36a-47e9-b201-9b7cd1cdb640" containerName="init" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.280852 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="e94778b1-d36a-47e9-b201-9b7cd1cdb640" containerName="init" Jan 21 11:26:21 crc kubenswrapper[4824]: E0121 11:26:21.280858 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9be37b35-7ad5-434a-8d16-c2f9f1661821" containerName="mariadb-database-create" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.280863 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="9be37b35-7ad5-434a-8d16-c2f9f1661821" containerName="mariadb-database-create" Jan 21 11:26:21 crc kubenswrapper[4824]: E0121 11:26:21.280876 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fe6e7565-793b-4555-b6d5-758f87a5b9c7" containerName="mariadb-account-create-update" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.280881 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="fe6e7565-793b-4555-b6d5-758f87a5b9c7" containerName="mariadb-account-create-update" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.281072 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="93c8ce3d-a4f0-4cc7-8592-61653eb09d3c" containerName="mariadb-account-create-update" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.281084 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="8600373d-f466-43d4-92fc-1fa938f6e91b" containerName="mariadb-database-create" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.281093 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="a10a8f17-c27b-4544-8f9d-a6afd9991a8e" containerName="mariadb-database-create" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.281102 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="9be37b35-7ad5-434a-8d16-c2f9f1661821" containerName="mariadb-database-create" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.281109 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="7b747df8-e61f-4689-a08c-0dc6688f1a0e" containerName="mariadb-database-create" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.281136 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="5240d35b-cbf1-472d-91da-debf418dd208" containerName="keystone-db-sync" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.281143 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="fe6e7565-793b-4555-b6d5-758f87a5b9c7" containerName="mariadb-account-create-update" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.281155 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="5c0ddc66-7696-4816-8e96-340aeb39d242" containerName="mariadb-account-create-update" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.281164 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="e94778b1-d36a-47e9-b201-9b7cd1cdb640" containerName="dnsmasq-dns" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.281173 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="b4773e7a-a50c-42e6-bb27-a25b0055b8e3" containerName="mariadb-account-create-update" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.285596 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-847c4cc679-kb6p9" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.296897 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-84f4b" event={"ID":"5240d35b-cbf1-472d-91da-debf418dd208","Type":"ContainerDied","Data":"9132d198700ff29f709588ab7e526f22887b3f414c249d3511b9eb2a27c47581"} Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.296939 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9132d198700ff29f709588ab7e526f22887b3f414c249d3511b9eb2a27c47581" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.297071 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-84f4b" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.313557 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-kb6p9"] Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.342801 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-sw5v2"] Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.343700 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-sw5v2" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.353723 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.353817 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.353969 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.354078 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.354588 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-qxhzv" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.363877 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bcfd76c3-d71d-4f02-af00-d01b113c4d2e-dns-svc\") pod \"dnsmasq-dns-847c4cc679-kb6p9\" (UID: \"bcfd76c3-d71d-4f02-af00-d01b113c4d2e\") " pod="openstack/dnsmasq-dns-847c4cc679-kb6p9" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.363923 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bcfd76c3-d71d-4f02-af00-d01b113c4d2e-ovsdbserver-sb\") pod \"dnsmasq-dns-847c4cc679-kb6p9\" (UID: \"bcfd76c3-d71d-4f02-af00-d01b113c4d2e\") " pod="openstack/dnsmasq-dns-847c4cc679-kb6p9" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.363993 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h5zvt\" (UniqueName: \"kubernetes.io/projected/bcfd76c3-d71d-4f02-af00-d01b113c4d2e-kube-api-access-h5zvt\") pod \"dnsmasq-dns-847c4cc679-kb6p9\" (UID: \"bcfd76c3-d71d-4f02-af00-d01b113c4d2e\") " pod="openstack/dnsmasq-dns-847c4cc679-kb6p9" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.364016 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bcfd76c3-d71d-4f02-af00-d01b113c4d2e-config\") pod \"dnsmasq-dns-847c4cc679-kb6p9\" (UID: \"bcfd76c3-d71d-4f02-af00-d01b113c4d2e\") " pod="openstack/dnsmasq-dns-847c4cc679-kb6p9" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.364063 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bcfd76c3-d71d-4f02-af00-d01b113c4d2e-dns-swift-storage-0\") pod \"dnsmasq-dns-847c4cc679-kb6p9\" (UID: \"bcfd76c3-d71d-4f02-af00-d01b113c4d2e\") " pod="openstack/dnsmasq-dns-847c4cc679-kb6p9" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.364102 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bcfd76c3-d71d-4f02-af00-d01b113c4d2e-ovsdbserver-nb\") pod \"dnsmasq-dns-847c4cc679-kb6p9\" (UID: \"bcfd76c3-d71d-4f02-af00-d01b113c4d2e\") " pod="openstack/dnsmasq-dns-847c4cc679-kb6p9" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.369056 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-sw5v2"] Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.444108 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-db-sync-jrsn8"] Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.445022 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-jrsn8" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.446698 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-heat-dockercfg-cj2g7" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.447052 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-config-data" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.455183 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-sync-jrsn8"] Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.465264 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4f0bb2ed-9f0d-4d49-b224-8d05b5de0829-config-data\") pod \"keystone-bootstrap-sw5v2\" (UID: \"4f0bb2ed-9f0d-4d49-b224-8d05b5de0829\") " pod="openstack/keystone-bootstrap-sw5v2" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.465323 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ztr5x\" (UniqueName: \"kubernetes.io/projected/4f0bb2ed-9f0d-4d49-b224-8d05b5de0829-kube-api-access-ztr5x\") pod \"keystone-bootstrap-sw5v2\" (UID: \"4f0bb2ed-9f0d-4d49-b224-8d05b5de0829\") " pod="openstack/keystone-bootstrap-sw5v2" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.465453 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bcfd76c3-d71d-4f02-af00-d01b113c4d2e-dns-svc\") pod \"dnsmasq-dns-847c4cc679-kb6p9\" (UID: \"bcfd76c3-d71d-4f02-af00-d01b113c4d2e\") " pod="openstack/dnsmasq-dns-847c4cc679-kb6p9" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.465495 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bcfd76c3-d71d-4f02-af00-d01b113c4d2e-ovsdbserver-sb\") pod \"dnsmasq-dns-847c4cc679-kb6p9\" (UID: \"bcfd76c3-d71d-4f02-af00-d01b113c4d2e\") " pod="openstack/dnsmasq-dns-847c4cc679-kb6p9" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.465518 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/4f0bb2ed-9f0d-4d49-b224-8d05b5de0829-fernet-keys\") pod \"keystone-bootstrap-sw5v2\" (UID: \"4f0bb2ed-9f0d-4d49-b224-8d05b5de0829\") " pod="openstack/keystone-bootstrap-sw5v2" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.465580 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h5zvt\" (UniqueName: \"kubernetes.io/projected/bcfd76c3-d71d-4f02-af00-d01b113c4d2e-kube-api-access-h5zvt\") pod \"dnsmasq-dns-847c4cc679-kb6p9\" (UID: \"bcfd76c3-d71d-4f02-af00-d01b113c4d2e\") " pod="openstack/dnsmasq-dns-847c4cc679-kb6p9" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.465605 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bcfd76c3-d71d-4f02-af00-d01b113c4d2e-config\") pod \"dnsmasq-dns-847c4cc679-kb6p9\" (UID: \"bcfd76c3-d71d-4f02-af00-d01b113c4d2e\") " pod="openstack/dnsmasq-dns-847c4cc679-kb6p9" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.465641 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f0bb2ed-9f0d-4d49-b224-8d05b5de0829-combined-ca-bundle\") pod \"keystone-bootstrap-sw5v2\" (UID: \"4f0bb2ed-9f0d-4d49-b224-8d05b5de0829\") " pod="openstack/keystone-bootstrap-sw5v2" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.465671 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bcfd76c3-d71d-4f02-af00-d01b113c4d2e-dns-swift-storage-0\") pod \"dnsmasq-dns-847c4cc679-kb6p9\" (UID: \"bcfd76c3-d71d-4f02-af00-d01b113c4d2e\") " pod="openstack/dnsmasq-dns-847c4cc679-kb6p9" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.465733 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bcfd76c3-d71d-4f02-af00-d01b113c4d2e-ovsdbserver-nb\") pod \"dnsmasq-dns-847c4cc679-kb6p9\" (UID: \"bcfd76c3-d71d-4f02-af00-d01b113c4d2e\") " pod="openstack/dnsmasq-dns-847c4cc679-kb6p9" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.465770 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4f0bb2ed-9f0d-4d49-b224-8d05b5de0829-scripts\") pod \"keystone-bootstrap-sw5v2\" (UID: \"4f0bb2ed-9f0d-4d49-b224-8d05b5de0829\") " pod="openstack/keystone-bootstrap-sw5v2" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.465791 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/4f0bb2ed-9f0d-4d49-b224-8d05b5de0829-credential-keys\") pod \"keystone-bootstrap-sw5v2\" (UID: \"4f0bb2ed-9f0d-4d49-b224-8d05b5de0829\") " pod="openstack/keystone-bootstrap-sw5v2" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.466316 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bcfd76c3-d71d-4f02-af00-d01b113c4d2e-dns-svc\") pod \"dnsmasq-dns-847c4cc679-kb6p9\" (UID: \"bcfd76c3-d71d-4f02-af00-d01b113c4d2e\") " pod="openstack/dnsmasq-dns-847c4cc679-kb6p9" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.466516 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bcfd76c3-d71d-4f02-af00-d01b113c4d2e-ovsdbserver-nb\") pod \"dnsmasq-dns-847c4cc679-kb6p9\" (UID: \"bcfd76c3-d71d-4f02-af00-d01b113c4d2e\") " pod="openstack/dnsmasq-dns-847c4cc679-kb6p9" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.467005 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bcfd76c3-d71d-4f02-af00-d01b113c4d2e-ovsdbserver-sb\") pod \"dnsmasq-dns-847c4cc679-kb6p9\" (UID: \"bcfd76c3-d71d-4f02-af00-d01b113c4d2e\") " pod="openstack/dnsmasq-dns-847c4cc679-kb6p9" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.467549 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bcfd76c3-d71d-4f02-af00-d01b113c4d2e-config\") pod \"dnsmasq-dns-847c4cc679-kb6p9\" (UID: \"bcfd76c3-d71d-4f02-af00-d01b113c4d2e\") " pod="openstack/dnsmasq-dns-847c4cc679-kb6p9" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.467602 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bcfd76c3-d71d-4f02-af00-d01b113c4d2e-dns-swift-storage-0\") pod \"dnsmasq-dns-847c4cc679-kb6p9\" (UID: \"bcfd76c3-d71d-4f02-af00-d01b113c4d2e\") " pod="openstack/dnsmasq-dns-847c4cc679-kb6p9" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.475556 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.478801 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.480201 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.480570 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.485724 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h5zvt\" (UniqueName: \"kubernetes.io/projected/bcfd76c3-d71d-4f02-af00-d01b113c4d2e-kube-api-access-h5zvt\") pod \"dnsmasq-dns-847c4cc679-kb6p9\" (UID: \"bcfd76c3-d71d-4f02-af00-d01b113c4d2e\") " pod="openstack/dnsmasq-dns-847c4cc679-kb6p9" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.491687 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.567320 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sgmtk\" (UniqueName: \"kubernetes.io/projected/bbefda93-dc56-43c5-b131-56af96ca832b-kube-api-access-sgmtk\") pod \"ceilometer-0\" (UID: \"bbefda93-dc56-43c5-b131-56af96ca832b\") " pod="openstack/ceilometer-0" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.568035 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xrdwp\" (UniqueName: \"kubernetes.io/projected/cadf2d38-24f9-43ab-8e29-803d97a0ff06-kube-api-access-xrdwp\") pod \"heat-db-sync-jrsn8\" (UID: \"cadf2d38-24f9-43ab-8e29-803d97a0ff06\") " pod="openstack/heat-db-sync-jrsn8" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.568138 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cadf2d38-24f9-43ab-8e29-803d97a0ff06-config-data\") pod \"heat-db-sync-jrsn8\" (UID: \"cadf2d38-24f9-43ab-8e29-803d97a0ff06\") " pod="openstack/heat-db-sync-jrsn8" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.568223 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/4f0bb2ed-9f0d-4d49-b224-8d05b5de0829-fernet-keys\") pod \"keystone-bootstrap-sw5v2\" (UID: \"4f0bb2ed-9f0d-4d49-b224-8d05b5de0829\") " pod="openstack/keystone-bootstrap-sw5v2" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.568287 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bbefda93-dc56-43c5-b131-56af96ca832b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"bbefda93-dc56-43c5-b131-56af96ca832b\") " pod="openstack/ceilometer-0" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.568387 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bbefda93-dc56-43c5-b131-56af96ca832b-config-data\") pod \"ceilometer-0\" (UID: \"bbefda93-dc56-43c5-b131-56af96ca832b\") " pod="openstack/ceilometer-0" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.568466 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f0bb2ed-9f0d-4d49-b224-8d05b5de0829-combined-ca-bundle\") pod \"keystone-bootstrap-sw5v2\" (UID: \"4f0bb2ed-9f0d-4d49-b224-8d05b5de0829\") " pod="openstack/keystone-bootstrap-sw5v2" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.568531 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bbefda93-dc56-43c5-b131-56af96ca832b-run-httpd\") pod \"ceilometer-0\" (UID: \"bbefda93-dc56-43c5-b131-56af96ca832b\") " pod="openstack/ceilometer-0" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.568612 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bbefda93-dc56-43c5-b131-56af96ca832b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"bbefda93-dc56-43c5-b131-56af96ca832b\") " pod="openstack/ceilometer-0" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.568710 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4f0bb2ed-9f0d-4d49-b224-8d05b5de0829-scripts\") pod \"keystone-bootstrap-sw5v2\" (UID: \"4f0bb2ed-9f0d-4d49-b224-8d05b5de0829\") " pod="openstack/keystone-bootstrap-sw5v2" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.568845 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/4f0bb2ed-9f0d-4d49-b224-8d05b5de0829-credential-keys\") pod \"keystone-bootstrap-sw5v2\" (UID: \"4f0bb2ed-9f0d-4d49-b224-8d05b5de0829\") " pod="openstack/keystone-bootstrap-sw5v2" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.568909 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bbefda93-dc56-43c5-b131-56af96ca832b-scripts\") pod \"ceilometer-0\" (UID: \"bbefda93-dc56-43c5-b131-56af96ca832b\") " pod="openstack/ceilometer-0" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.569006 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4f0bb2ed-9f0d-4d49-b224-8d05b5de0829-config-data\") pod \"keystone-bootstrap-sw5v2\" (UID: \"4f0bb2ed-9f0d-4d49-b224-8d05b5de0829\") " pod="openstack/keystone-bootstrap-sw5v2" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.569075 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cadf2d38-24f9-43ab-8e29-803d97a0ff06-combined-ca-bundle\") pod \"heat-db-sync-jrsn8\" (UID: \"cadf2d38-24f9-43ab-8e29-803d97a0ff06\") " pod="openstack/heat-db-sync-jrsn8" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.569145 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bbefda93-dc56-43c5-b131-56af96ca832b-log-httpd\") pod \"ceilometer-0\" (UID: \"bbefda93-dc56-43c5-b131-56af96ca832b\") " pod="openstack/ceilometer-0" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.569223 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ztr5x\" (UniqueName: \"kubernetes.io/projected/4f0bb2ed-9f0d-4d49-b224-8d05b5de0829-kube-api-access-ztr5x\") pod \"keystone-bootstrap-sw5v2\" (UID: \"4f0bb2ed-9f0d-4d49-b224-8d05b5de0829\") " pod="openstack/keystone-bootstrap-sw5v2" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.574230 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4f0bb2ed-9f0d-4d49-b224-8d05b5de0829-scripts\") pod \"keystone-bootstrap-sw5v2\" (UID: \"4f0bb2ed-9f0d-4d49-b224-8d05b5de0829\") " pod="openstack/keystone-bootstrap-sw5v2" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.574747 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/4f0bb2ed-9f0d-4d49-b224-8d05b5de0829-fernet-keys\") pod \"keystone-bootstrap-sw5v2\" (UID: \"4f0bb2ed-9f0d-4d49-b224-8d05b5de0829\") " pod="openstack/keystone-bootstrap-sw5v2" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.575086 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4f0bb2ed-9f0d-4d49-b224-8d05b5de0829-config-data\") pod \"keystone-bootstrap-sw5v2\" (UID: \"4f0bb2ed-9f0d-4d49-b224-8d05b5de0829\") " pod="openstack/keystone-bootstrap-sw5v2" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.575274 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f0bb2ed-9f0d-4d49-b224-8d05b5de0829-combined-ca-bundle\") pod \"keystone-bootstrap-sw5v2\" (UID: \"4f0bb2ed-9f0d-4d49-b224-8d05b5de0829\") " pod="openstack/keystone-bootstrap-sw5v2" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.580639 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/4f0bb2ed-9f0d-4d49-b224-8d05b5de0829-credential-keys\") pod \"keystone-bootstrap-sw5v2\" (UID: \"4f0bb2ed-9f0d-4d49-b224-8d05b5de0829\") " pod="openstack/keystone-bootstrap-sw5v2" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.590315 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ztr5x\" (UniqueName: \"kubernetes.io/projected/4f0bb2ed-9f0d-4d49-b224-8d05b5de0829-kube-api-access-ztr5x\") pod \"keystone-bootstrap-sw5v2\" (UID: \"4f0bb2ed-9f0d-4d49-b224-8d05b5de0829\") " pod="openstack/keystone-bootstrap-sw5v2" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.601253 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-847c4cc679-kb6p9" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.619750 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-kb6p9"] Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.631900 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-fh8q6"] Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.634546 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-785d8bcb8c-fh8q6" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.641290 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-fh8q6"] Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.663772 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-sw5v2" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.665461 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-6vgh6"] Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.666384 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-6vgh6" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.676929 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xrdwp\" (UniqueName: \"kubernetes.io/projected/cadf2d38-24f9-43ab-8e29-803d97a0ff06-kube-api-access-xrdwp\") pod \"heat-db-sync-jrsn8\" (UID: \"cadf2d38-24f9-43ab-8e29-803d97a0ff06\") " pod="openstack/heat-db-sync-jrsn8" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.677010 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cadf2d38-24f9-43ab-8e29-803d97a0ff06-config-data\") pod \"heat-db-sync-jrsn8\" (UID: \"cadf2d38-24f9-43ab-8e29-803d97a0ff06\") " pod="openstack/heat-db-sync-jrsn8" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.677035 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8214df67-a984-4a5b-bcbb-6b16ab42f5fe-dns-svc\") pod \"dnsmasq-dns-785d8bcb8c-fh8q6\" (UID: \"8214df67-a984-4a5b-bcbb-6b16ab42f5fe\") " pod="openstack/dnsmasq-dns-785d8bcb8c-fh8q6" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.677053 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bbefda93-dc56-43c5-b131-56af96ca832b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"bbefda93-dc56-43c5-b131-56af96ca832b\") " pod="openstack/ceilometer-0" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.677089 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8214df67-a984-4a5b-bcbb-6b16ab42f5fe-ovsdbserver-sb\") pod \"dnsmasq-dns-785d8bcb8c-fh8q6\" (UID: \"8214df67-a984-4a5b-bcbb-6b16ab42f5fe\") " pod="openstack/dnsmasq-dns-785d8bcb8c-fh8q6" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.677119 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8214df67-a984-4a5b-bcbb-6b16ab42f5fe-config\") pod \"dnsmasq-dns-785d8bcb8c-fh8q6\" (UID: \"8214df67-a984-4a5b-bcbb-6b16ab42f5fe\") " pod="openstack/dnsmasq-dns-785d8bcb8c-fh8q6" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.677140 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bbefda93-dc56-43c5-b131-56af96ca832b-config-data\") pod \"ceilometer-0\" (UID: \"bbefda93-dc56-43c5-b131-56af96ca832b\") " pod="openstack/ceilometer-0" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.677164 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bbefda93-dc56-43c5-b131-56af96ca832b-run-httpd\") pod \"ceilometer-0\" (UID: \"bbefda93-dc56-43c5-b131-56af96ca832b\") " pod="openstack/ceilometer-0" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.677183 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8214df67-a984-4a5b-bcbb-6b16ab42f5fe-dns-swift-storage-0\") pod \"dnsmasq-dns-785d8bcb8c-fh8q6\" (UID: \"8214df67-a984-4a5b-bcbb-6b16ab42f5fe\") " pod="openstack/dnsmasq-dns-785d8bcb8c-fh8q6" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.677229 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bbefda93-dc56-43c5-b131-56af96ca832b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"bbefda93-dc56-43c5-b131-56af96ca832b\") " pod="openstack/ceilometer-0" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.677267 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bbefda93-dc56-43c5-b131-56af96ca832b-scripts\") pod \"ceilometer-0\" (UID: \"bbefda93-dc56-43c5-b131-56af96ca832b\") " pod="openstack/ceilometer-0" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.677297 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8214df67-a984-4a5b-bcbb-6b16ab42f5fe-ovsdbserver-nb\") pod \"dnsmasq-dns-785d8bcb8c-fh8q6\" (UID: \"8214df67-a984-4a5b-bcbb-6b16ab42f5fe\") " pod="openstack/dnsmasq-dns-785d8bcb8c-fh8q6" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.677329 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cadf2d38-24f9-43ab-8e29-803d97a0ff06-combined-ca-bundle\") pod \"heat-db-sync-jrsn8\" (UID: \"cadf2d38-24f9-43ab-8e29-803d97a0ff06\") " pod="openstack/heat-db-sync-jrsn8" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.677360 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bbefda93-dc56-43c5-b131-56af96ca832b-log-httpd\") pod \"ceilometer-0\" (UID: \"bbefda93-dc56-43c5-b131-56af96ca832b\") " pod="openstack/ceilometer-0" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.677380 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x8x9b\" (UniqueName: \"kubernetes.io/projected/8214df67-a984-4a5b-bcbb-6b16ab42f5fe-kube-api-access-x8x9b\") pod \"dnsmasq-dns-785d8bcb8c-fh8q6\" (UID: \"8214df67-a984-4a5b-bcbb-6b16ab42f5fe\") " pod="openstack/dnsmasq-dns-785d8bcb8c-fh8q6" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.677422 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sgmtk\" (UniqueName: \"kubernetes.io/projected/bbefda93-dc56-43c5-b131-56af96ca832b-kube-api-access-sgmtk\") pod \"ceilometer-0\" (UID: \"bbefda93-dc56-43c5-b131-56af96ca832b\") " pod="openstack/ceilometer-0" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.679424 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bbefda93-dc56-43c5-b131-56af96ca832b-run-httpd\") pod \"ceilometer-0\" (UID: \"bbefda93-dc56-43c5-b131-56af96ca832b\") " pod="openstack/ceilometer-0" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.680009 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.680343 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-n8hrk" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.680539 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.680870 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bbefda93-dc56-43c5-b131-56af96ca832b-log-httpd\") pod \"ceilometer-0\" (UID: \"bbefda93-dc56-43c5-b131-56af96ca832b\") " pod="openstack/ceilometer-0" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.683769 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bbefda93-dc56-43c5-b131-56af96ca832b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"bbefda93-dc56-43c5-b131-56af96ca832b\") " pod="openstack/ceilometer-0" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.684387 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bbefda93-dc56-43c5-b131-56af96ca832b-scripts\") pod \"ceilometer-0\" (UID: \"bbefda93-dc56-43c5-b131-56af96ca832b\") " pod="openstack/ceilometer-0" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.684734 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bbefda93-dc56-43c5-b131-56af96ca832b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"bbefda93-dc56-43c5-b131-56af96ca832b\") " pod="openstack/ceilometer-0" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.685535 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bbefda93-dc56-43c5-b131-56af96ca832b-config-data\") pod \"ceilometer-0\" (UID: \"bbefda93-dc56-43c5-b131-56af96ca832b\") " pod="openstack/ceilometer-0" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.686347 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cadf2d38-24f9-43ab-8e29-803d97a0ff06-combined-ca-bundle\") pod \"heat-db-sync-jrsn8\" (UID: \"cadf2d38-24f9-43ab-8e29-803d97a0ff06\") " pod="openstack/heat-db-sync-jrsn8" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.690827 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cadf2d38-24f9-43ab-8e29-803d97a0ff06-config-data\") pod \"heat-db-sync-jrsn8\" (UID: \"cadf2d38-24f9-43ab-8e29-803d97a0ff06\") " pod="openstack/heat-db-sync-jrsn8" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.698585 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xrdwp\" (UniqueName: \"kubernetes.io/projected/cadf2d38-24f9-43ab-8e29-803d97a0ff06-kube-api-access-xrdwp\") pod \"heat-db-sync-jrsn8\" (UID: \"cadf2d38-24f9-43ab-8e29-803d97a0ff06\") " pod="openstack/heat-db-sync-jrsn8" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.699664 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sgmtk\" (UniqueName: \"kubernetes.io/projected/bbefda93-dc56-43c5-b131-56af96ca832b-kube-api-access-sgmtk\") pod \"ceilometer-0\" (UID: \"bbefda93-dc56-43c5-b131-56af96ca832b\") " pod="openstack/ceilometer-0" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.702535 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-6vgh6"] Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.762740 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-jrsn8" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.778290 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8214df67-a984-4a5b-bcbb-6b16ab42f5fe-ovsdbserver-sb\") pod \"dnsmasq-dns-785d8bcb8c-fh8q6\" (UID: \"8214df67-a984-4a5b-bcbb-6b16ab42f5fe\") " pod="openstack/dnsmasq-dns-785d8bcb8c-fh8q6" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.778333 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gkb9f\" (UniqueName: \"kubernetes.io/projected/08b8c363-a8c1-4dfe-8fd1-1af2d055a305-kube-api-access-gkb9f\") pod \"placement-db-sync-6vgh6\" (UID: \"08b8c363-a8c1-4dfe-8fd1-1af2d055a305\") " pod="openstack/placement-db-sync-6vgh6" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.778365 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8214df67-a984-4a5b-bcbb-6b16ab42f5fe-config\") pod \"dnsmasq-dns-785d8bcb8c-fh8q6\" (UID: \"8214df67-a984-4a5b-bcbb-6b16ab42f5fe\") " pod="openstack/dnsmasq-dns-785d8bcb8c-fh8q6" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.778390 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08b8c363-a8c1-4dfe-8fd1-1af2d055a305-combined-ca-bundle\") pod \"placement-db-sync-6vgh6\" (UID: \"08b8c363-a8c1-4dfe-8fd1-1af2d055a305\") " pod="openstack/placement-db-sync-6vgh6" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.778422 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8214df67-a984-4a5b-bcbb-6b16ab42f5fe-dns-swift-storage-0\") pod \"dnsmasq-dns-785d8bcb8c-fh8q6\" (UID: \"8214df67-a984-4a5b-bcbb-6b16ab42f5fe\") " pod="openstack/dnsmasq-dns-785d8bcb8c-fh8q6" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.778456 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/08b8c363-a8c1-4dfe-8fd1-1af2d055a305-scripts\") pod \"placement-db-sync-6vgh6\" (UID: \"08b8c363-a8c1-4dfe-8fd1-1af2d055a305\") " pod="openstack/placement-db-sync-6vgh6" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.778510 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/08b8c363-a8c1-4dfe-8fd1-1af2d055a305-config-data\") pod \"placement-db-sync-6vgh6\" (UID: \"08b8c363-a8c1-4dfe-8fd1-1af2d055a305\") " pod="openstack/placement-db-sync-6vgh6" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.778545 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8214df67-a984-4a5b-bcbb-6b16ab42f5fe-ovsdbserver-nb\") pod \"dnsmasq-dns-785d8bcb8c-fh8q6\" (UID: \"8214df67-a984-4a5b-bcbb-6b16ab42f5fe\") " pod="openstack/dnsmasq-dns-785d8bcb8c-fh8q6" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.778595 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x8x9b\" (UniqueName: \"kubernetes.io/projected/8214df67-a984-4a5b-bcbb-6b16ab42f5fe-kube-api-access-x8x9b\") pod \"dnsmasq-dns-785d8bcb8c-fh8q6\" (UID: \"8214df67-a984-4a5b-bcbb-6b16ab42f5fe\") " pod="openstack/dnsmasq-dns-785d8bcb8c-fh8q6" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.778758 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/08b8c363-a8c1-4dfe-8fd1-1af2d055a305-logs\") pod \"placement-db-sync-6vgh6\" (UID: \"08b8c363-a8c1-4dfe-8fd1-1af2d055a305\") " pod="openstack/placement-db-sync-6vgh6" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.778812 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8214df67-a984-4a5b-bcbb-6b16ab42f5fe-dns-svc\") pod \"dnsmasq-dns-785d8bcb8c-fh8q6\" (UID: \"8214df67-a984-4a5b-bcbb-6b16ab42f5fe\") " pod="openstack/dnsmasq-dns-785d8bcb8c-fh8q6" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.779520 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8214df67-a984-4a5b-bcbb-6b16ab42f5fe-config\") pod \"dnsmasq-dns-785d8bcb8c-fh8q6\" (UID: \"8214df67-a984-4a5b-bcbb-6b16ab42f5fe\") " pod="openstack/dnsmasq-dns-785d8bcb8c-fh8q6" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.779536 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8214df67-a984-4a5b-bcbb-6b16ab42f5fe-dns-svc\") pod \"dnsmasq-dns-785d8bcb8c-fh8q6\" (UID: \"8214df67-a984-4a5b-bcbb-6b16ab42f5fe\") " pod="openstack/dnsmasq-dns-785d8bcb8c-fh8q6" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.779717 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8214df67-a984-4a5b-bcbb-6b16ab42f5fe-dns-swift-storage-0\") pod \"dnsmasq-dns-785d8bcb8c-fh8q6\" (UID: \"8214df67-a984-4a5b-bcbb-6b16ab42f5fe\") " pod="openstack/dnsmasq-dns-785d8bcb8c-fh8q6" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.779796 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8214df67-a984-4a5b-bcbb-6b16ab42f5fe-ovsdbserver-nb\") pod \"dnsmasq-dns-785d8bcb8c-fh8q6\" (UID: \"8214df67-a984-4a5b-bcbb-6b16ab42f5fe\") " pod="openstack/dnsmasq-dns-785d8bcb8c-fh8q6" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.780319 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8214df67-a984-4a5b-bcbb-6b16ab42f5fe-ovsdbserver-sb\") pod \"dnsmasq-dns-785d8bcb8c-fh8q6\" (UID: \"8214df67-a984-4a5b-bcbb-6b16ab42f5fe\") " pod="openstack/dnsmasq-dns-785d8bcb8c-fh8q6" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.794886 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x8x9b\" (UniqueName: \"kubernetes.io/projected/8214df67-a984-4a5b-bcbb-6b16ab42f5fe-kube-api-access-x8x9b\") pod \"dnsmasq-dns-785d8bcb8c-fh8q6\" (UID: \"8214df67-a984-4a5b-bcbb-6b16ab42f5fe\") " pod="openstack/dnsmasq-dns-785d8bcb8c-fh8q6" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.823235 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.880324 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/08b8c363-a8c1-4dfe-8fd1-1af2d055a305-logs\") pod \"placement-db-sync-6vgh6\" (UID: \"08b8c363-a8c1-4dfe-8fd1-1af2d055a305\") " pod="openstack/placement-db-sync-6vgh6" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.880646 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/08b8c363-a8c1-4dfe-8fd1-1af2d055a305-logs\") pod \"placement-db-sync-6vgh6\" (UID: \"08b8c363-a8c1-4dfe-8fd1-1af2d055a305\") " pod="openstack/placement-db-sync-6vgh6" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.880989 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gkb9f\" (UniqueName: \"kubernetes.io/projected/08b8c363-a8c1-4dfe-8fd1-1af2d055a305-kube-api-access-gkb9f\") pod \"placement-db-sync-6vgh6\" (UID: \"08b8c363-a8c1-4dfe-8fd1-1af2d055a305\") " pod="openstack/placement-db-sync-6vgh6" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.881040 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08b8c363-a8c1-4dfe-8fd1-1af2d055a305-combined-ca-bundle\") pod \"placement-db-sync-6vgh6\" (UID: \"08b8c363-a8c1-4dfe-8fd1-1af2d055a305\") " pod="openstack/placement-db-sync-6vgh6" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.881453 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/08b8c363-a8c1-4dfe-8fd1-1af2d055a305-scripts\") pod \"placement-db-sync-6vgh6\" (UID: \"08b8c363-a8c1-4dfe-8fd1-1af2d055a305\") " pod="openstack/placement-db-sync-6vgh6" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.881507 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/08b8c363-a8c1-4dfe-8fd1-1af2d055a305-config-data\") pod \"placement-db-sync-6vgh6\" (UID: \"08b8c363-a8c1-4dfe-8fd1-1af2d055a305\") " pod="openstack/placement-db-sync-6vgh6" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.884140 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08b8c363-a8c1-4dfe-8fd1-1af2d055a305-combined-ca-bundle\") pod \"placement-db-sync-6vgh6\" (UID: \"08b8c363-a8c1-4dfe-8fd1-1af2d055a305\") " pod="openstack/placement-db-sync-6vgh6" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.884344 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/08b8c363-a8c1-4dfe-8fd1-1af2d055a305-scripts\") pod \"placement-db-sync-6vgh6\" (UID: \"08b8c363-a8c1-4dfe-8fd1-1af2d055a305\") " pod="openstack/placement-db-sync-6vgh6" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.885019 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/08b8c363-a8c1-4dfe-8fd1-1af2d055a305-config-data\") pod \"placement-db-sync-6vgh6\" (UID: \"08b8c363-a8c1-4dfe-8fd1-1af2d055a305\") " pod="openstack/placement-db-sync-6vgh6" Jan 21 11:26:21 crc kubenswrapper[4824]: I0121 11:26:21.896298 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gkb9f\" (UniqueName: \"kubernetes.io/projected/08b8c363-a8c1-4dfe-8fd1-1af2d055a305-kube-api-access-gkb9f\") pod \"placement-db-sync-6vgh6\" (UID: \"08b8c363-a8c1-4dfe-8fd1-1af2d055a305\") " pod="openstack/placement-db-sync-6vgh6" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.009255 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-785d8bcb8c-fh8q6" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.017508 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-6vgh6" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.118671 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-kb6p9"] Jan 21 11:26:22 crc kubenswrapper[4824]: W0121 11:26:22.143195 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbcfd76c3_d71d_4f02_af00_d01b113c4d2e.slice/crio-158d502ef9a1830d418ad9c0bd801b29d56915eafcb443ee997a5f301edbf91f WatchSource:0}: Error finding container 158d502ef9a1830d418ad9c0bd801b29d56915eafcb443ee997a5f301edbf91f: Status 404 returned error can't find the container with id 158d502ef9a1830d418ad9c0bd801b29d56915eafcb443ee997a5f301edbf91f Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.171726 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-sw5v2"] Jan 21 11:26:22 crc kubenswrapper[4824]: W0121 11:26:22.178132 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4f0bb2ed_9f0d_4d49_b224_8d05b5de0829.slice/crio-0307ac7b768d862977960b3a1fde7cfaed49319b93b9406e8b79d485ba5b5853 WatchSource:0}: Error finding container 0307ac7b768d862977960b3a1fde7cfaed49319b93b9406e8b79d485ba5b5853: Status 404 returned error can't find the container with id 0307ac7b768d862977960b3a1fde7cfaed49319b93b9406e8b79d485ba5b5853 Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.245792 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-sync-jrsn8"] Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.374122 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.379233 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-847c4cc679-kb6p9" event={"ID":"bcfd76c3-d71d-4f02-af00-d01b113c4d2e","Type":"ContainerStarted","Data":"158d502ef9a1830d418ad9c0bd801b29d56915eafcb443ee997a5f301edbf91f"} Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.382868 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-jrsn8" event={"ID":"cadf2d38-24f9-43ab-8e29-803d97a0ff06","Type":"ContainerStarted","Data":"f8ce50a0aaa511cfcd5c91e987379f42c2650248768a77a65a4355ed457a06c8"} Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.386078 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-sw5v2" event={"ID":"4f0bb2ed-9f0d-4d49-b224-8d05b5de0829","Type":"ContainerStarted","Data":"0307ac7b768d862977960b3a1fde7cfaed49319b93b9406e8b79d485ba5b5853"} Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.404934 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.406377 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.409239 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-4bzft" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.409435 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.409657 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.409787 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.417677 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.432474 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-ngq69"] Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.433753 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-ngq69" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.439498 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-8q76f" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.439751 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.490076 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.508365 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6a5a3f84-2320-4f8f-b8d4-d6ee9053f859-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"6a5a3f84-2320-4f8f-b8d4-d6ee9053f859\") " pod="openstack/glance-default-external-api-0" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.508558 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mvbdv\" (UniqueName: \"kubernetes.io/projected/6a5a3f84-2320-4f8f-b8d4-d6ee9053f859-kube-api-access-mvbdv\") pod \"glance-default-external-api-0\" (UID: \"6a5a3f84-2320-4f8f-b8d4-d6ee9053f859\") " pod="openstack/glance-default-external-api-0" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.508659 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"6a5a3f84-2320-4f8f-b8d4-d6ee9053f859\") " pod="openstack/glance-default-external-api-0" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.508755 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/262733b9-4179-49e3-aee9-b62197cc89ba-db-sync-config-data\") pod \"barbican-db-sync-ngq69\" (UID: \"262733b9-4179-49e3-aee9-b62197cc89ba\") " pod="openstack/barbican-db-sync-ngq69" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.508823 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6a5a3f84-2320-4f8f-b8d4-d6ee9053f859-logs\") pod \"glance-default-external-api-0\" (UID: \"6a5a3f84-2320-4f8f-b8d4-d6ee9053f859\") " pod="openstack/glance-default-external-api-0" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.508926 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-flpqh\" (UniqueName: \"kubernetes.io/projected/262733b9-4179-49e3-aee9-b62197cc89ba-kube-api-access-flpqh\") pod \"barbican-db-sync-ngq69\" (UID: \"262733b9-4179-49e3-aee9-b62197cc89ba\") " pod="openstack/barbican-db-sync-ngq69" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.509024 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6a5a3f84-2320-4f8f-b8d4-d6ee9053f859-config-data\") pod \"glance-default-external-api-0\" (UID: \"6a5a3f84-2320-4f8f-b8d4-d6ee9053f859\") " pod="openstack/glance-default-external-api-0" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.509189 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6a5a3f84-2320-4f8f-b8d4-d6ee9053f859-scripts\") pod \"glance-default-external-api-0\" (UID: \"6a5a3f84-2320-4f8f-b8d4-d6ee9053f859\") " pod="openstack/glance-default-external-api-0" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.509239 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a5a3f84-2320-4f8f-b8d4-d6ee9053f859-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"6a5a3f84-2320-4f8f-b8d4-d6ee9053f859\") " pod="openstack/glance-default-external-api-0" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.509282 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/262733b9-4179-49e3-aee9-b62197cc89ba-combined-ca-bundle\") pod \"barbican-db-sync-ngq69\" (UID: \"262733b9-4179-49e3-aee9-b62197cc89ba\") " pod="openstack/barbican-db-sync-ngq69" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.509318 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/6a5a3f84-2320-4f8f-b8d4-d6ee9053f859-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"6a5a3f84-2320-4f8f-b8d4-d6ee9053f859\") " pod="openstack/glance-default-external-api-0" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.526470 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.531084 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.531264 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.541245 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-ngq69"] Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.563211 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.573553 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-rlp2t"] Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.574444 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-rlp2t" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.576384 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.576906 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-2xjd5" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.577035 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.583814 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-rlp2t"] Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.591104 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-6vgh6"] Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.610294 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-fh8q6"] Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.610502 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7d0e8430-4bfe-4e37-8c22-502dd5444af4-logs\") pod \"glance-default-internal-api-0\" (UID: \"7d0e8430-4bfe-4e37-8c22-502dd5444af4\") " pod="openstack/glance-default-internal-api-0" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.610637 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0a6c0628-4522-4bb7-8a82-cc2e019eca2a-etc-machine-id\") pod \"cinder-db-sync-rlp2t\" (UID: \"0a6c0628-4522-4bb7-8a82-cc2e019eca2a\") " pod="openstack/cinder-db-sync-rlp2t" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.610782 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/0a6c0628-4522-4bb7-8a82-cc2e019eca2a-db-sync-config-data\") pod \"cinder-db-sync-rlp2t\" (UID: \"0a6c0628-4522-4bb7-8a82-cc2e019eca2a\") " pod="openstack/cinder-db-sync-rlp2t" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.610922 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6a5a3f84-2320-4f8f-b8d4-d6ee9053f859-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"6a5a3f84-2320-4f8f-b8d4-d6ee9053f859\") " pod="openstack/glance-default-external-api-0" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.611550 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mvbdv\" (UniqueName: \"kubernetes.io/projected/6a5a3f84-2320-4f8f-b8d4-d6ee9053f859-kube-api-access-mvbdv\") pod \"glance-default-external-api-0\" (UID: \"6a5a3f84-2320-4f8f-b8d4-d6ee9053f859\") " pod="openstack/glance-default-external-api-0" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.611700 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7d0e8430-4bfe-4e37-8c22-502dd5444af4-config-data\") pod \"glance-default-internal-api-0\" (UID: \"7d0e8430-4bfe-4e37-8c22-502dd5444af4\") " pod="openstack/glance-default-internal-api-0" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.611793 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-msmnx\" (UniqueName: \"kubernetes.io/projected/0a6c0628-4522-4bb7-8a82-cc2e019eca2a-kube-api-access-msmnx\") pod \"cinder-db-sync-rlp2t\" (UID: \"0a6c0628-4522-4bb7-8a82-cc2e019eca2a\") " pod="openstack/cinder-db-sync-rlp2t" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.611875 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"6a5a3f84-2320-4f8f-b8d4-d6ee9053f859\") " pod="openstack/glance-default-external-api-0" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.611971 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7d0e8430-4bfe-4e37-8c22-502dd5444af4-scripts\") pod \"glance-default-internal-api-0\" (UID: \"7d0e8430-4bfe-4e37-8c22-502dd5444af4\") " pod="openstack/glance-default-internal-api-0" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.612041 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/262733b9-4179-49e3-aee9-b62197cc89ba-db-sync-config-data\") pod \"barbican-db-sync-ngq69\" (UID: \"262733b9-4179-49e3-aee9-b62197cc89ba\") " pod="openstack/barbican-db-sync-ngq69" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.612161 4824 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"6a5a3f84-2320-4f8f-b8d4-d6ee9053f859\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/glance-default-external-api-0" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.612400 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6a5a3f84-2320-4f8f-b8d4-d6ee9053f859-logs\") pod \"glance-default-external-api-0\" (UID: \"6a5a3f84-2320-4f8f-b8d4-d6ee9053f859\") " pod="openstack/glance-default-external-api-0" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.612497 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0a6c0628-4522-4bb7-8a82-cc2e019eca2a-scripts\") pod \"cinder-db-sync-rlp2t\" (UID: \"0a6c0628-4522-4bb7-8a82-cc2e019eca2a\") " pod="openstack/cinder-db-sync-rlp2t" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.612574 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"7d0e8430-4bfe-4e37-8c22-502dd5444af4\") " pod="openstack/glance-default-internal-api-0" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.612636 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f4wc2\" (UniqueName: \"kubernetes.io/projected/7d0e8430-4bfe-4e37-8c22-502dd5444af4-kube-api-access-f4wc2\") pod \"glance-default-internal-api-0\" (UID: \"7d0e8430-4bfe-4e37-8c22-502dd5444af4\") " pod="openstack/glance-default-internal-api-0" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.612741 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-flpqh\" (UniqueName: \"kubernetes.io/projected/262733b9-4179-49e3-aee9-b62197cc89ba-kube-api-access-flpqh\") pod \"barbican-db-sync-ngq69\" (UID: \"262733b9-4179-49e3-aee9-b62197cc89ba\") " pod="openstack/barbican-db-sync-ngq69" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.612823 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6a5a3f84-2320-4f8f-b8d4-d6ee9053f859-config-data\") pod \"glance-default-external-api-0\" (UID: \"6a5a3f84-2320-4f8f-b8d4-d6ee9053f859\") " pod="openstack/glance-default-external-api-0" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.612915 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a6c0628-4522-4bb7-8a82-cc2e019eca2a-combined-ca-bundle\") pod \"cinder-db-sync-rlp2t\" (UID: \"0a6c0628-4522-4bb7-8a82-cc2e019eca2a\") " pod="openstack/cinder-db-sync-rlp2t" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.613021 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d0e8430-4bfe-4e37-8c22-502dd5444af4-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"7d0e8430-4bfe-4e37-8c22-502dd5444af4\") " pod="openstack/glance-default-internal-api-0" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.613104 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6a5a3f84-2320-4f8f-b8d4-d6ee9053f859-scripts\") pod \"glance-default-external-api-0\" (UID: \"6a5a3f84-2320-4f8f-b8d4-d6ee9053f859\") " pod="openstack/glance-default-external-api-0" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.613178 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a5a3f84-2320-4f8f-b8d4-d6ee9053f859-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"6a5a3f84-2320-4f8f-b8d4-d6ee9053f859\") " pod="openstack/glance-default-external-api-0" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.612934 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6a5a3f84-2320-4f8f-b8d4-d6ee9053f859-logs\") pod \"glance-default-external-api-0\" (UID: \"6a5a3f84-2320-4f8f-b8d4-d6ee9053f859\") " pod="openstack/glance-default-external-api-0" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.613318 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7d0e8430-4bfe-4e37-8c22-502dd5444af4-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"7d0e8430-4bfe-4e37-8c22-502dd5444af4\") " pod="openstack/glance-default-internal-api-0" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.613395 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/262733b9-4179-49e3-aee9-b62197cc89ba-combined-ca-bundle\") pod \"barbican-db-sync-ngq69\" (UID: \"262733b9-4179-49e3-aee9-b62197cc89ba\") " pod="openstack/barbican-db-sync-ngq69" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.613475 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7d0e8430-4bfe-4e37-8c22-502dd5444af4-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"7d0e8430-4bfe-4e37-8c22-502dd5444af4\") " pod="openstack/glance-default-internal-api-0" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.613558 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/6a5a3f84-2320-4f8f-b8d4-d6ee9053f859-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"6a5a3f84-2320-4f8f-b8d4-d6ee9053f859\") " pod="openstack/glance-default-external-api-0" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.613653 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a6c0628-4522-4bb7-8a82-cc2e019eca2a-config-data\") pod \"cinder-db-sync-rlp2t\" (UID: \"0a6c0628-4522-4bb7-8a82-cc2e019eca2a\") " pod="openstack/cinder-db-sync-rlp2t" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.615578 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/262733b9-4179-49e3-aee9-b62197cc89ba-db-sync-config-data\") pod \"barbican-db-sync-ngq69\" (UID: \"262733b9-4179-49e3-aee9-b62197cc89ba\") " pod="openstack/barbican-db-sync-ngq69" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.617308 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/6a5a3f84-2320-4f8f-b8d4-d6ee9053f859-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"6a5a3f84-2320-4f8f-b8d4-d6ee9053f859\") " pod="openstack/glance-default-external-api-0" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.618200 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/262733b9-4179-49e3-aee9-b62197cc89ba-combined-ca-bundle\") pod \"barbican-db-sync-ngq69\" (UID: \"262733b9-4179-49e3-aee9-b62197cc89ba\") " pod="openstack/barbican-db-sync-ngq69" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.618437 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6a5a3f84-2320-4f8f-b8d4-d6ee9053f859-scripts\") pod \"glance-default-external-api-0\" (UID: \"6a5a3f84-2320-4f8f-b8d4-d6ee9053f859\") " pod="openstack/glance-default-external-api-0" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.618514 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6a5a3f84-2320-4f8f-b8d4-d6ee9053f859-config-data\") pod \"glance-default-external-api-0\" (UID: \"6a5a3f84-2320-4f8f-b8d4-d6ee9053f859\") " pod="openstack/glance-default-external-api-0" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.623728 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a5a3f84-2320-4f8f-b8d4-d6ee9053f859-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"6a5a3f84-2320-4f8f-b8d4-d6ee9053f859\") " pod="openstack/glance-default-external-api-0" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.625561 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mvbdv\" (UniqueName: \"kubernetes.io/projected/6a5a3f84-2320-4f8f-b8d4-d6ee9053f859-kube-api-access-mvbdv\") pod \"glance-default-external-api-0\" (UID: \"6a5a3f84-2320-4f8f-b8d4-d6ee9053f859\") " pod="openstack/glance-default-external-api-0" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.632974 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6a5a3f84-2320-4f8f-b8d4-d6ee9053f859-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"6a5a3f84-2320-4f8f-b8d4-d6ee9053f859\") " pod="openstack/glance-default-external-api-0" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.633386 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-flpqh\" (UniqueName: \"kubernetes.io/projected/262733b9-4179-49e3-aee9-b62197cc89ba-kube-api-access-flpqh\") pod \"barbican-db-sync-ngq69\" (UID: \"262733b9-4179-49e3-aee9-b62197cc89ba\") " pod="openstack/barbican-db-sync-ngq69" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.635583 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"6a5a3f84-2320-4f8f-b8d4-d6ee9053f859\") " pod="openstack/glance-default-external-api-0" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.715147 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7d0e8430-4bfe-4e37-8c22-502dd5444af4-config-data\") pod \"glance-default-internal-api-0\" (UID: \"7d0e8430-4bfe-4e37-8c22-502dd5444af4\") " pod="openstack/glance-default-internal-api-0" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.715249 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-msmnx\" (UniqueName: \"kubernetes.io/projected/0a6c0628-4522-4bb7-8a82-cc2e019eca2a-kube-api-access-msmnx\") pod \"cinder-db-sync-rlp2t\" (UID: \"0a6c0628-4522-4bb7-8a82-cc2e019eca2a\") " pod="openstack/cinder-db-sync-rlp2t" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.715287 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7d0e8430-4bfe-4e37-8c22-502dd5444af4-scripts\") pod \"glance-default-internal-api-0\" (UID: \"7d0e8430-4bfe-4e37-8c22-502dd5444af4\") " pod="openstack/glance-default-internal-api-0" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.715318 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0a6c0628-4522-4bb7-8a82-cc2e019eca2a-scripts\") pod \"cinder-db-sync-rlp2t\" (UID: \"0a6c0628-4522-4bb7-8a82-cc2e019eca2a\") " pod="openstack/cinder-db-sync-rlp2t" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.715336 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"7d0e8430-4bfe-4e37-8c22-502dd5444af4\") " pod="openstack/glance-default-internal-api-0" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.715351 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f4wc2\" (UniqueName: \"kubernetes.io/projected/7d0e8430-4bfe-4e37-8c22-502dd5444af4-kube-api-access-f4wc2\") pod \"glance-default-internal-api-0\" (UID: \"7d0e8430-4bfe-4e37-8c22-502dd5444af4\") " pod="openstack/glance-default-internal-api-0" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.715401 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a6c0628-4522-4bb7-8a82-cc2e019eca2a-combined-ca-bundle\") pod \"cinder-db-sync-rlp2t\" (UID: \"0a6c0628-4522-4bb7-8a82-cc2e019eca2a\") " pod="openstack/cinder-db-sync-rlp2t" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.715425 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d0e8430-4bfe-4e37-8c22-502dd5444af4-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"7d0e8430-4bfe-4e37-8c22-502dd5444af4\") " pod="openstack/glance-default-internal-api-0" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.715464 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7d0e8430-4bfe-4e37-8c22-502dd5444af4-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"7d0e8430-4bfe-4e37-8c22-502dd5444af4\") " pod="openstack/glance-default-internal-api-0" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.715498 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7d0e8430-4bfe-4e37-8c22-502dd5444af4-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"7d0e8430-4bfe-4e37-8c22-502dd5444af4\") " pod="openstack/glance-default-internal-api-0" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.715532 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a6c0628-4522-4bb7-8a82-cc2e019eca2a-config-data\") pod \"cinder-db-sync-rlp2t\" (UID: \"0a6c0628-4522-4bb7-8a82-cc2e019eca2a\") " pod="openstack/cinder-db-sync-rlp2t" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.715556 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7d0e8430-4bfe-4e37-8c22-502dd5444af4-logs\") pod \"glance-default-internal-api-0\" (UID: \"7d0e8430-4bfe-4e37-8c22-502dd5444af4\") " pod="openstack/glance-default-internal-api-0" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.715583 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0a6c0628-4522-4bb7-8a82-cc2e019eca2a-etc-machine-id\") pod \"cinder-db-sync-rlp2t\" (UID: \"0a6c0628-4522-4bb7-8a82-cc2e019eca2a\") " pod="openstack/cinder-db-sync-rlp2t" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.715598 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/0a6c0628-4522-4bb7-8a82-cc2e019eca2a-db-sync-config-data\") pod \"cinder-db-sync-rlp2t\" (UID: \"0a6c0628-4522-4bb7-8a82-cc2e019eca2a\") " pod="openstack/cinder-db-sync-rlp2t" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.715649 4824 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"7d0e8430-4bfe-4e37-8c22-502dd5444af4\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/glance-default-internal-api-0" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.716035 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0a6c0628-4522-4bb7-8a82-cc2e019eca2a-etc-machine-id\") pod \"cinder-db-sync-rlp2t\" (UID: \"0a6c0628-4522-4bb7-8a82-cc2e019eca2a\") " pod="openstack/cinder-db-sync-rlp2t" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.716229 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7d0e8430-4bfe-4e37-8c22-502dd5444af4-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"7d0e8430-4bfe-4e37-8c22-502dd5444af4\") " pod="openstack/glance-default-internal-api-0" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.716289 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7d0e8430-4bfe-4e37-8c22-502dd5444af4-logs\") pod \"glance-default-internal-api-0\" (UID: \"7d0e8430-4bfe-4e37-8c22-502dd5444af4\") " pod="openstack/glance-default-internal-api-0" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.721422 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/0a6c0628-4522-4bb7-8a82-cc2e019eca2a-db-sync-config-data\") pod \"cinder-db-sync-rlp2t\" (UID: \"0a6c0628-4522-4bb7-8a82-cc2e019eca2a\") " pod="openstack/cinder-db-sync-rlp2t" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.721619 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d0e8430-4bfe-4e37-8c22-502dd5444af4-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"7d0e8430-4bfe-4e37-8c22-502dd5444af4\") " pod="openstack/glance-default-internal-api-0" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.721849 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a6c0628-4522-4bb7-8a82-cc2e019eca2a-combined-ca-bundle\") pod \"cinder-db-sync-rlp2t\" (UID: \"0a6c0628-4522-4bb7-8a82-cc2e019eca2a\") " pod="openstack/cinder-db-sync-rlp2t" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.722617 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0a6c0628-4522-4bb7-8a82-cc2e019eca2a-scripts\") pod \"cinder-db-sync-rlp2t\" (UID: \"0a6c0628-4522-4bb7-8a82-cc2e019eca2a\") " pod="openstack/cinder-db-sync-rlp2t" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.722704 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7d0e8430-4bfe-4e37-8c22-502dd5444af4-scripts\") pod \"glance-default-internal-api-0\" (UID: \"7d0e8430-4bfe-4e37-8c22-502dd5444af4\") " pod="openstack/glance-default-internal-api-0" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.724142 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7d0e8430-4bfe-4e37-8c22-502dd5444af4-config-data\") pod \"glance-default-internal-api-0\" (UID: \"7d0e8430-4bfe-4e37-8c22-502dd5444af4\") " pod="openstack/glance-default-internal-api-0" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.725291 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a6c0628-4522-4bb7-8a82-cc2e019eca2a-config-data\") pod \"cinder-db-sync-rlp2t\" (UID: \"0a6c0628-4522-4bb7-8a82-cc2e019eca2a\") " pod="openstack/cinder-db-sync-rlp2t" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.727574 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7d0e8430-4bfe-4e37-8c22-502dd5444af4-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"7d0e8430-4bfe-4e37-8c22-502dd5444af4\") " pod="openstack/glance-default-internal-api-0" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.734076 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-msmnx\" (UniqueName: \"kubernetes.io/projected/0a6c0628-4522-4bb7-8a82-cc2e019eca2a-kube-api-access-msmnx\") pod \"cinder-db-sync-rlp2t\" (UID: \"0a6c0628-4522-4bb7-8a82-cc2e019eca2a\") " pod="openstack/cinder-db-sync-rlp2t" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.734619 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f4wc2\" (UniqueName: \"kubernetes.io/projected/7d0e8430-4bfe-4e37-8c22-502dd5444af4-kube-api-access-f4wc2\") pod \"glance-default-internal-api-0\" (UID: \"7d0e8430-4bfe-4e37-8c22-502dd5444af4\") " pod="openstack/glance-default-internal-api-0" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.749511 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-rlp2t" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.754712 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-t2txg"] Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.756000 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-t2txg" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.758324 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.758374 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.758387 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-49xnt" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.762481 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"7d0e8430-4bfe-4e37-8c22-502dd5444af4\") " pod="openstack/glance-default-internal-api-0" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.773737 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-t2txg"] Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.780583 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.816838 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/a19b6730-b7aa-4319-8127-c32d7874a471-config\") pod \"neutron-db-sync-t2txg\" (UID: \"a19b6730-b7aa-4319-8127-c32d7874a471\") " pod="openstack/neutron-db-sync-t2txg" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.817437 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dpjkl\" (UniqueName: \"kubernetes.io/projected/a19b6730-b7aa-4319-8127-c32d7874a471-kube-api-access-dpjkl\") pod \"neutron-db-sync-t2txg\" (UID: \"a19b6730-b7aa-4319-8127-c32d7874a471\") " pod="openstack/neutron-db-sync-t2txg" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.817466 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a19b6730-b7aa-4319-8127-c32d7874a471-combined-ca-bundle\") pod \"neutron-db-sync-t2txg\" (UID: \"a19b6730-b7aa-4319-8127-c32d7874a471\") " pod="openstack/neutron-db-sync-t2txg" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.850298 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-ngq69" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.919435 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dpjkl\" (UniqueName: \"kubernetes.io/projected/a19b6730-b7aa-4319-8127-c32d7874a471-kube-api-access-dpjkl\") pod \"neutron-db-sync-t2txg\" (UID: \"a19b6730-b7aa-4319-8127-c32d7874a471\") " pod="openstack/neutron-db-sync-t2txg" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.919473 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a19b6730-b7aa-4319-8127-c32d7874a471-combined-ca-bundle\") pod \"neutron-db-sync-t2txg\" (UID: \"a19b6730-b7aa-4319-8127-c32d7874a471\") " pod="openstack/neutron-db-sync-t2txg" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.919535 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/a19b6730-b7aa-4319-8127-c32d7874a471-config\") pod \"neutron-db-sync-t2txg\" (UID: \"a19b6730-b7aa-4319-8127-c32d7874a471\") " pod="openstack/neutron-db-sync-t2txg" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.925519 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a19b6730-b7aa-4319-8127-c32d7874a471-combined-ca-bundle\") pod \"neutron-db-sync-t2txg\" (UID: \"a19b6730-b7aa-4319-8127-c32d7874a471\") " pod="openstack/neutron-db-sync-t2txg" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.929570 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/a19b6730-b7aa-4319-8127-c32d7874a471-config\") pod \"neutron-db-sync-t2txg\" (UID: \"a19b6730-b7aa-4319-8127-c32d7874a471\") " pod="openstack/neutron-db-sync-t2txg" Jan 21 11:26:22 crc kubenswrapper[4824]: I0121 11:26:22.936794 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dpjkl\" (UniqueName: \"kubernetes.io/projected/a19b6730-b7aa-4319-8127-c32d7874a471-kube-api-access-dpjkl\") pod \"neutron-db-sync-t2txg\" (UID: \"a19b6730-b7aa-4319-8127-c32d7874a471\") " pod="openstack/neutron-db-sync-t2txg" Jan 21 11:26:23 crc kubenswrapper[4824]: I0121 11:26:23.021724 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Jan 21 11:26:23 crc kubenswrapper[4824]: I0121 11:26:23.043666 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Jan 21 11:26:23 crc kubenswrapper[4824]: I0121 11:26:23.070049 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Jan 21 11:26:23 crc kubenswrapper[4824]: I0121 11:26:23.080422 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-t2txg" Jan 21 11:26:23 crc kubenswrapper[4824]: I0121 11:26:23.166434 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Jan 21 11:26:23 crc kubenswrapper[4824]: I0121 11:26:23.193758 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-rlp2t"] Jan 21 11:26:23 crc kubenswrapper[4824]: I0121 11:26:23.335632 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Jan 21 11:26:23 crc kubenswrapper[4824]: W0121 11:26:23.372019 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod262733b9_4179_49e3_aee9_b62197cc89ba.slice/crio-2bab663bc55719655bbe7aad12a3a6a58c0374bdaecaed83d199e55c2baa7ddf WatchSource:0}: Error finding container 2bab663bc55719655bbe7aad12a3a6a58c0374bdaecaed83d199e55c2baa7ddf: Status 404 returned error can't find the container with id 2bab663bc55719655bbe7aad12a3a6a58c0374bdaecaed83d199e55c2baa7ddf Jan 21 11:26:23 crc kubenswrapper[4824]: I0121 11:26:23.375484 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-ngq69"] Jan 21 11:26:23 crc kubenswrapper[4824]: I0121 11:26:23.405166 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-rlp2t" event={"ID":"0a6c0628-4522-4bb7-8a82-cc2e019eca2a","Type":"ContainerStarted","Data":"877238fc09eb99e400629467da99c2ba4a6e53c0afc16677b1308d3c52849ae9"} Jan 21 11:26:23 crc kubenswrapper[4824]: I0121 11:26:23.409442 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-sw5v2" event={"ID":"4f0bb2ed-9f0d-4d49-b224-8d05b5de0829","Type":"ContainerStarted","Data":"68285af994e4fe82e817010ea6bff330d5655f091788a58ca5b521d0e79564f1"} Jan 21 11:26:23 crc kubenswrapper[4824]: I0121 11:26:23.411134 4824 generic.go:334] "Generic (PLEG): container finished" podID="8214df67-a984-4a5b-bcbb-6b16ab42f5fe" containerID="7af315bc544e0ed568b3db042720a7c28e2f59b1fb260d2208c816705728d7f5" exitCode=0 Jan 21 11:26:23 crc kubenswrapper[4824]: I0121 11:26:23.411192 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-fh8q6" event={"ID":"8214df67-a984-4a5b-bcbb-6b16ab42f5fe","Type":"ContainerDied","Data":"7af315bc544e0ed568b3db042720a7c28e2f59b1fb260d2208c816705728d7f5"} Jan 21 11:26:23 crc kubenswrapper[4824]: I0121 11:26:23.411216 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-fh8q6" event={"ID":"8214df67-a984-4a5b-bcbb-6b16ab42f5fe","Type":"ContainerStarted","Data":"a7ed7bee402dc3db5565fe550b060e37c1771f55a94528a7a39dbd2741c1ec27"} Jan 21 11:26:23 crc kubenswrapper[4824]: I0121 11:26:23.415321 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-6vgh6" event={"ID":"08b8c363-a8c1-4dfe-8fd1-1af2d055a305","Type":"ContainerStarted","Data":"97721ccfe8abdf2324794ac67debd67a8fc655d3d7ab538bf2cb4b35e35f8386"} Jan 21 11:26:23 crc kubenswrapper[4824]: I0121 11:26:23.418241 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bbefda93-dc56-43c5-b131-56af96ca832b","Type":"ContainerStarted","Data":"9fa11b86f012b31aa61fb0528e706362f80104db47ab25f797420eb04f92a3fa"} Jan 21 11:26:23 crc kubenswrapper[4824]: I0121 11:26:23.420215 4824 generic.go:334] "Generic (PLEG): container finished" podID="bcfd76c3-d71d-4f02-af00-d01b113c4d2e" containerID="27ee2d89193499812ebe436242e9b3d7756adae2afdbb8a226ae95bcc2f9a4f1" exitCode=0 Jan 21 11:26:23 crc kubenswrapper[4824]: I0121 11:26:23.420824 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-847c4cc679-kb6p9" event={"ID":"bcfd76c3-d71d-4f02-af00-d01b113c4d2e","Type":"ContainerDied","Data":"27ee2d89193499812ebe436242e9b3d7756adae2afdbb8a226ae95bcc2f9a4f1"} Jan 21 11:26:23 crc kubenswrapper[4824]: I0121 11:26:23.423555 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-ngq69" event={"ID":"262733b9-4179-49e3-aee9-b62197cc89ba","Type":"ContainerStarted","Data":"2bab663bc55719655bbe7aad12a3a6a58c0374bdaecaed83d199e55c2baa7ddf"} Jan 21 11:26:23 crc kubenswrapper[4824]: I0121 11:26:23.429631 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-sw5v2" podStartSLOduration=2.429616409 podStartE2EDuration="2.429616409s" podCreationTimestamp="2026-01-21 11:26:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:26:23.423702974 +0000 UTC m=+925.716732267" watchObservedRunningTime="2026-01-21 11:26:23.429616409 +0000 UTC m=+925.722645701" Jan 21 11:26:23 crc kubenswrapper[4824]: I0121 11:26:23.432058 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"6a5a3f84-2320-4f8f-b8d4-d6ee9053f859","Type":"ContainerStarted","Data":"cb87ef02c610bd346ad76ef8c74796d1ec1bf79d6d471b5705c3e5bb8a69ec3b"} Jan 21 11:26:23 crc kubenswrapper[4824]: I0121 11:26:23.650670 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Jan 21 11:26:23 crc kubenswrapper[4824]: I0121 11:26:23.657807 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-t2txg"] Jan 21 11:26:23 crc kubenswrapper[4824]: W0121 11:26:23.665795 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7d0e8430_4bfe_4e37_8c22_502dd5444af4.slice/crio-51a42c61ed30de7289f1aa596c3516c76d7926804f43324d5132c6f93e7e5c66 WatchSource:0}: Error finding container 51a42c61ed30de7289f1aa596c3516c76d7926804f43324d5132c6f93e7e5c66: Status 404 returned error can't find the container with id 51a42c61ed30de7289f1aa596c3516c76d7926804f43324d5132c6f93e7e5c66 Jan 21 11:26:23 crc kubenswrapper[4824]: W0121 11:26:23.684757 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda19b6730_b7aa_4319_8127_c32d7874a471.slice/crio-936d5e5c16f699412bd3905d0974b8b47a243fdf030015ab8663f5d0120a7209 WatchSource:0}: Error finding container 936d5e5c16f699412bd3905d0974b8b47a243fdf030015ab8663f5d0120a7209: Status 404 returned error can't find the container with id 936d5e5c16f699412bd3905d0974b8b47a243fdf030015ab8663f5d0120a7209 Jan 21 11:26:23 crc kubenswrapper[4824]: I0121 11:26:23.741668 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-847c4cc679-kb6p9" Jan 21 11:26:23 crc kubenswrapper[4824]: I0121 11:26:23.844568 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h5zvt\" (UniqueName: \"kubernetes.io/projected/bcfd76c3-d71d-4f02-af00-d01b113c4d2e-kube-api-access-h5zvt\") pod \"bcfd76c3-d71d-4f02-af00-d01b113c4d2e\" (UID: \"bcfd76c3-d71d-4f02-af00-d01b113c4d2e\") " Jan 21 11:26:23 crc kubenswrapper[4824]: I0121 11:26:23.844652 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bcfd76c3-d71d-4f02-af00-d01b113c4d2e-ovsdbserver-nb\") pod \"bcfd76c3-d71d-4f02-af00-d01b113c4d2e\" (UID: \"bcfd76c3-d71d-4f02-af00-d01b113c4d2e\") " Jan 21 11:26:23 crc kubenswrapper[4824]: I0121 11:26:23.844768 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bcfd76c3-d71d-4f02-af00-d01b113c4d2e-config\") pod \"bcfd76c3-d71d-4f02-af00-d01b113c4d2e\" (UID: \"bcfd76c3-d71d-4f02-af00-d01b113c4d2e\") " Jan 21 11:26:23 crc kubenswrapper[4824]: I0121 11:26:23.844868 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bcfd76c3-d71d-4f02-af00-d01b113c4d2e-ovsdbserver-sb\") pod \"bcfd76c3-d71d-4f02-af00-d01b113c4d2e\" (UID: \"bcfd76c3-d71d-4f02-af00-d01b113c4d2e\") " Jan 21 11:26:23 crc kubenswrapper[4824]: I0121 11:26:23.844936 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bcfd76c3-d71d-4f02-af00-d01b113c4d2e-dns-swift-storage-0\") pod \"bcfd76c3-d71d-4f02-af00-d01b113c4d2e\" (UID: \"bcfd76c3-d71d-4f02-af00-d01b113c4d2e\") " Jan 21 11:26:23 crc kubenswrapper[4824]: I0121 11:26:23.844997 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bcfd76c3-d71d-4f02-af00-d01b113c4d2e-dns-svc\") pod \"bcfd76c3-d71d-4f02-af00-d01b113c4d2e\" (UID: \"bcfd76c3-d71d-4f02-af00-d01b113c4d2e\") " Jan 21 11:26:23 crc kubenswrapper[4824]: I0121 11:26:23.853761 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bcfd76c3-d71d-4f02-af00-d01b113c4d2e-kube-api-access-h5zvt" (OuterVolumeSpecName: "kube-api-access-h5zvt") pod "bcfd76c3-d71d-4f02-af00-d01b113c4d2e" (UID: "bcfd76c3-d71d-4f02-af00-d01b113c4d2e"). InnerVolumeSpecName "kube-api-access-h5zvt". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:26:23 crc kubenswrapper[4824]: I0121 11:26:23.864619 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bcfd76c3-d71d-4f02-af00-d01b113c4d2e-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "bcfd76c3-d71d-4f02-af00-d01b113c4d2e" (UID: "bcfd76c3-d71d-4f02-af00-d01b113c4d2e"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:26:23 crc kubenswrapper[4824]: I0121 11:26:23.868313 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bcfd76c3-d71d-4f02-af00-d01b113c4d2e-config" (OuterVolumeSpecName: "config") pod "bcfd76c3-d71d-4f02-af00-d01b113c4d2e" (UID: "bcfd76c3-d71d-4f02-af00-d01b113c4d2e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:26:23 crc kubenswrapper[4824]: I0121 11:26:23.868487 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bcfd76c3-d71d-4f02-af00-d01b113c4d2e-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "bcfd76c3-d71d-4f02-af00-d01b113c4d2e" (UID: "bcfd76c3-d71d-4f02-af00-d01b113c4d2e"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:26:23 crc kubenswrapper[4824]: I0121 11:26:23.871602 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bcfd76c3-d71d-4f02-af00-d01b113c4d2e-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "bcfd76c3-d71d-4f02-af00-d01b113c4d2e" (UID: "bcfd76c3-d71d-4f02-af00-d01b113c4d2e"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:26:23 crc kubenswrapper[4824]: I0121 11:26:23.878639 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bcfd76c3-d71d-4f02-af00-d01b113c4d2e-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "bcfd76c3-d71d-4f02-af00-d01b113c4d2e" (UID: "bcfd76c3-d71d-4f02-af00-d01b113c4d2e"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:26:23 crc kubenswrapper[4824]: I0121 11:26:23.947196 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h5zvt\" (UniqueName: \"kubernetes.io/projected/bcfd76c3-d71d-4f02-af00-d01b113c4d2e-kube-api-access-h5zvt\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:23 crc kubenswrapper[4824]: I0121 11:26:23.947221 4824 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bcfd76c3-d71d-4f02-af00-d01b113c4d2e-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:23 crc kubenswrapper[4824]: I0121 11:26:23.947231 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bcfd76c3-d71d-4f02-af00-d01b113c4d2e-config\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:23 crc kubenswrapper[4824]: I0121 11:26:23.947239 4824 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bcfd76c3-d71d-4f02-af00-d01b113c4d2e-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:23 crc kubenswrapper[4824]: I0121 11:26:23.947247 4824 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bcfd76c3-d71d-4f02-af00-d01b113c4d2e-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:23 crc kubenswrapper[4824]: I0121 11:26:23.947254 4824 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bcfd76c3-d71d-4f02-af00-d01b113c4d2e-dns-svc\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:25 crc kubenswrapper[4824]: I0121 11:26:24.451600 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"6a5a3f84-2320-4f8f-b8d4-d6ee9053f859","Type":"ContainerStarted","Data":"b624461c7fd817bff45a1813fc59b08421416d9686db5ebe66efe4049b30534b"} Jan 21 11:26:25 crc kubenswrapper[4824]: I0121 11:26:24.454417 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"7d0e8430-4bfe-4e37-8c22-502dd5444af4","Type":"ContainerStarted","Data":"859f2e0dceb5ba0b5dc20dc8542580aaaceec1b327c7c16135166be30bdad2a1"} Jan 21 11:26:25 crc kubenswrapper[4824]: I0121 11:26:24.454453 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"7d0e8430-4bfe-4e37-8c22-502dd5444af4","Type":"ContainerStarted","Data":"51a42c61ed30de7289f1aa596c3516c76d7926804f43324d5132c6f93e7e5c66"} Jan 21 11:26:25 crc kubenswrapper[4824]: I0121 11:26:24.458200 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-fh8q6" event={"ID":"8214df67-a984-4a5b-bcbb-6b16ab42f5fe","Type":"ContainerStarted","Data":"56442b913a4db71125dc75408dc50b059e8519640a2cc312a0c8277e0b1aefab"} Jan 21 11:26:25 crc kubenswrapper[4824]: I0121 11:26:24.458995 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-785d8bcb8c-fh8q6" Jan 21 11:26:25 crc kubenswrapper[4824]: I0121 11:26:24.461309 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-t2txg" event={"ID":"a19b6730-b7aa-4319-8127-c32d7874a471","Type":"ContainerStarted","Data":"49abf3af76b6411b5212e8cfa13e8c4145a6700e1d310d3281c5d847c1a1f457"} Jan 21 11:26:25 crc kubenswrapper[4824]: I0121 11:26:24.461337 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-t2txg" event={"ID":"a19b6730-b7aa-4319-8127-c32d7874a471","Type":"ContainerStarted","Data":"936d5e5c16f699412bd3905d0974b8b47a243fdf030015ab8663f5d0120a7209"} Jan 21 11:26:25 crc kubenswrapper[4824]: I0121 11:26:24.475759 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-785d8bcb8c-fh8q6" podStartSLOduration=3.475748282 podStartE2EDuration="3.475748282s" podCreationTimestamp="2026-01-21 11:26:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:26:24.475139695 +0000 UTC m=+926.768168987" watchObservedRunningTime="2026-01-21 11:26:24.475748282 +0000 UTC m=+926.768777574" Jan 21 11:26:25 crc kubenswrapper[4824]: I0121 11:26:24.484034 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-847c4cc679-kb6p9" Jan 21 11:26:25 crc kubenswrapper[4824]: I0121 11:26:24.484064 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-847c4cc679-kb6p9" event={"ID":"bcfd76c3-d71d-4f02-af00-d01b113c4d2e","Type":"ContainerDied","Data":"158d502ef9a1830d418ad9c0bd801b29d56915eafcb443ee997a5f301edbf91f"} Jan 21 11:26:25 crc kubenswrapper[4824]: I0121 11:26:24.484089 4824 scope.go:117] "RemoveContainer" containerID="27ee2d89193499812ebe436242e9b3d7756adae2afdbb8a226ae95bcc2f9a4f1" Jan 21 11:26:25 crc kubenswrapper[4824]: I0121 11:26:24.491021 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-t2txg" podStartSLOduration=2.491005481 podStartE2EDuration="2.491005481s" podCreationTimestamp="2026-01-21 11:26:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:26:24.485381483 +0000 UTC m=+926.778410775" watchObservedRunningTime="2026-01-21 11:26:24.491005481 +0000 UTC m=+926.784034773" Jan 21 11:26:25 crc kubenswrapper[4824]: I0121 11:26:24.624024 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-kb6p9"] Jan 21 11:26:25 crc kubenswrapper[4824]: I0121 11:26:24.633308 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-kb6p9"] Jan 21 11:26:25 crc kubenswrapper[4824]: I0121 11:26:25.500370 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"6a5a3f84-2320-4f8f-b8d4-d6ee9053f859","Type":"ContainerStarted","Data":"7f3d59f931487b81345a89a9579c3ee29598f0123e3fc5337b9d33bd8702f80f"} Jan 21 11:26:25 crc kubenswrapper[4824]: I0121 11:26:25.500455 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="6a5a3f84-2320-4f8f-b8d4-d6ee9053f859" containerName="glance-log" containerID="cri-o://b624461c7fd817bff45a1813fc59b08421416d9686db5ebe66efe4049b30534b" gracePeriod=30 Jan 21 11:26:25 crc kubenswrapper[4824]: I0121 11:26:25.500517 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="6a5a3f84-2320-4f8f-b8d4-d6ee9053f859" containerName="glance-httpd" containerID="cri-o://7f3d59f931487b81345a89a9579c3ee29598f0123e3fc5337b9d33bd8702f80f" gracePeriod=30 Jan 21 11:26:25 crc kubenswrapper[4824]: I0121 11:26:25.510900 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"7d0e8430-4bfe-4e37-8c22-502dd5444af4","Type":"ContainerStarted","Data":"0f7acbd726575b92b89853489927b20157e8c557e9675f9935d803c1e04cda43"} Jan 21 11:26:25 crc kubenswrapper[4824]: I0121 11:26:25.510969 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="7d0e8430-4bfe-4e37-8c22-502dd5444af4" containerName="glance-log" containerID="cri-o://859f2e0dceb5ba0b5dc20dc8542580aaaceec1b327c7c16135166be30bdad2a1" gracePeriod=30 Jan 21 11:26:25 crc kubenswrapper[4824]: I0121 11:26:25.510986 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="7d0e8430-4bfe-4e37-8c22-502dd5444af4" containerName="glance-httpd" containerID="cri-o://0f7acbd726575b92b89853489927b20157e8c557e9675f9935d803c1e04cda43" gracePeriod=30 Jan 21 11:26:25 crc kubenswrapper[4824]: I0121 11:26:25.540258 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=4.540241329 podStartE2EDuration="4.540241329s" podCreationTimestamp="2026-01-21 11:26:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:26:25.539648792 +0000 UTC m=+927.832678084" watchObservedRunningTime="2026-01-21 11:26:25.540241329 +0000 UTC m=+927.833270621" Jan 21 11:26:25 crc kubenswrapper[4824]: I0121 11:26:25.542642 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=4.542629685 podStartE2EDuration="4.542629685s" podCreationTimestamp="2026-01-21 11:26:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:26:25.521819491 +0000 UTC m=+927.814848783" watchObservedRunningTime="2026-01-21 11:26:25.542629685 +0000 UTC m=+927.835658976" Jan 21 11:26:26 crc kubenswrapper[4824]: I0121 11:26:26.073537 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bcfd76c3-d71d-4f02-af00-d01b113c4d2e" path="/var/lib/kubelet/pods/bcfd76c3-d71d-4f02-af00-d01b113c4d2e/volumes" Jan 21 11:26:26 crc kubenswrapper[4824]: I0121 11:26:26.524558 4824 generic.go:334] "Generic (PLEG): container finished" podID="6a5a3f84-2320-4f8f-b8d4-d6ee9053f859" containerID="7f3d59f931487b81345a89a9579c3ee29598f0123e3fc5337b9d33bd8702f80f" exitCode=0 Jan 21 11:26:26 crc kubenswrapper[4824]: I0121 11:26:26.524772 4824 generic.go:334] "Generic (PLEG): container finished" podID="6a5a3f84-2320-4f8f-b8d4-d6ee9053f859" containerID="b624461c7fd817bff45a1813fc59b08421416d9686db5ebe66efe4049b30534b" exitCode=143 Jan 21 11:26:26 crc kubenswrapper[4824]: I0121 11:26:26.524635 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"6a5a3f84-2320-4f8f-b8d4-d6ee9053f859","Type":"ContainerDied","Data":"7f3d59f931487b81345a89a9579c3ee29598f0123e3fc5337b9d33bd8702f80f"} Jan 21 11:26:26 crc kubenswrapper[4824]: I0121 11:26:26.524837 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"6a5a3f84-2320-4f8f-b8d4-d6ee9053f859","Type":"ContainerDied","Data":"b624461c7fd817bff45a1813fc59b08421416d9686db5ebe66efe4049b30534b"} Jan 21 11:26:26 crc kubenswrapper[4824]: I0121 11:26:26.526324 4824 generic.go:334] "Generic (PLEG): container finished" podID="4f0bb2ed-9f0d-4d49-b224-8d05b5de0829" containerID="68285af994e4fe82e817010ea6bff330d5655f091788a58ca5b521d0e79564f1" exitCode=0 Jan 21 11:26:26 crc kubenswrapper[4824]: I0121 11:26:26.526389 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-sw5v2" event={"ID":"4f0bb2ed-9f0d-4d49-b224-8d05b5de0829","Type":"ContainerDied","Data":"68285af994e4fe82e817010ea6bff330d5655f091788a58ca5b521d0e79564f1"} Jan 21 11:26:26 crc kubenswrapper[4824]: I0121 11:26:26.530569 4824 generic.go:334] "Generic (PLEG): container finished" podID="7d0e8430-4bfe-4e37-8c22-502dd5444af4" containerID="0f7acbd726575b92b89853489927b20157e8c557e9675f9935d803c1e04cda43" exitCode=0 Jan 21 11:26:26 crc kubenswrapper[4824]: I0121 11:26:26.530610 4824 generic.go:334] "Generic (PLEG): container finished" podID="7d0e8430-4bfe-4e37-8c22-502dd5444af4" containerID="859f2e0dceb5ba0b5dc20dc8542580aaaceec1b327c7c16135166be30bdad2a1" exitCode=143 Jan 21 11:26:26 crc kubenswrapper[4824]: I0121 11:26:26.531514 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"7d0e8430-4bfe-4e37-8c22-502dd5444af4","Type":"ContainerDied","Data":"0f7acbd726575b92b89853489927b20157e8c557e9675f9935d803c1e04cda43"} Jan 21 11:26:26 crc kubenswrapper[4824]: I0121 11:26:26.531557 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"7d0e8430-4bfe-4e37-8c22-502dd5444af4","Type":"ContainerDied","Data":"859f2e0dceb5ba0b5dc20dc8542580aaaceec1b327c7c16135166be30bdad2a1"} Jan 21 11:26:30 crc kubenswrapper[4824]: I0121 11:26:30.574221 4824 generic.go:334] "Generic (PLEG): container finished" podID="a19b6730-b7aa-4319-8127-c32d7874a471" containerID="49abf3af76b6411b5212e8cfa13e8c4145a6700e1d310d3281c5d847c1a1f457" exitCode=0 Jan 21 11:26:30 crc kubenswrapper[4824]: I0121 11:26:30.574256 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-t2txg" event={"ID":"a19b6730-b7aa-4319-8127-c32d7874a471","Type":"ContainerDied","Data":"49abf3af76b6411b5212e8cfa13e8c4145a6700e1d310d3281c5d847c1a1f457"} Jan 21 11:26:32 crc kubenswrapper[4824]: I0121 11:26:32.011075 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-785d8bcb8c-fh8q6" Jan 21 11:26:32 crc kubenswrapper[4824]: I0121 11:26:32.066527 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-sj79w"] Jan 21 11:26:32 crc kubenswrapper[4824]: I0121 11:26:32.066732 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-74f6bcbc87-sj79w" podUID="c4a81903-5e1f-4d9c-bacf-60639069db18" containerName="dnsmasq-dns" containerID="cri-o://862050906c3faeb783c3cf1a55874b97964bf3e73dcab5972c6e49f1ea08c530" gracePeriod=10 Jan 21 11:26:32 crc kubenswrapper[4824]: I0121 11:26:32.593378 4824 generic.go:334] "Generic (PLEG): container finished" podID="c4a81903-5e1f-4d9c-bacf-60639069db18" containerID="862050906c3faeb783c3cf1a55874b97964bf3e73dcab5972c6e49f1ea08c530" exitCode=0 Jan 21 11:26:32 crc kubenswrapper[4824]: I0121 11:26:32.593408 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74f6bcbc87-sj79w" event={"ID":"c4a81903-5e1f-4d9c-bacf-60639069db18","Type":"ContainerDied","Data":"862050906c3faeb783c3cf1a55874b97964bf3e73dcab5972c6e49f1ea08c530"} Jan 21 11:26:34 crc kubenswrapper[4824]: I0121 11:26:34.498094 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-sw5v2" Jan 21 11:26:34 crc kubenswrapper[4824]: I0121 11:26:34.531564 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4f0bb2ed-9f0d-4d49-b224-8d05b5de0829-config-data\") pod \"4f0bb2ed-9f0d-4d49-b224-8d05b5de0829\" (UID: \"4f0bb2ed-9f0d-4d49-b224-8d05b5de0829\") " Jan 21 11:26:34 crc kubenswrapper[4824]: I0121 11:26:34.531607 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f0bb2ed-9f0d-4d49-b224-8d05b5de0829-combined-ca-bundle\") pod \"4f0bb2ed-9f0d-4d49-b224-8d05b5de0829\" (UID: \"4f0bb2ed-9f0d-4d49-b224-8d05b5de0829\") " Jan 21 11:26:34 crc kubenswrapper[4824]: I0121 11:26:34.531648 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ztr5x\" (UniqueName: \"kubernetes.io/projected/4f0bb2ed-9f0d-4d49-b224-8d05b5de0829-kube-api-access-ztr5x\") pod \"4f0bb2ed-9f0d-4d49-b224-8d05b5de0829\" (UID: \"4f0bb2ed-9f0d-4d49-b224-8d05b5de0829\") " Jan 21 11:26:34 crc kubenswrapper[4824]: I0121 11:26:34.531683 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/4f0bb2ed-9f0d-4d49-b224-8d05b5de0829-credential-keys\") pod \"4f0bb2ed-9f0d-4d49-b224-8d05b5de0829\" (UID: \"4f0bb2ed-9f0d-4d49-b224-8d05b5de0829\") " Jan 21 11:26:34 crc kubenswrapper[4824]: I0121 11:26:34.531731 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4f0bb2ed-9f0d-4d49-b224-8d05b5de0829-scripts\") pod \"4f0bb2ed-9f0d-4d49-b224-8d05b5de0829\" (UID: \"4f0bb2ed-9f0d-4d49-b224-8d05b5de0829\") " Jan 21 11:26:34 crc kubenswrapper[4824]: I0121 11:26:34.531767 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/4f0bb2ed-9f0d-4d49-b224-8d05b5de0829-fernet-keys\") pod \"4f0bb2ed-9f0d-4d49-b224-8d05b5de0829\" (UID: \"4f0bb2ed-9f0d-4d49-b224-8d05b5de0829\") " Jan 21 11:26:34 crc kubenswrapper[4824]: I0121 11:26:34.542466 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4f0bb2ed-9f0d-4d49-b224-8d05b5de0829-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "4f0bb2ed-9f0d-4d49-b224-8d05b5de0829" (UID: "4f0bb2ed-9f0d-4d49-b224-8d05b5de0829"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:26:34 crc kubenswrapper[4824]: I0121 11:26:34.542495 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4f0bb2ed-9f0d-4d49-b224-8d05b5de0829-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "4f0bb2ed-9f0d-4d49-b224-8d05b5de0829" (UID: "4f0bb2ed-9f0d-4d49-b224-8d05b5de0829"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:26:34 crc kubenswrapper[4824]: I0121 11:26:34.542570 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4f0bb2ed-9f0d-4d49-b224-8d05b5de0829-scripts" (OuterVolumeSpecName: "scripts") pod "4f0bb2ed-9f0d-4d49-b224-8d05b5de0829" (UID: "4f0bb2ed-9f0d-4d49-b224-8d05b5de0829"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:26:34 crc kubenswrapper[4824]: I0121 11:26:34.542619 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4f0bb2ed-9f0d-4d49-b224-8d05b5de0829-kube-api-access-ztr5x" (OuterVolumeSpecName: "kube-api-access-ztr5x") pod "4f0bb2ed-9f0d-4d49-b224-8d05b5de0829" (UID: "4f0bb2ed-9f0d-4d49-b224-8d05b5de0829"). InnerVolumeSpecName "kube-api-access-ztr5x". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:26:34 crc kubenswrapper[4824]: I0121 11:26:34.550203 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4f0bb2ed-9f0d-4d49-b224-8d05b5de0829-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4f0bb2ed-9f0d-4d49-b224-8d05b5de0829" (UID: "4f0bb2ed-9f0d-4d49-b224-8d05b5de0829"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:26:34 crc kubenswrapper[4824]: I0121 11:26:34.551130 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4f0bb2ed-9f0d-4d49-b224-8d05b5de0829-config-data" (OuterVolumeSpecName: "config-data") pod "4f0bb2ed-9f0d-4d49-b224-8d05b5de0829" (UID: "4f0bb2ed-9f0d-4d49-b224-8d05b5de0829"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:26:34 crc kubenswrapper[4824]: I0121 11:26:34.607734 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-sw5v2" event={"ID":"4f0bb2ed-9f0d-4d49-b224-8d05b5de0829","Type":"ContainerDied","Data":"0307ac7b768d862977960b3a1fde7cfaed49319b93b9406e8b79d485ba5b5853"} Jan 21 11:26:34 crc kubenswrapper[4824]: I0121 11:26:34.607769 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0307ac7b768d862977960b3a1fde7cfaed49319b93b9406e8b79d485ba5b5853" Jan 21 11:26:34 crc kubenswrapper[4824]: I0121 11:26:34.607776 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-sw5v2" Jan 21 11:26:34 crc kubenswrapper[4824]: I0121 11:26:34.633036 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4f0bb2ed-9f0d-4d49-b224-8d05b5de0829-config-data\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:34 crc kubenswrapper[4824]: I0121 11:26:34.633062 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f0bb2ed-9f0d-4d49-b224-8d05b5de0829-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:34 crc kubenswrapper[4824]: I0121 11:26:34.633071 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ztr5x\" (UniqueName: \"kubernetes.io/projected/4f0bb2ed-9f0d-4d49-b224-8d05b5de0829-kube-api-access-ztr5x\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:34 crc kubenswrapper[4824]: I0121 11:26:34.633079 4824 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/4f0bb2ed-9f0d-4d49-b224-8d05b5de0829-credential-keys\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:34 crc kubenswrapper[4824]: I0121 11:26:34.633088 4824 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4f0bb2ed-9f0d-4d49-b224-8d05b5de0829-scripts\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:34 crc kubenswrapper[4824]: I0121 11:26:34.633097 4824 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/4f0bb2ed-9f0d-4d49-b224-8d05b5de0829-fernet-keys\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:35 crc kubenswrapper[4824]: I0121 11:26:35.563748 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-sw5v2"] Jan 21 11:26:35 crc kubenswrapper[4824]: I0121 11:26:35.569012 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-sw5v2"] Jan 21 11:26:35 crc kubenswrapper[4824]: I0121 11:26:35.665191 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-5l2gg"] Jan 21 11:26:35 crc kubenswrapper[4824]: E0121 11:26:35.665885 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bcfd76c3-d71d-4f02-af00-d01b113c4d2e" containerName="init" Jan 21 11:26:35 crc kubenswrapper[4824]: I0121 11:26:35.665904 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="bcfd76c3-d71d-4f02-af00-d01b113c4d2e" containerName="init" Jan 21 11:26:35 crc kubenswrapper[4824]: E0121 11:26:35.665950 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4f0bb2ed-9f0d-4d49-b224-8d05b5de0829" containerName="keystone-bootstrap" Jan 21 11:26:35 crc kubenswrapper[4824]: I0121 11:26:35.665976 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="4f0bb2ed-9f0d-4d49-b224-8d05b5de0829" containerName="keystone-bootstrap" Jan 21 11:26:35 crc kubenswrapper[4824]: I0121 11:26:35.666191 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="4f0bb2ed-9f0d-4d49-b224-8d05b5de0829" containerName="keystone-bootstrap" Jan 21 11:26:35 crc kubenswrapper[4824]: I0121 11:26:35.666225 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="bcfd76c3-d71d-4f02-af00-d01b113c4d2e" containerName="init" Jan 21 11:26:35 crc kubenswrapper[4824]: I0121 11:26:35.666876 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-5l2gg" Jan 21 11:26:35 crc kubenswrapper[4824]: I0121 11:26:35.668411 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Jan 21 11:26:35 crc kubenswrapper[4824]: I0121 11:26:35.668853 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-qxhzv" Jan 21 11:26:35 crc kubenswrapper[4824]: I0121 11:26:35.668914 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Jan 21 11:26:35 crc kubenswrapper[4824]: I0121 11:26:35.669007 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Jan 21 11:26:35 crc kubenswrapper[4824]: I0121 11:26:35.669738 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Jan 21 11:26:35 crc kubenswrapper[4824]: I0121 11:26:35.671981 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-5l2gg"] Jan 21 11:26:35 crc kubenswrapper[4824]: I0121 11:26:35.719065 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-74f6bcbc87-sj79w" podUID="c4a81903-5e1f-4d9c-bacf-60639069db18" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.126:5353: connect: connection refused" Jan 21 11:26:35 crc kubenswrapper[4824]: I0121 11:26:35.748515 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7e75011-d633-4ccc-951c-d019cb2100f9-combined-ca-bundle\") pod \"keystone-bootstrap-5l2gg\" (UID: \"a7e75011-d633-4ccc-951c-d019cb2100f9\") " pod="openstack/keystone-bootstrap-5l2gg" Jan 21 11:26:35 crc kubenswrapper[4824]: I0121 11:26:35.748585 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8lj6r\" (UniqueName: \"kubernetes.io/projected/a7e75011-d633-4ccc-951c-d019cb2100f9-kube-api-access-8lj6r\") pod \"keystone-bootstrap-5l2gg\" (UID: \"a7e75011-d633-4ccc-951c-d019cb2100f9\") " pod="openstack/keystone-bootstrap-5l2gg" Jan 21 11:26:35 crc kubenswrapper[4824]: I0121 11:26:35.748616 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a7e75011-d633-4ccc-951c-d019cb2100f9-config-data\") pod \"keystone-bootstrap-5l2gg\" (UID: \"a7e75011-d633-4ccc-951c-d019cb2100f9\") " pod="openstack/keystone-bootstrap-5l2gg" Jan 21 11:26:35 crc kubenswrapper[4824]: I0121 11:26:35.748647 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a7e75011-d633-4ccc-951c-d019cb2100f9-fernet-keys\") pod \"keystone-bootstrap-5l2gg\" (UID: \"a7e75011-d633-4ccc-951c-d019cb2100f9\") " pod="openstack/keystone-bootstrap-5l2gg" Jan 21 11:26:35 crc kubenswrapper[4824]: I0121 11:26:35.748660 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/a7e75011-d633-4ccc-951c-d019cb2100f9-credential-keys\") pod \"keystone-bootstrap-5l2gg\" (UID: \"a7e75011-d633-4ccc-951c-d019cb2100f9\") " pod="openstack/keystone-bootstrap-5l2gg" Jan 21 11:26:35 crc kubenswrapper[4824]: I0121 11:26:35.748817 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a7e75011-d633-4ccc-951c-d019cb2100f9-scripts\") pod \"keystone-bootstrap-5l2gg\" (UID: \"a7e75011-d633-4ccc-951c-d019cb2100f9\") " pod="openstack/keystone-bootstrap-5l2gg" Jan 21 11:26:35 crc kubenswrapper[4824]: I0121 11:26:35.849994 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7e75011-d633-4ccc-951c-d019cb2100f9-combined-ca-bundle\") pod \"keystone-bootstrap-5l2gg\" (UID: \"a7e75011-d633-4ccc-951c-d019cb2100f9\") " pod="openstack/keystone-bootstrap-5l2gg" Jan 21 11:26:35 crc kubenswrapper[4824]: I0121 11:26:35.850064 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8lj6r\" (UniqueName: \"kubernetes.io/projected/a7e75011-d633-4ccc-951c-d019cb2100f9-kube-api-access-8lj6r\") pod \"keystone-bootstrap-5l2gg\" (UID: \"a7e75011-d633-4ccc-951c-d019cb2100f9\") " pod="openstack/keystone-bootstrap-5l2gg" Jan 21 11:26:35 crc kubenswrapper[4824]: I0121 11:26:35.850098 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a7e75011-d633-4ccc-951c-d019cb2100f9-config-data\") pod \"keystone-bootstrap-5l2gg\" (UID: \"a7e75011-d633-4ccc-951c-d019cb2100f9\") " pod="openstack/keystone-bootstrap-5l2gg" Jan 21 11:26:35 crc kubenswrapper[4824]: I0121 11:26:35.850130 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a7e75011-d633-4ccc-951c-d019cb2100f9-fernet-keys\") pod \"keystone-bootstrap-5l2gg\" (UID: \"a7e75011-d633-4ccc-951c-d019cb2100f9\") " pod="openstack/keystone-bootstrap-5l2gg" Jan 21 11:26:35 crc kubenswrapper[4824]: I0121 11:26:35.850144 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/a7e75011-d633-4ccc-951c-d019cb2100f9-credential-keys\") pod \"keystone-bootstrap-5l2gg\" (UID: \"a7e75011-d633-4ccc-951c-d019cb2100f9\") " pod="openstack/keystone-bootstrap-5l2gg" Jan 21 11:26:35 crc kubenswrapper[4824]: I0121 11:26:35.850210 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a7e75011-d633-4ccc-951c-d019cb2100f9-scripts\") pod \"keystone-bootstrap-5l2gg\" (UID: \"a7e75011-d633-4ccc-951c-d019cb2100f9\") " pod="openstack/keystone-bootstrap-5l2gg" Jan 21 11:26:35 crc kubenswrapper[4824]: I0121 11:26:35.853402 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a7e75011-d633-4ccc-951c-d019cb2100f9-scripts\") pod \"keystone-bootstrap-5l2gg\" (UID: \"a7e75011-d633-4ccc-951c-d019cb2100f9\") " pod="openstack/keystone-bootstrap-5l2gg" Jan 21 11:26:35 crc kubenswrapper[4824]: I0121 11:26:35.853461 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/a7e75011-d633-4ccc-951c-d019cb2100f9-credential-keys\") pod \"keystone-bootstrap-5l2gg\" (UID: \"a7e75011-d633-4ccc-951c-d019cb2100f9\") " pod="openstack/keystone-bootstrap-5l2gg" Jan 21 11:26:35 crc kubenswrapper[4824]: I0121 11:26:35.853944 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a7e75011-d633-4ccc-951c-d019cb2100f9-config-data\") pod \"keystone-bootstrap-5l2gg\" (UID: \"a7e75011-d633-4ccc-951c-d019cb2100f9\") " pod="openstack/keystone-bootstrap-5l2gg" Jan 21 11:26:35 crc kubenswrapper[4824]: I0121 11:26:35.854254 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7e75011-d633-4ccc-951c-d019cb2100f9-combined-ca-bundle\") pod \"keystone-bootstrap-5l2gg\" (UID: \"a7e75011-d633-4ccc-951c-d019cb2100f9\") " pod="openstack/keystone-bootstrap-5l2gg" Jan 21 11:26:35 crc kubenswrapper[4824]: I0121 11:26:35.855039 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a7e75011-d633-4ccc-951c-d019cb2100f9-fernet-keys\") pod \"keystone-bootstrap-5l2gg\" (UID: \"a7e75011-d633-4ccc-951c-d019cb2100f9\") " pod="openstack/keystone-bootstrap-5l2gg" Jan 21 11:26:35 crc kubenswrapper[4824]: I0121 11:26:35.863272 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8lj6r\" (UniqueName: \"kubernetes.io/projected/a7e75011-d633-4ccc-951c-d019cb2100f9-kube-api-access-8lj6r\") pod \"keystone-bootstrap-5l2gg\" (UID: \"a7e75011-d633-4ccc-951c-d019cb2100f9\") " pod="openstack/keystone-bootstrap-5l2gg" Jan 21 11:26:35 crc kubenswrapper[4824]: I0121 11:26:35.985949 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-5l2gg" Jan 21 11:26:36 crc kubenswrapper[4824]: I0121 11:26:36.060377 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4f0bb2ed-9f0d-4d49-b224-8d05b5de0829" path="/var/lib/kubelet/pods/4f0bb2ed-9f0d-4d49-b224-8d05b5de0829/volumes" Jan 21 11:26:39 crc kubenswrapper[4824]: I0121 11:26:39.935992 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Jan 21 11:26:39 crc kubenswrapper[4824]: I0121 11:26:39.954283 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Jan 21 11:26:39 crc kubenswrapper[4824]: I0121 11:26:39.957372 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-t2txg" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.105850 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7d0e8430-4bfe-4e37-8c22-502dd5444af4-logs\") pod \"7d0e8430-4bfe-4e37-8c22-502dd5444af4\" (UID: \"7d0e8430-4bfe-4e37-8c22-502dd5444af4\") " Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.105893 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dpjkl\" (UniqueName: \"kubernetes.io/projected/a19b6730-b7aa-4319-8127-c32d7874a471-kube-api-access-dpjkl\") pod \"a19b6730-b7aa-4319-8127-c32d7874a471\" (UID: \"a19b6730-b7aa-4319-8127-c32d7874a471\") " Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.105911 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d0e8430-4bfe-4e37-8c22-502dd5444af4-combined-ca-bundle\") pod \"7d0e8430-4bfe-4e37-8c22-502dd5444af4\" (UID: \"7d0e8430-4bfe-4e37-8c22-502dd5444af4\") " Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.105949 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f4wc2\" (UniqueName: \"kubernetes.io/projected/7d0e8430-4bfe-4e37-8c22-502dd5444af4-kube-api-access-f4wc2\") pod \"7d0e8430-4bfe-4e37-8c22-502dd5444af4\" (UID: \"7d0e8430-4bfe-4e37-8c22-502dd5444af4\") " Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.105987 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a19b6730-b7aa-4319-8127-c32d7874a471-combined-ca-bundle\") pod \"a19b6730-b7aa-4319-8127-c32d7874a471\" (UID: \"a19b6730-b7aa-4319-8127-c32d7874a471\") " Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.106017 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"6a5a3f84-2320-4f8f-b8d4-d6ee9053f859\" (UID: \"6a5a3f84-2320-4f8f-b8d4-d6ee9053f859\") " Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.106034 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6a5a3f84-2320-4f8f-b8d4-d6ee9053f859-logs\") pod \"6a5a3f84-2320-4f8f-b8d4-d6ee9053f859\" (UID: \"6a5a3f84-2320-4f8f-b8d4-d6ee9053f859\") " Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.106058 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6a5a3f84-2320-4f8f-b8d4-d6ee9053f859-public-tls-certs\") pod \"6a5a3f84-2320-4f8f-b8d4-d6ee9053f859\" (UID: \"6a5a3f84-2320-4f8f-b8d4-d6ee9053f859\") " Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.106095 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"7d0e8430-4bfe-4e37-8c22-502dd5444af4\" (UID: \"7d0e8430-4bfe-4e37-8c22-502dd5444af4\") " Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.106118 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6a5a3f84-2320-4f8f-b8d4-d6ee9053f859-config-data\") pod \"6a5a3f84-2320-4f8f-b8d4-d6ee9053f859\" (UID: \"6a5a3f84-2320-4f8f-b8d4-d6ee9053f859\") " Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.106157 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/a19b6730-b7aa-4319-8127-c32d7874a471-config\") pod \"a19b6730-b7aa-4319-8127-c32d7874a471\" (UID: \"a19b6730-b7aa-4319-8127-c32d7874a471\") " Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.106177 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/6a5a3f84-2320-4f8f-b8d4-d6ee9053f859-httpd-run\") pod \"6a5a3f84-2320-4f8f-b8d4-d6ee9053f859\" (UID: \"6a5a3f84-2320-4f8f-b8d4-d6ee9053f859\") " Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.106645 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7d0e8430-4bfe-4e37-8c22-502dd5444af4-internal-tls-certs\") pod \"7d0e8430-4bfe-4e37-8c22-502dd5444af4\" (UID: \"7d0e8430-4bfe-4e37-8c22-502dd5444af4\") " Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.106694 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mvbdv\" (UniqueName: \"kubernetes.io/projected/6a5a3f84-2320-4f8f-b8d4-d6ee9053f859-kube-api-access-mvbdv\") pod \"6a5a3f84-2320-4f8f-b8d4-d6ee9053f859\" (UID: \"6a5a3f84-2320-4f8f-b8d4-d6ee9053f859\") " Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.106733 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a5a3f84-2320-4f8f-b8d4-d6ee9053f859-combined-ca-bundle\") pod \"6a5a3f84-2320-4f8f-b8d4-d6ee9053f859\" (UID: \"6a5a3f84-2320-4f8f-b8d4-d6ee9053f859\") " Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.106771 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7d0e8430-4bfe-4e37-8c22-502dd5444af4-config-data\") pod \"7d0e8430-4bfe-4e37-8c22-502dd5444af4\" (UID: \"7d0e8430-4bfe-4e37-8c22-502dd5444af4\") " Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.106776 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7d0e8430-4bfe-4e37-8c22-502dd5444af4-logs" (OuterVolumeSpecName: "logs") pod "7d0e8430-4bfe-4e37-8c22-502dd5444af4" (UID: "7d0e8430-4bfe-4e37-8c22-502dd5444af4"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.106792 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6a5a3f84-2320-4f8f-b8d4-d6ee9053f859-scripts\") pod \"6a5a3f84-2320-4f8f-b8d4-d6ee9053f859\" (UID: \"6a5a3f84-2320-4f8f-b8d4-d6ee9053f859\") " Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.106810 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7d0e8430-4bfe-4e37-8c22-502dd5444af4-scripts\") pod \"7d0e8430-4bfe-4e37-8c22-502dd5444af4\" (UID: \"7d0e8430-4bfe-4e37-8c22-502dd5444af4\") " Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.106811 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6a5a3f84-2320-4f8f-b8d4-d6ee9053f859-logs" (OuterVolumeSpecName: "logs") pod "6a5a3f84-2320-4f8f-b8d4-d6ee9053f859" (UID: "6a5a3f84-2320-4f8f-b8d4-d6ee9053f859"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.106824 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7d0e8430-4bfe-4e37-8c22-502dd5444af4-httpd-run\") pod \"7d0e8430-4bfe-4e37-8c22-502dd5444af4\" (UID: \"7d0e8430-4bfe-4e37-8c22-502dd5444af4\") " Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.107142 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6a5a3f84-2320-4f8f-b8d4-d6ee9053f859-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "6a5a3f84-2320-4f8f-b8d4-d6ee9053f859" (UID: "6a5a3f84-2320-4f8f-b8d4-d6ee9053f859"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.107369 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7d0e8430-4bfe-4e37-8c22-502dd5444af4-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "7d0e8430-4bfe-4e37-8c22-502dd5444af4" (UID: "7d0e8430-4bfe-4e37-8c22-502dd5444af4"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.108445 4824 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7d0e8430-4bfe-4e37-8c22-502dd5444af4-httpd-run\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.108463 4824 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7d0e8430-4bfe-4e37-8c22-502dd5444af4-logs\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.108473 4824 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6a5a3f84-2320-4f8f-b8d4-d6ee9053f859-logs\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.108480 4824 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/6a5a3f84-2320-4f8f-b8d4-d6ee9053f859-httpd-run\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.110665 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6a5a3f84-2320-4f8f-b8d4-d6ee9053f859-kube-api-access-mvbdv" (OuterVolumeSpecName: "kube-api-access-mvbdv") pod "6a5a3f84-2320-4f8f-b8d4-d6ee9053f859" (UID: "6a5a3f84-2320-4f8f-b8d4-d6ee9053f859"). InnerVolumeSpecName "kube-api-access-mvbdv". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.112439 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage06-crc" (OuterVolumeSpecName: "glance") pod "6a5a3f84-2320-4f8f-b8d4-d6ee9053f859" (UID: "6a5a3f84-2320-4f8f-b8d4-d6ee9053f859"). InnerVolumeSpecName "local-storage06-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.112543 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7d0e8430-4bfe-4e37-8c22-502dd5444af4-kube-api-access-f4wc2" (OuterVolumeSpecName: "kube-api-access-f4wc2") pod "7d0e8430-4bfe-4e37-8c22-502dd5444af4" (UID: "7d0e8430-4bfe-4e37-8c22-502dd5444af4"). InnerVolumeSpecName "kube-api-access-f4wc2". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.112836 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage07-crc" (OuterVolumeSpecName: "glance") pod "7d0e8430-4bfe-4e37-8c22-502dd5444af4" (UID: "7d0e8430-4bfe-4e37-8c22-502dd5444af4"). InnerVolumeSpecName "local-storage07-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.115250 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7d0e8430-4bfe-4e37-8c22-502dd5444af4-scripts" (OuterVolumeSpecName: "scripts") pod "7d0e8430-4bfe-4e37-8c22-502dd5444af4" (UID: "7d0e8430-4bfe-4e37-8c22-502dd5444af4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.115323 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a19b6730-b7aa-4319-8127-c32d7874a471-kube-api-access-dpjkl" (OuterVolumeSpecName: "kube-api-access-dpjkl") pod "a19b6730-b7aa-4319-8127-c32d7874a471" (UID: "a19b6730-b7aa-4319-8127-c32d7874a471"). InnerVolumeSpecName "kube-api-access-dpjkl". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.115256 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6a5a3f84-2320-4f8f-b8d4-d6ee9053f859-scripts" (OuterVolumeSpecName: "scripts") pod "6a5a3f84-2320-4f8f-b8d4-d6ee9053f859" (UID: "6a5a3f84-2320-4f8f-b8d4-d6ee9053f859"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.132211 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6a5a3f84-2320-4f8f-b8d4-d6ee9053f859-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6a5a3f84-2320-4f8f-b8d4-d6ee9053f859" (UID: "6a5a3f84-2320-4f8f-b8d4-d6ee9053f859"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.134905 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7d0e8430-4bfe-4e37-8c22-502dd5444af4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7d0e8430-4bfe-4e37-8c22-502dd5444af4" (UID: "7d0e8430-4bfe-4e37-8c22-502dd5444af4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.143106 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a19b6730-b7aa-4319-8127-c32d7874a471-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a19b6730-b7aa-4319-8127-c32d7874a471" (UID: "a19b6730-b7aa-4319-8127-c32d7874a471"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.145896 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7d0e8430-4bfe-4e37-8c22-502dd5444af4-config-data" (OuterVolumeSpecName: "config-data") pod "7d0e8430-4bfe-4e37-8c22-502dd5444af4" (UID: "7d0e8430-4bfe-4e37-8c22-502dd5444af4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.146316 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a19b6730-b7aa-4319-8127-c32d7874a471-config" (OuterVolumeSpecName: "config") pod "a19b6730-b7aa-4319-8127-c32d7874a471" (UID: "a19b6730-b7aa-4319-8127-c32d7874a471"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.149099 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6a5a3f84-2320-4f8f-b8d4-d6ee9053f859-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "6a5a3f84-2320-4f8f-b8d4-d6ee9053f859" (UID: "6a5a3f84-2320-4f8f-b8d4-d6ee9053f859"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.153915 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7d0e8430-4bfe-4e37-8c22-502dd5444af4-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "7d0e8430-4bfe-4e37-8c22-502dd5444af4" (UID: "7d0e8430-4bfe-4e37-8c22-502dd5444af4"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.155443 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6a5a3f84-2320-4f8f-b8d4-d6ee9053f859-config-data" (OuterVolumeSpecName: "config-data") pod "6a5a3f84-2320-4f8f-b8d4-d6ee9053f859" (UID: "6a5a3f84-2320-4f8f-b8d4-d6ee9053f859"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.210391 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a5a3f84-2320-4f8f-b8d4-d6ee9053f859-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.210417 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7d0e8430-4bfe-4e37-8c22-502dd5444af4-config-data\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.210426 4824 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6a5a3f84-2320-4f8f-b8d4-d6ee9053f859-scripts\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.210434 4824 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7d0e8430-4bfe-4e37-8c22-502dd5444af4-scripts\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.210442 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dpjkl\" (UniqueName: \"kubernetes.io/projected/a19b6730-b7aa-4319-8127-c32d7874a471-kube-api-access-dpjkl\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.210452 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d0e8430-4bfe-4e37-8c22-502dd5444af4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.210460 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f4wc2\" (UniqueName: \"kubernetes.io/projected/7d0e8430-4bfe-4e37-8c22-502dd5444af4-kube-api-access-f4wc2\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.210467 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a19b6730-b7aa-4319-8127-c32d7874a471-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.210494 4824 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" " Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.210502 4824 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6a5a3f84-2320-4f8f-b8d4-d6ee9053f859-public-tls-certs\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.210514 4824 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" " Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.210521 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6a5a3f84-2320-4f8f-b8d4-d6ee9053f859-config-data\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.210529 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/a19b6730-b7aa-4319-8127-c32d7874a471-config\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.210536 4824 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7d0e8430-4bfe-4e37-8c22-502dd5444af4-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.210543 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mvbdv\" (UniqueName: \"kubernetes.io/projected/6a5a3f84-2320-4f8f-b8d4-d6ee9053f859-kube-api-access-mvbdv\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.226442 4824 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage06-crc" (UniqueName: "kubernetes.io/local-volume/local-storage06-crc") on node "crc" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.231270 4824 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage07-crc" (UniqueName: "kubernetes.io/local-volume/local-storage07-crc") on node "crc" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.311672 4824 reconciler_common.go:293] "Volume detached for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.311712 4824 reconciler_common.go:293] "Volume detached for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:40 crc kubenswrapper[4824]: E0121 11:26:40.321123 4824 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified" Jan 21 11:26:40 crc kubenswrapper[4824]: E0121 11:26:40.321249 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:barbican-db-sync,Image:quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified,Command:[/bin/bash],Args:[-c barbican-manage db upgrade],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/barbican/barbican.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-flpqh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42403,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42403,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-db-sync-ngq69_openstack(262733b9-4179-49e3-aee9-b62197cc89ba): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Jan 21 11:26:40 crc kubenswrapper[4824]: E0121 11:26:40.322760 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/barbican-db-sync-ngq69" podUID="262733b9-4179-49e3-aee9-b62197cc89ba" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.644899 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"7d0e8430-4bfe-4e37-8c22-502dd5444af4","Type":"ContainerDied","Data":"51a42c61ed30de7289f1aa596c3516c76d7926804f43324d5132c6f93e7e5c66"} Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.644921 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.644945 4824 scope.go:117] "RemoveContainer" containerID="0f7acbd726575b92b89853489927b20157e8c557e9675f9935d803c1e04cda43" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.646254 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-t2txg" event={"ID":"a19b6730-b7aa-4319-8127-c32d7874a471","Type":"ContainerDied","Data":"936d5e5c16f699412bd3905d0974b8b47a243fdf030015ab8663f5d0120a7209"} Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.646275 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="936d5e5c16f699412bd3905d0974b8b47a243fdf030015ab8663f5d0120a7209" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.646334 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-t2txg" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.653071 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.653870 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"6a5a3f84-2320-4f8f-b8d4-d6ee9053f859","Type":"ContainerDied","Data":"cb87ef02c610bd346ad76ef8c74796d1ec1bf79d6d471b5705c3e5bb8a69ec3b"} Jan 21 11:26:40 crc kubenswrapper[4824]: E0121 11:26:40.654420 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified\\\"\"" pod="openstack/barbican-db-sync-ngq69" podUID="262733b9-4179-49e3-aee9-b62197cc89ba" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.689970 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.699905 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.707332 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.712884 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.722683 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Jan 21 11:26:40 crc kubenswrapper[4824]: E0121 11:26:40.723064 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a19b6730-b7aa-4319-8127-c32d7874a471" containerName="neutron-db-sync" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.723078 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="a19b6730-b7aa-4319-8127-c32d7874a471" containerName="neutron-db-sync" Jan 21 11:26:40 crc kubenswrapper[4824]: E0121 11:26:40.723092 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d0e8430-4bfe-4e37-8c22-502dd5444af4" containerName="glance-httpd" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.723097 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d0e8430-4bfe-4e37-8c22-502dd5444af4" containerName="glance-httpd" Jan 21 11:26:40 crc kubenswrapper[4824]: E0121 11:26:40.723113 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d0e8430-4bfe-4e37-8c22-502dd5444af4" containerName="glance-log" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.723118 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d0e8430-4bfe-4e37-8c22-502dd5444af4" containerName="glance-log" Jan 21 11:26:40 crc kubenswrapper[4824]: E0121 11:26:40.723130 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a5a3f84-2320-4f8f-b8d4-d6ee9053f859" containerName="glance-log" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.723135 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a5a3f84-2320-4f8f-b8d4-d6ee9053f859" containerName="glance-log" Jan 21 11:26:40 crc kubenswrapper[4824]: E0121 11:26:40.723153 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a5a3f84-2320-4f8f-b8d4-d6ee9053f859" containerName="glance-httpd" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.723158 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a5a3f84-2320-4f8f-b8d4-d6ee9053f859" containerName="glance-httpd" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.723364 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="7d0e8430-4bfe-4e37-8c22-502dd5444af4" containerName="glance-httpd" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.723378 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="7d0e8430-4bfe-4e37-8c22-502dd5444af4" containerName="glance-log" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.723387 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="6a5a3f84-2320-4f8f-b8d4-d6ee9053f859" containerName="glance-httpd" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.723398 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="a19b6730-b7aa-4319-8127-c32d7874a471" containerName="neutron-db-sync" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.723405 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="6a5a3f84-2320-4f8f-b8d4-d6ee9053f859" containerName="glance-log" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.727155 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.728200 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.729216 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.732511 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.732743 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.732881 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.733024 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-4bzft" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.733387 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.733474 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.739767 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.746390 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.820057 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/46f0fa4b-479a-42c5-b9b2-3f30a079be48-logs\") pod \"glance-default-internal-api-0\" (UID: \"46f0fa4b-479a-42c5-b9b2-3f30a079be48\") " pod="openstack/glance-default-internal-api-0" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.820111 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/46f0fa4b-479a-42c5-b9b2-3f30a079be48-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"46f0fa4b-479a-42c5-b9b2-3f30a079be48\") " pod="openstack/glance-default-internal-api-0" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.820183 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/46f0fa4b-479a-42c5-b9b2-3f30a079be48-config-data\") pod \"glance-default-internal-api-0\" (UID: \"46f0fa4b-479a-42c5-b9b2-3f30a079be48\") " pod="openstack/glance-default-internal-api-0" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.820210 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"46f0fa4b-479a-42c5-b9b2-3f30a079be48\") " pod="openstack/glance-default-internal-api-0" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.820252 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/46f0fa4b-479a-42c5-b9b2-3f30a079be48-scripts\") pod \"glance-default-internal-api-0\" (UID: \"46f0fa4b-479a-42c5-b9b2-3f30a079be48\") " pod="openstack/glance-default-internal-api-0" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.820275 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46f0fa4b-479a-42c5-b9b2-3f30a079be48-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"46f0fa4b-479a-42c5-b9b2-3f30a079be48\") " pod="openstack/glance-default-internal-api-0" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.820389 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v8grh\" (UniqueName: \"kubernetes.io/projected/46f0fa4b-479a-42c5-b9b2-3f30a079be48-kube-api-access-v8grh\") pod \"glance-default-internal-api-0\" (UID: \"46f0fa4b-479a-42c5-b9b2-3f30a079be48\") " pod="openstack/glance-default-internal-api-0" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.820425 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/46f0fa4b-479a-42c5-b9b2-3f30a079be48-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"46f0fa4b-479a-42c5-b9b2-3f30a079be48\") " pod="openstack/glance-default-internal-api-0" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.921596 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/12776f5e-c372-41bd-91fe-ba7d0915050a-logs\") pod \"glance-default-external-api-0\" (UID: \"12776f5e-c372-41bd-91fe-ba7d0915050a\") " pod="openstack/glance-default-external-api-0" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.922012 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/12776f5e-c372-41bd-91fe-ba7d0915050a-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"12776f5e-c372-41bd-91fe-ba7d0915050a\") " pod="openstack/glance-default-external-api-0" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.922079 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/46f0fa4b-479a-42c5-b9b2-3f30a079be48-scripts\") pod \"glance-default-internal-api-0\" (UID: \"46f0fa4b-479a-42c5-b9b2-3f30a079be48\") " pod="openstack/glance-default-internal-api-0" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.922118 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46f0fa4b-479a-42c5-b9b2-3f30a079be48-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"46f0fa4b-479a-42c5-b9b2-3f30a079be48\") " pod="openstack/glance-default-internal-api-0" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.922164 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7qwlh\" (UniqueName: \"kubernetes.io/projected/12776f5e-c372-41bd-91fe-ba7d0915050a-kube-api-access-7qwlh\") pod \"glance-default-external-api-0\" (UID: \"12776f5e-c372-41bd-91fe-ba7d0915050a\") " pod="openstack/glance-default-external-api-0" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.922198 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12776f5e-c372-41bd-91fe-ba7d0915050a-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"12776f5e-c372-41bd-91fe-ba7d0915050a\") " pod="openstack/glance-default-external-api-0" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.922344 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v8grh\" (UniqueName: \"kubernetes.io/projected/46f0fa4b-479a-42c5-b9b2-3f30a079be48-kube-api-access-v8grh\") pod \"glance-default-internal-api-0\" (UID: \"46f0fa4b-479a-42c5-b9b2-3f30a079be48\") " pod="openstack/glance-default-internal-api-0" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.922388 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/46f0fa4b-479a-42c5-b9b2-3f30a079be48-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"46f0fa4b-479a-42c5-b9b2-3f30a079be48\") " pod="openstack/glance-default-internal-api-0" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.922518 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/46f0fa4b-479a-42c5-b9b2-3f30a079be48-logs\") pod \"glance-default-internal-api-0\" (UID: \"46f0fa4b-479a-42c5-b9b2-3f30a079be48\") " pod="openstack/glance-default-internal-api-0" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.922727 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/46f0fa4b-479a-42c5-b9b2-3f30a079be48-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"46f0fa4b-479a-42c5-b9b2-3f30a079be48\") " pod="openstack/glance-default-internal-api-0" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.922787 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/12776f5e-c372-41bd-91fe-ba7d0915050a-config-data\") pod \"glance-default-external-api-0\" (UID: \"12776f5e-c372-41bd-91fe-ba7d0915050a\") " pod="openstack/glance-default-external-api-0" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.922824 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"12776f5e-c372-41bd-91fe-ba7d0915050a\") " pod="openstack/glance-default-external-api-0" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.922854 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/46f0fa4b-479a-42c5-b9b2-3f30a079be48-logs\") pod \"glance-default-internal-api-0\" (UID: \"46f0fa4b-479a-42c5-b9b2-3f30a079be48\") " pod="openstack/glance-default-internal-api-0" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.922866 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/12776f5e-c372-41bd-91fe-ba7d0915050a-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"12776f5e-c372-41bd-91fe-ba7d0915050a\") " pod="openstack/glance-default-external-api-0" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.922921 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/46f0fa4b-479a-42c5-b9b2-3f30a079be48-config-data\") pod \"glance-default-internal-api-0\" (UID: \"46f0fa4b-479a-42c5-b9b2-3f30a079be48\") " pod="openstack/glance-default-internal-api-0" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.922944 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/46f0fa4b-479a-42c5-b9b2-3f30a079be48-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"46f0fa4b-479a-42c5-b9b2-3f30a079be48\") " pod="openstack/glance-default-internal-api-0" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.922976 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"46f0fa4b-479a-42c5-b9b2-3f30a079be48\") " pod="openstack/glance-default-internal-api-0" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.923002 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/12776f5e-c372-41bd-91fe-ba7d0915050a-scripts\") pod \"glance-default-external-api-0\" (UID: \"12776f5e-c372-41bd-91fe-ba7d0915050a\") " pod="openstack/glance-default-external-api-0" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.923132 4824 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"46f0fa4b-479a-42c5-b9b2-3f30a079be48\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/glance-default-internal-api-0" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.927599 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/46f0fa4b-479a-42c5-b9b2-3f30a079be48-scripts\") pod \"glance-default-internal-api-0\" (UID: \"46f0fa4b-479a-42c5-b9b2-3f30a079be48\") " pod="openstack/glance-default-internal-api-0" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.927876 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46f0fa4b-479a-42c5-b9b2-3f30a079be48-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"46f0fa4b-479a-42c5-b9b2-3f30a079be48\") " pod="openstack/glance-default-internal-api-0" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.928853 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/46f0fa4b-479a-42c5-b9b2-3f30a079be48-config-data\") pod \"glance-default-internal-api-0\" (UID: \"46f0fa4b-479a-42c5-b9b2-3f30a079be48\") " pod="openstack/glance-default-internal-api-0" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.928904 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/46f0fa4b-479a-42c5-b9b2-3f30a079be48-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"46f0fa4b-479a-42c5-b9b2-3f30a079be48\") " pod="openstack/glance-default-internal-api-0" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.935340 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v8grh\" (UniqueName: \"kubernetes.io/projected/46f0fa4b-479a-42c5-b9b2-3f30a079be48-kube-api-access-v8grh\") pod \"glance-default-internal-api-0\" (UID: \"46f0fa4b-479a-42c5-b9b2-3f30a079be48\") " pod="openstack/glance-default-internal-api-0" Jan 21 11:26:40 crc kubenswrapper[4824]: I0121 11:26:40.940946 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"46f0fa4b-479a-42c5-b9b2-3f30a079be48\") " pod="openstack/glance-default-internal-api-0" Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.024013 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/12776f5e-c372-41bd-91fe-ba7d0915050a-config-data\") pod \"glance-default-external-api-0\" (UID: \"12776f5e-c372-41bd-91fe-ba7d0915050a\") " pod="openstack/glance-default-external-api-0" Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.024053 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"12776f5e-c372-41bd-91fe-ba7d0915050a\") " pod="openstack/glance-default-external-api-0" Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.024082 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/12776f5e-c372-41bd-91fe-ba7d0915050a-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"12776f5e-c372-41bd-91fe-ba7d0915050a\") " pod="openstack/glance-default-external-api-0" Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.024115 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/12776f5e-c372-41bd-91fe-ba7d0915050a-scripts\") pod \"glance-default-external-api-0\" (UID: \"12776f5e-c372-41bd-91fe-ba7d0915050a\") " pod="openstack/glance-default-external-api-0" Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.024129 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/12776f5e-c372-41bd-91fe-ba7d0915050a-logs\") pod \"glance-default-external-api-0\" (UID: \"12776f5e-c372-41bd-91fe-ba7d0915050a\") " pod="openstack/glance-default-external-api-0" Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.024151 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/12776f5e-c372-41bd-91fe-ba7d0915050a-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"12776f5e-c372-41bd-91fe-ba7d0915050a\") " pod="openstack/glance-default-external-api-0" Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.024185 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7qwlh\" (UniqueName: \"kubernetes.io/projected/12776f5e-c372-41bd-91fe-ba7d0915050a-kube-api-access-7qwlh\") pod \"glance-default-external-api-0\" (UID: \"12776f5e-c372-41bd-91fe-ba7d0915050a\") " pod="openstack/glance-default-external-api-0" Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.024185 4824 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"12776f5e-c372-41bd-91fe-ba7d0915050a\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/glance-default-external-api-0" Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.024206 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12776f5e-c372-41bd-91fe-ba7d0915050a-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"12776f5e-c372-41bd-91fe-ba7d0915050a\") " pod="openstack/glance-default-external-api-0" Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.024623 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/12776f5e-c372-41bd-91fe-ba7d0915050a-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"12776f5e-c372-41bd-91fe-ba7d0915050a\") " pod="openstack/glance-default-external-api-0" Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.024760 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/12776f5e-c372-41bd-91fe-ba7d0915050a-logs\") pod \"glance-default-external-api-0\" (UID: \"12776f5e-c372-41bd-91fe-ba7d0915050a\") " pod="openstack/glance-default-external-api-0" Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.027768 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/12776f5e-c372-41bd-91fe-ba7d0915050a-scripts\") pod \"glance-default-external-api-0\" (UID: \"12776f5e-c372-41bd-91fe-ba7d0915050a\") " pod="openstack/glance-default-external-api-0" Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.028337 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/12776f5e-c372-41bd-91fe-ba7d0915050a-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"12776f5e-c372-41bd-91fe-ba7d0915050a\") " pod="openstack/glance-default-external-api-0" Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.028654 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/12776f5e-c372-41bd-91fe-ba7d0915050a-config-data\") pod \"glance-default-external-api-0\" (UID: \"12776f5e-c372-41bd-91fe-ba7d0915050a\") " pod="openstack/glance-default-external-api-0" Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.028811 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12776f5e-c372-41bd-91fe-ba7d0915050a-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"12776f5e-c372-41bd-91fe-ba7d0915050a\") " pod="openstack/glance-default-external-api-0" Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.037467 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7qwlh\" (UniqueName: \"kubernetes.io/projected/12776f5e-c372-41bd-91fe-ba7d0915050a-kube-api-access-7qwlh\") pod \"glance-default-external-api-0\" (UID: \"12776f5e-c372-41bd-91fe-ba7d0915050a\") " pod="openstack/glance-default-external-api-0" Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.041020 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"12776f5e-c372-41bd-91fe-ba7d0915050a\") " pod="openstack/glance-default-external-api-0" Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.050031 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.063517 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.160306 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-vbmkp"] Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.166150 4824 scope.go:117] "RemoveContainer" containerID="859f2e0dceb5ba0b5dc20dc8542580aaaceec1b327c7c16135166be30bdad2a1" Jan 21 11:26:41 crc kubenswrapper[4824]: E0121 11:26:41.166150 4824 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified" Jan 21 11:26:41 crc kubenswrapper[4824]: E0121 11:26:41.166343 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-msmnx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-rlp2t_openstack(0a6c0628-4522-4bb7-8a82-cc2e019eca2a): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Jan 21 11:26:41 crc kubenswrapper[4824]: E0121 11:26:41.167919 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-rlp2t" podUID="0a6c0628-4522-4bb7-8a82-cc2e019eca2a" Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.170009 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55f844cf75-vbmkp" Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.174261 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-vbmkp"] Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.201982 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74f6bcbc87-sj79w" Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.243831 4824 scope.go:117] "RemoveContainer" containerID="7f3d59f931487b81345a89a9579c3ee29598f0123e3fc5337b9d33bd8702f80f" Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.319068 4824 scope.go:117] "RemoveContainer" containerID="b624461c7fd817bff45a1813fc59b08421416d9686db5ebe66efe4049b30534b" Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.342665 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c4a81903-5e1f-4d9c-bacf-60639069db18-config\") pod \"c4a81903-5e1f-4d9c-bacf-60639069db18\" (UID: \"c4a81903-5e1f-4d9c-bacf-60639069db18\") " Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.342739 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c4a81903-5e1f-4d9c-bacf-60639069db18-ovsdbserver-nb\") pod \"c4a81903-5e1f-4d9c-bacf-60639069db18\" (UID: \"c4a81903-5e1f-4d9c-bacf-60639069db18\") " Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.342829 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c4a81903-5e1f-4d9c-bacf-60639069db18-ovsdbserver-sb\") pod \"c4a81903-5e1f-4d9c-bacf-60639069db18\" (UID: \"c4a81903-5e1f-4d9c-bacf-60639069db18\") " Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.342918 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c4a81903-5e1f-4d9c-bacf-60639069db18-dns-svc\") pod \"c4a81903-5e1f-4d9c-bacf-60639069db18\" (UID: \"c4a81903-5e1f-4d9c-bacf-60639069db18\") " Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.342996 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xgrw\" (UniqueName: \"kubernetes.io/projected/c4a81903-5e1f-4d9c-bacf-60639069db18-kube-api-access-9xgrw\") pod \"c4a81903-5e1f-4d9c-bacf-60639069db18\" (UID: \"c4a81903-5e1f-4d9c-bacf-60639069db18\") " Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.343028 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c4a81903-5e1f-4d9c-bacf-60639069db18-dns-swift-storage-0\") pod \"c4a81903-5e1f-4d9c-bacf-60639069db18\" (UID: \"c4a81903-5e1f-4d9c-bacf-60639069db18\") " Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.343362 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bb4ea3d2-63c7-4829-b6b8-8112476493fa-ovsdbserver-nb\") pod \"dnsmasq-dns-55f844cf75-vbmkp\" (UID: \"bb4ea3d2-63c7-4829-b6b8-8112476493fa\") " pod="openstack/dnsmasq-dns-55f844cf75-vbmkp" Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.343385 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bb4ea3d2-63c7-4829-b6b8-8112476493fa-dns-svc\") pod \"dnsmasq-dns-55f844cf75-vbmkp\" (UID: \"bb4ea3d2-63c7-4829-b6b8-8112476493fa\") " pod="openstack/dnsmasq-dns-55f844cf75-vbmkp" Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.343403 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bb4ea3d2-63c7-4829-b6b8-8112476493fa-dns-swift-storage-0\") pod \"dnsmasq-dns-55f844cf75-vbmkp\" (UID: \"bb4ea3d2-63c7-4829-b6b8-8112476493fa\") " pod="openstack/dnsmasq-dns-55f844cf75-vbmkp" Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.343807 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bb4ea3d2-63c7-4829-b6b8-8112476493fa-config\") pod \"dnsmasq-dns-55f844cf75-vbmkp\" (UID: \"bb4ea3d2-63c7-4829-b6b8-8112476493fa\") " pod="openstack/dnsmasq-dns-55f844cf75-vbmkp" Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.343834 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-clj2h\" (UniqueName: \"kubernetes.io/projected/bb4ea3d2-63c7-4829-b6b8-8112476493fa-kube-api-access-clj2h\") pod \"dnsmasq-dns-55f844cf75-vbmkp\" (UID: \"bb4ea3d2-63c7-4829-b6b8-8112476493fa\") " pod="openstack/dnsmasq-dns-55f844cf75-vbmkp" Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.343872 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bb4ea3d2-63c7-4829-b6b8-8112476493fa-ovsdbserver-sb\") pod \"dnsmasq-dns-55f844cf75-vbmkp\" (UID: \"bb4ea3d2-63c7-4829-b6b8-8112476493fa\") " pod="openstack/dnsmasq-dns-55f844cf75-vbmkp" Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.353762 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c4a81903-5e1f-4d9c-bacf-60639069db18-kube-api-access-9xgrw" (OuterVolumeSpecName: "kube-api-access-9xgrw") pod "c4a81903-5e1f-4d9c-bacf-60639069db18" (UID: "c4a81903-5e1f-4d9c-bacf-60639069db18"). InnerVolumeSpecName "kube-api-access-9xgrw". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.399296 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c4a81903-5e1f-4d9c-bacf-60639069db18-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "c4a81903-5e1f-4d9c-bacf-60639069db18" (UID: "c4a81903-5e1f-4d9c-bacf-60639069db18"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.406873 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c4a81903-5e1f-4d9c-bacf-60639069db18-config" (OuterVolumeSpecName: "config") pod "c4a81903-5e1f-4d9c-bacf-60639069db18" (UID: "c4a81903-5e1f-4d9c-bacf-60639069db18"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.430364 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-79bfcdcfbd-s2v5h"] Jan 21 11:26:41 crc kubenswrapper[4824]: E0121 11:26:41.430769 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c4a81903-5e1f-4d9c-bacf-60639069db18" containerName="dnsmasq-dns" Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.430787 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="c4a81903-5e1f-4d9c-bacf-60639069db18" containerName="dnsmasq-dns" Jan 21 11:26:41 crc kubenswrapper[4824]: E0121 11:26:41.430800 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c4a81903-5e1f-4d9c-bacf-60639069db18" containerName="init" Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.430805 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="c4a81903-5e1f-4d9c-bacf-60639069db18" containerName="init" Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.430984 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="c4a81903-5e1f-4d9c-bacf-60639069db18" containerName="dnsmasq-dns" Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.443061 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-79bfcdcfbd-s2v5h" Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.446434 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bb4ea3d2-63c7-4829-b6b8-8112476493fa-config\") pod \"dnsmasq-dns-55f844cf75-vbmkp\" (UID: \"bb4ea3d2-63c7-4829-b6b8-8112476493fa\") " pod="openstack/dnsmasq-dns-55f844cf75-vbmkp" Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.446816 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-clj2h\" (UniqueName: \"kubernetes.io/projected/bb4ea3d2-63c7-4829-b6b8-8112476493fa-kube-api-access-clj2h\") pod \"dnsmasq-dns-55f844cf75-vbmkp\" (UID: \"bb4ea3d2-63c7-4829-b6b8-8112476493fa\") " pod="openstack/dnsmasq-dns-55f844cf75-vbmkp" Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.446866 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bb4ea3d2-63c7-4829-b6b8-8112476493fa-ovsdbserver-sb\") pod \"dnsmasq-dns-55f844cf75-vbmkp\" (UID: \"bb4ea3d2-63c7-4829-b6b8-8112476493fa\") " pod="openstack/dnsmasq-dns-55f844cf75-vbmkp" Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.446946 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.447078 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bb4ea3d2-63c7-4829-b6b8-8112476493fa-ovsdbserver-nb\") pod \"dnsmasq-dns-55f844cf75-vbmkp\" (UID: \"bb4ea3d2-63c7-4829-b6b8-8112476493fa\") " pod="openstack/dnsmasq-dns-55f844cf75-vbmkp" Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.447109 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bb4ea3d2-63c7-4829-b6b8-8112476493fa-dns-svc\") pod \"dnsmasq-dns-55f844cf75-vbmkp\" (UID: \"bb4ea3d2-63c7-4829-b6b8-8112476493fa\") " pod="openstack/dnsmasq-dns-55f844cf75-vbmkp" Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.447166 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bb4ea3d2-63c7-4829-b6b8-8112476493fa-dns-swift-storage-0\") pod \"dnsmasq-dns-55f844cf75-vbmkp\" (UID: \"bb4ea3d2-63c7-4829-b6b8-8112476493fa\") " pod="openstack/dnsmasq-dns-55f844cf75-vbmkp" Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.447562 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xgrw\" (UniqueName: \"kubernetes.io/projected/c4a81903-5e1f-4d9c-bacf-60639069db18-kube-api-access-9xgrw\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.447582 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c4a81903-5e1f-4d9c-bacf-60639069db18-config\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.447591 4824 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c4a81903-5e1f-4d9c-bacf-60639069db18-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.447920 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bb4ea3d2-63c7-4829-b6b8-8112476493fa-config\") pod \"dnsmasq-dns-55f844cf75-vbmkp\" (UID: \"bb4ea3d2-63c7-4829-b6b8-8112476493fa\") " pod="openstack/dnsmasq-dns-55f844cf75-vbmkp" Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.447948 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bb4ea3d2-63c7-4829-b6b8-8112476493fa-ovsdbserver-sb\") pod \"dnsmasq-dns-55f844cf75-vbmkp\" (UID: \"bb4ea3d2-63c7-4829-b6b8-8112476493fa\") " pod="openstack/dnsmasq-dns-55f844cf75-vbmkp" Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.448000 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-49xnt" Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.448559 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.448755 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.456478 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-79bfcdcfbd-s2v5h"] Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.464970 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bb4ea3d2-63c7-4829-b6b8-8112476493fa-dns-svc\") pod \"dnsmasq-dns-55f844cf75-vbmkp\" (UID: \"bb4ea3d2-63c7-4829-b6b8-8112476493fa\") " pod="openstack/dnsmasq-dns-55f844cf75-vbmkp" Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.465644 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bb4ea3d2-63c7-4829-b6b8-8112476493fa-dns-swift-storage-0\") pod \"dnsmasq-dns-55f844cf75-vbmkp\" (UID: \"bb4ea3d2-63c7-4829-b6b8-8112476493fa\") " pod="openstack/dnsmasq-dns-55f844cf75-vbmkp" Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.467492 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bb4ea3d2-63c7-4829-b6b8-8112476493fa-ovsdbserver-nb\") pod \"dnsmasq-dns-55f844cf75-vbmkp\" (UID: \"bb4ea3d2-63c7-4829-b6b8-8112476493fa\") " pod="openstack/dnsmasq-dns-55f844cf75-vbmkp" Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.469900 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c4a81903-5e1f-4d9c-bacf-60639069db18-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "c4a81903-5e1f-4d9c-bacf-60639069db18" (UID: "c4a81903-5e1f-4d9c-bacf-60639069db18"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.474842 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-clj2h\" (UniqueName: \"kubernetes.io/projected/bb4ea3d2-63c7-4829-b6b8-8112476493fa-kube-api-access-clj2h\") pod \"dnsmasq-dns-55f844cf75-vbmkp\" (UID: \"bb4ea3d2-63c7-4829-b6b8-8112476493fa\") " pod="openstack/dnsmasq-dns-55f844cf75-vbmkp" Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.510105 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55f844cf75-vbmkp" Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.513080 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c4a81903-5e1f-4d9c-bacf-60639069db18-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "c4a81903-5e1f-4d9c-bacf-60639069db18" (UID: "c4a81903-5e1f-4d9c-bacf-60639069db18"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.526859 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c4a81903-5e1f-4d9c-bacf-60639069db18-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "c4a81903-5e1f-4d9c-bacf-60639069db18" (UID: "c4a81903-5e1f-4d9c-bacf-60639069db18"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.550831 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/d9302662-76b1-4acc-b9e3-ffde925d34b2-httpd-config\") pod \"neutron-79bfcdcfbd-s2v5h\" (UID: \"d9302662-76b1-4acc-b9e3-ffde925d34b2\") " pod="openstack/neutron-79bfcdcfbd-s2v5h" Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.550907 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9302662-76b1-4acc-b9e3-ffde925d34b2-combined-ca-bundle\") pod \"neutron-79bfcdcfbd-s2v5h\" (UID: \"d9302662-76b1-4acc-b9e3-ffde925d34b2\") " pod="openstack/neutron-79bfcdcfbd-s2v5h" Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.551922 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/d9302662-76b1-4acc-b9e3-ffde925d34b2-ovndb-tls-certs\") pod \"neutron-79bfcdcfbd-s2v5h\" (UID: \"d9302662-76b1-4acc-b9e3-ffde925d34b2\") " pod="openstack/neutron-79bfcdcfbd-s2v5h" Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.552004 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/d9302662-76b1-4acc-b9e3-ffde925d34b2-config\") pod \"neutron-79bfcdcfbd-s2v5h\" (UID: \"d9302662-76b1-4acc-b9e3-ffde925d34b2\") " pod="openstack/neutron-79bfcdcfbd-s2v5h" Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.552032 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9x8cf\" (UniqueName: \"kubernetes.io/projected/d9302662-76b1-4acc-b9e3-ffde925d34b2-kube-api-access-9x8cf\") pod \"neutron-79bfcdcfbd-s2v5h\" (UID: \"d9302662-76b1-4acc-b9e3-ffde925d34b2\") " pod="openstack/neutron-79bfcdcfbd-s2v5h" Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.552112 4824 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c4a81903-5e1f-4d9c-bacf-60639069db18-dns-svc\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.552153 4824 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c4a81903-5e1f-4d9c-bacf-60639069db18-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.552165 4824 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c4a81903-5e1f-4d9c-bacf-60639069db18-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.654654 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/d9302662-76b1-4acc-b9e3-ffde925d34b2-httpd-config\") pod \"neutron-79bfcdcfbd-s2v5h\" (UID: \"d9302662-76b1-4acc-b9e3-ffde925d34b2\") " pod="openstack/neutron-79bfcdcfbd-s2v5h" Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.654737 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9302662-76b1-4acc-b9e3-ffde925d34b2-combined-ca-bundle\") pod \"neutron-79bfcdcfbd-s2v5h\" (UID: \"d9302662-76b1-4acc-b9e3-ffde925d34b2\") " pod="openstack/neutron-79bfcdcfbd-s2v5h" Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.654888 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/d9302662-76b1-4acc-b9e3-ffde925d34b2-ovndb-tls-certs\") pod \"neutron-79bfcdcfbd-s2v5h\" (UID: \"d9302662-76b1-4acc-b9e3-ffde925d34b2\") " pod="openstack/neutron-79bfcdcfbd-s2v5h" Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.654936 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/d9302662-76b1-4acc-b9e3-ffde925d34b2-config\") pod \"neutron-79bfcdcfbd-s2v5h\" (UID: \"d9302662-76b1-4acc-b9e3-ffde925d34b2\") " pod="openstack/neutron-79bfcdcfbd-s2v5h" Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.654972 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9x8cf\" (UniqueName: \"kubernetes.io/projected/d9302662-76b1-4acc-b9e3-ffde925d34b2-kube-api-access-9x8cf\") pod \"neutron-79bfcdcfbd-s2v5h\" (UID: \"d9302662-76b1-4acc-b9e3-ffde925d34b2\") " pod="openstack/neutron-79bfcdcfbd-s2v5h" Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.660834 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/d9302662-76b1-4acc-b9e3-ffde925d34b2-config\") pod \"neutron-79bfcdcfbd-s2v5h\" (UID: \"d9302662-76b1-4acc-b9e3-ffde925d34b2\") " pod="openstack/neutron-79bfcdcfbd-s2v5h" Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.662412 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9302662-76b1-4acc-b9e3-ffde925d34b2-combined-ca-bundle\") pod \"neutron-79bfcdcfbd-s2v5h\" (UID: \"d9302662-76b1-4acc-b9e3-ffde925d34b2\") " pod="openstack/neutron-79bfcdcfbd-s2v5h" Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.662457 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/d9302662-76b1-4acc-b9e3-ffde925d34b2-ovndb-tls-certs\") pod \"neutron-79bfcdcfbd-s2v5h\" (UID: \"d9302662-76b1-4acc-b9e3-ffde925d34b2\") " pod="openstack/neutron-79bfcdcfbd-s2v5h" Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.672689 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/d9302662-76b1-4acc-b9e3-ffde925d34b2-httpd-config\") pod \"neutron-79bfcdcfbd-s2v5h\" (UID: \"d9302662-76b1-4acc-b9e3-ffde925d34b2\") " pod="openstack/neutron-79bfcdcfbd-s2v5h" Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.680030 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-5l2gg"] Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.688481 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9x8cf\" (UniqueName: \"kubernetes.io/projected/d9302662-76b1-4acc-b9e3-ffde925d34b2-kube-api-access-9x8cf\") pod \"neutron-79bfcdcfbd-s2v5h\" (UID: \"d9302662-76b1-4acc-b9e3-ffde925d34b2\") " pod="openstack/neutron-79bfcdcfbd-s2v5h" Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.694498 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-jrsn8" event={"ID":"cadf2d38-24f9-43ab-8e29-803d97a0ff06","Type":"ContainerStarted","Data":"5dcf0246f4ab2f2f4f3230d6a47ec6ac1102fcb34b4edb4d89a73c99d792f5ff"} Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.714164 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74f6bcbc87-sj79w" event={"ID":"c4a81903-5e1f-4d9c-bacf-60639069db18","Type":"ContainerDied","Data":"19d64cd123fe9ef3646e70cf5808fefa9acba69592569cd21640e9399cbca5d9"} Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.714209 4824 scope.go:117] "RemoveContainer" containerID="862050906c3faeb783c3cf1a55874b97964bf3e73dcab5972c6e49f1ea08c530" Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.714327 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74f6bcbc87-sj79w" Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.746634 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-6vgh6" event={"ID":"08b8c363-a8c1-4dfe-8fd1-1af2d055a305","Type":"ContainerStarted","Data":"5f4a1e6bcd0c138e59cd1e69d47cce9046bae28ab6b9ce57bd7be2ad0c6a4e61"} Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.748592 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-db-sync-jrsn8" podStartSLOduration=1.8425048130000001 podStartE2EDuration="20.748578322s" podCreationTimestamp="2026-01-21 11:26:21 +0000 UTC" firstStartedPulling="2026-01-21 11:26:22.260048638 +0000 UTC m=+924.553077930" lastFinishedPulling="2026-01-21 11:26:41.166122146 +0000 UTC m=+943.459151439" observedRunningTime="2026-01-21 11:26:41.7104694 +0000 UTC m=+944.003498681" watchObservedRunningTime="2026-01-21 11:26:41.748578322 +0000 UTC m=+944.041607613" Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.756938 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-sj79w"] Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.758402 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bbefda93-dc56-43c5-b131-56af96ca832b","Type":"ContainerStarted","Data":"c4e76cd725e81ba223348c70595837e4caaba881aaa9eba8437cfc5280bd859c"} Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.764735 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-79bfcdcfbd-s2v5h" Jan 21 11:26:41 crc kubenswrapper[4824]: E0121 11:26:41.765666 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified\\\"\"" pod="openstack/cinder-db-sync-rlp2t" podUID="0a6c0628-4522-4bb7-8a82-cc2e019eca2a" Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.765671 4824 scope.go:117] "RemoveContainer" containerID="0aacefa5ca6057f616f2afc320e43a0e95b25311f579f1cc69ade834fb63b2f0" Jan 21 11:26:41 crc kubenswrapper[4824]: W0121 11:26:41.769070 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod46f0fa4b_479a_42c5_b9b2_3f30a079be48.slice/crio-5dc758f1e11243d2a910301ab89754d05c98dfa5446bb49c248a5833a44c6ce9 WatchSource:0}: Error finding container 5dc758f1e11243d2a910301ab89754d05c98dfa5446bb49c248a5833a44c6ce9: Status 404 returned error can't find the container with id 5dc758f1e11243d2a910301ab89754d05c98dfa5446bb49c248a5833a44c6ce9 Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.771171 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-sj79w"] Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.779981 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.780332 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-6vgh6" podStartSLOduration=2.210949144 podStartE2EDuration="20.780315956s" podCreationTimestamp="2026-01-21 11:26:21 +0000 UTC" firstStartedPulling="2026-01-21 11:26:22.566259702 +0000 UTC m=+924.859288994" lastFinishedPulling="2026-01-21 11:26:41.135626515 +0000 UTC m=+943.428655806" observedRunningTime="2026-01-21 11:26:41.762243607 +0000 UTC m=+944.055272899" watchObservedRunningTime="2026-01-21 11:26:41.780315956 +0000 UTC m=+944.073345248" Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.848867 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Jan 21 11:26:41 crc kubenswrapper[4824]: I0121 11:26:41.939206 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-vbmkp"] Jan 21 11:26:42 crc kubenswrapper[4824]: I0121 11:26:42.062879 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6a5a3f84-2320-4f8f-b8d4-d6ee9053f859" path="/var/lib/kubelet/pods/6a5a3f84-2320-4f8f-b8d4-d6ee9053f859/volumes" Jan 21 11:26:42 crc kubenswrapper[4824]: I0121 11:26:42.063692 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7d0e8430-4bfe-4e37-8c22-502dd5444af4" path="/var/lib/kubelet/pods/7d0e8430-4bfe-4e37-8c22-502dd5444af4/volumes" Jan 21 11:26:42 crc kubenswrapper[4824]: I0121 11:26:42.064718 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c4a81903-5e1f-4d9c-bacf-60639069db18" path="/var/lib/kubelet/pods/c4a81903-5e1f-4d9c-bacf-60639069db18/volumes" Jan 21 11:26:42 crc kubenswrapper[4824]: I0121 11:26:42.288037 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-79bfcdcfbd-s2v5h"] Jan 21 11:26:42 crc kubenswrapper[4824]: W0121 11:26:42.303499 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd9302662_76b1_4acc_b9e3_ffde925d34b2.slice/crio-2df3211773129d1b5d01ef3e6ca36fb1965741e7764384f48d28c04cb0d0626d WatchSource:0}: Error finding container 2df3211773129d1b5d01ef3e6ca36fb1965741e7764384f48d28c04cb0d0626d: Status 404 returned error can't find the container with id 2df3211773129d1b5d01ef3e6ca36fb1965741e7764384f48d28c04cb0d0626d Jan 21 11:26:42 crc kubenswrapper[4824]: I0121 11:26:42.767636 4824 generic.go:334] "Generic (PLEG): container finished" podID="bb4ea3d2-63c7-4829-b6b8-8112476493fa" containerID="f5715a5c2dbadf161e906011876459ee94ba9a1523d9455c1d55ac180e574139" exitCode=0 Jan 21 11:26:42 crc kubenswrapper[4824]: I0121 11:26:42.767923 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-vbmkp" event={"ID":"bb4ea3d2-63c7-4829-b6b8-8112476493fa","Type":"ContainerDied","Data":"f5715a5c2dbadf161e906011876459ee94ba9a1523d9455c1d55ac180e574139"} Jan 21 11:26:42 crc kubenswrapper[4824]: I0121 11:26:42.767951 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-vbmkp" event={"ID":"bb4ea3d2-63c7-4829-b6b8-8112476493fa","Type":"ContainerStarted","Data":"5e519680bc7c55a0ecc86ba490dcbb3969a313e5b2bc06cbb94da1a485a26d3b"} Jan 21 11:26:42 crc kubenswrapper[4824]: I0121 11:26:42.770903 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"12776f5e-c372-41bd-91fe-ba7d0915050a","Type":"ContainerStarted","Data":"9f11951e86971107f0a8d485c5dead64b960ce2e124caba92b7e241dae731aca"} Jan 21 11:26:42 crc kubenswrapper[4824]: I0121 11:26:42.770942 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"12776f5e-c372-41bd-91fe-ba7d0915050a","Type":"ContainerStarted","Data":"eaaf72feb767aa6ebf839a27a99ea70de0682afc00ba0ae8e8e10dabfcaf1864"} Jan 21 11:26:42 crc kubenswrapper[4824]: I0121 11:26:42.772472 4824 generic.go:334] "Generic (PLEG): container finished" podID="08b8c363-a8c1-4dfe-8fd1-1af2d055a305" containerID="5f4a1e6bcd0c138e59cd1e69d47cce9046bae28ab6b9ce57bd7be2ad0c6a4e61" exitCode=0 Jan 21 11:26:42 crc kubenswrapper[4824]: I0121 11:26:42.772517 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-6vgh6" event={"ID":"08b8c363-a8c1-4dfe-8fd1-1af2d055a305","Type":"ContainerDied","Data":"5f4a1e6bcd0c138e59cd1e69d47cce9046bae28ab6b9ce57bd7be2ad0c6a4e61"} Jan 21 11:26:42 crc kubenswrapper[4824]: I0121 11:26:42.773759 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"46f0fa4b-479a-42c5-b9b2-3f30a079be48","Type":"ContainerStarted","Data":"953a68ff3f5b22fd88f3a1a2babc253f6f9d0648a9b5bca14c07b85059d64e1a"} Jan 21 11:26:42 crc kubenswrapper[4824]: I0121 11:26:42.773782 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"46f0fa4b-479a-42c5-b9b2-3f30a079be48","Type":"ContainerStarted","Data":"5dc758f1e11243d2a910301ab89754d05c98dfa5446bb49c248a5833a44c6ce9"} Jan 21 11:26:42 crc kubenswrapper[4824]: I0121 11:26:42.774846 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-5l2gg" event={"ID":"a7e75011-d633-4ccc-951c-d019cb2100f9","Type":"ContainerStarted","Data":"2534da156868c09c8dd8dee0424c34fab5c7b1401695a79ef623bde9f1ef00f0"} Jan 21 11:26:42 crc kubenswrapper[4824]: I0121 11:26:42.774870 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-5l2gg" event={"ID":"a7e75011-d633-4ccc-951c-d019cb2100f9","Type":"ContainerStarted","Data":"f73eda2acba113b106a0d01407072387e56c3b86b9c9c6135b58a1185903e93b"} Jan 21 11:26:42 crc kubenswrapper[4824]: I0121 11:26:42.778015 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-79bfcdcfbd-s2v5h" event={"ID":"d9302662-76b1-4acc-b9e3-ffde925d34b2","Type":"ContainerStarted","Data":"ff804ebe93733fa9e697132922f8e30c6ee75d058b7cb8afecec7da791271c35"} Jan 21 11:26:42 crc kubenswrapper[4824]: I0121 11:26:42.778041 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-79bfcdcfbd-s2v5h" event={"ID":"d9302662-76b1-4acc-b9e3-ffde925d34b2","Type":"ContainerStarted","Data":"2df3211773129d1b5d01ef3e6ca36fb1965741e7764384f48d28c04cb0d0626d"} Jan 21 11:26:42 crc kubenswrapper[4824]: I0121 11:26:42.802124 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-5l2gg" podStartSLOduration=7.8021119169999995 podStartE2EDuration="7.802111917s" podCreationTimestamp="2026-01-21 11:26:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:26:42.795007916 +0000 UTC m=+945.088037209" watchObservedRunningTime="2026-01-21 11:26:42.802111917 +0000 UTC m=+945.095141209" Jan 21 11:26:43 crc kubenswrapper[4824]: I0121 11:26:43.486787 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-659565f76f-jdg2q"] Jan 21 11:26:43 crc kubenswrapper[4824]: I0121 11:26:43.489747 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-659565f76f-jdg2q" Jan 21 11:26:43 crc kubenswrapper[4824]: I0121 11:26:43.491454 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Jan 21 11:26:43 crc kubenswrapper[4824]: I0121 11:26:43.492619 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Jan 21 11:26:43 crc kubenswrapper[4824]: I0121 11:26:43.521347 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-659565f76f-jdg2q"] Jan 21 11:26:43 crc kubenswrapper[4824]: I0121 11:26:43.590474 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b6f2ba03-746e-465c-bde2-c917129fb125-internal-tls-certs\") pod \"neutron-659565f76f-jdg2q\" (UID: \"b6f2ba03-746e-465c-bde2-c917129fb125\") " pod="openstack/neutron-659565f76f-jdg2q" Jan 21 11:26:43 crc kubenswrapper[4824]: I0121 11:26:43.590515 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/b6f2ba03-746e-465c-bde2-c917129fb125-httpd-config\") pod \"neutron-659565f76f-jdg2q\" (UID: \"b6f2ba03-746e-465c-bde2-c917129fb125\") " pod="openstack/neutron-659565f76f-jdg2q" Jan 21 11:26:43 crc kubenswrapper[4824]: I0121 11:26:43.590545 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/b6f2ba03-746e-465c-bde2-c917129fb125-ovndb-tls-certs\") pod \"neutron-659565f76f-jdg2q\" (UID: \"b6f2ba03-746e-465c-bde2-c917129fb125\") " pod="openstack/neutron-659565f76f-jdg2q" Jan 21 11:26:43 crc kubenswrapper[4824]: I0121 11:26:43.590608 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/b6f2ba03-746e-465c-bde2-c917129fb125-config\") pod \"neutron-659565f76f-jdg2q\" (UID: \"b6f2ba03-746e-465c-bde2-c917129fb125\") " pod="openstack/neutron-659565f76f-jdg2q" Jan 21 11:26:43 crc kubenswrapper[4824]: I0121 11:26:43.590644 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c6dqp\" (UniqueName: \"kubernetes.io/projected/b6f2ba03-746e-465c-bde2-c917129fb125-kube-api-access-c6dqp\") pod \"neutron-659565f76f-jdg2q\" (UID: \"b6f2ba03-746e-465c-bde2-c917129fb125\") " pod="openstack/neutron-659565f76f-jdg2q" Jan 21 11:26:43 crc kubenswrapper[4824]: I0121 11:26:43.590660 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b6f2ba03-746e-465c-bde2-c917129fb125-public-tls-certs\") pod \"neutron-659565f76f-jdg2q\" (UID: \"b6f2ba03-746e-465c-bde2-c917129fb125\") " pod="openstack/neutron-659565f76f-jdg2q" Jan 21 11:26:43 crc kubenswrapper[4824]: I0121 11:26:43.590674 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6f2ba03-746e-465c-bde2-c917129fb125-combined-ca-bundle\") pod \"neutron-659565f76f-jdg2q\" (UID: \"b6f2ba03-746e-465c-bde2-c917129fb125\") " pod="openstack/neutron-659565f76f-jdg2q" Jan 21 11:26:43 crc kubenswrapper[4824]: I0121 11:26:43.692246 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b6f2ba03-746e-465c-bde2-c917129fb125-internal-tls-certs\") pod \"neutron-659565f76f-jdg2q\" (UID: \"b6f2ba03-746e-465c-bde2-c917129fb125\") " pod="openstack/neutron-659565f76f-jdg2q" Jan 21 11:26:43 crc kubenswrapper[4824]: I0121 11:26:43.692292 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/b6f2ba03-746e-465c-bde2-c917129fb125-httpd-config\") pod \"neutron-659565f76f-jdg2q\" (UID: \"b6f2ba03-746e-465c-bde2-c917129fb125\") " pod="openstack/neutron-659565f76f-jdg2q" Jan 21 11:26:43 crc kubenswrapper[4824]: I0121 11:26:43.692332 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/b6f2ba03-746e-465c-bde2-c917129fb125-ovndb-tls-certs\") pod \"neutron-659565f76f-jdg2q\" (UID: \"b6f2ba03-746e-465c-bde2-c917129fb125\") " pod="openstack/neutron-659565f76f-jdg2q" Jan 21 11:26:43 crc kubenswrapper[4824]: I0121 11:26:43.692364 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/b6f2ba03-746e-465c-bde2-c917129fb125-config\") pod \"neutron-659565f76f-jdg2q\" (UID: \"b6f2ba03-746e-465c-bde2-c917129fb125\") " pod="openstack/neutron-659565f76f-jdg2q" Jan 21 11:26:43 crc kubenswrapper[4824]: I0121 11:26:43.692408 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c6dqp\" (UniqueName: \"kubernetes.io/projected/b6f2ba03-746e-465c-bde2-c917129fb125-kube-api-access-c6dqp\") pod \"neutron-659565f76f-jdg2q\" (UID: \"b6f2ba03-746e-465c-bde2-c917129fb125\") " pod="openstack/neutron-659565f76f-jdg2q" Jan 21 11:26:43 crc kubenswrapper[4824]: I0121 11:26:43.692428 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b6f2ba03-746e-465c-bde2-c917129fb125-public-tls-certs\") pod \"neutron-659565f76f-jdg2q\" (UID: \"b6f2ba03-746e-465c-bde2-c917129fb125\") " pod="openstack/neutron-659565f76f-jdg2q" Jan 21 11:26:43 crc kubenswrapper[4824]: I0121 11:26:43.692443 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6f2ba03-746e-465c-bde2-c917129fb125-combined-ca-bundle\") pod \"neutron-659565f76f-jdg2q\" (UID: \"b6f2ba03-746e-465c-bde2-c917129fb125\") " pod="openstack/neutron-659565f76f-jdg2q" Jan 21 11:26:43 crc kubenswrapper[4824]: I0121 11:26:43.699753 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6f2ba03-746e-465c-bde2-c917129fb125-combined-ca-bundle\") pod \"neutron-659565f76f-jdg2q\" (UID: \"b6f2ba03-746e-465c-bde2-c917129fb125\") " pod="openstack/neutron-659565f76f-jdg2q" Jan 21 11:26:43 crc kubenswrapper[4824]: I0121 11:26:43.699877 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/b6f2ba03-746e-465c-bde2-c917129fb125-ovndb-tls-certs\") pod \"neutron-659565f76f-jdg2q\" (UID: \"b6f2ba03-746e-465c-bde2-c917129fb125\") " pod="openstack/neutron-659565f76f-jdg2q" Jan 21 11:26:43 crc kubenswrapper[4824]: I0121 11:26:43.700008 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b6f2ba03-746e-465c-bde2-c917129fb125-internal-tls-certs\") pod \"neutron-659565f76f-jdg2q\" (UID: \"b6f2ba03-746e-465c-bde2-c917129fb125\") " pod="openstack/neutron-659565f76f-jdg2q" Jan 21 11:26:43 crc kubenswrapper[4824]: I0121 11:26:43.701064 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b6f2ba03-746e-465c-bde2-c917129fb125-public-tls-certs\") pod \"neutron-659565f76f-jdg2q\" (UID: \"b6f2ba03-746e-465c-bde2-c917129fb125\") " pod="openstack/neutron-659565f76f-jdg2q" Jan 21 11:26:43 crc kubenswrapper[4824]: I0121 11:26:43.701601 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/b6f2ba03-746e-465c-bde2-c917129fb125-config\") pod \"neutron-659565f76f-jdg2q\" (UID: \"b6f2ba03-746e-465c-bde2-c917129fb125\") " pod="openstack/neutron-659565f76f-jdg2q" Jan 21 11:26:43 crc kubenswrapper[4824]: I0121 11:26:43.708427 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/b6f2ba03-746e-465c-bde2-c917129fb125-httpd-config\") pod \"neutron-659565f76f-jdg2q\" (UID: \"b6f2ba03-746e-465c-bde2-c917129fb125\") " pod="openstack/neutron-659565f76f-jdg2q" Jan 21 11:26:43 crc kubenswrapper[4824]: I0121 11:26:43.711473 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c6dqp\" (UniqueName: \"kubernetes.io/projected/b6f2ba03-746e-465c-bde2-c917129fb125-kube-api-access-c6dqp\") pod \"neutron-659565f76f-jdg2q\" (UID: \"b6f2ba03-746e-465c-bde2-c917129fb125\") " pod="openstack/neutron-659565f76f-jdg2q" Jan 21 11:26:43 crc kubenswrapper[4824]: I0121 11:26:43.784317 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-79bfcdcfbd-s2v5h" event={"ID":"d9302662-76b1-4acc-b9e3-ffde925d34b2","Type":"ContainerStarted","Data":"eb6d7862490b4afbdbbe422c8984287dc94d6e4a39261b3846440fb920c542fb"} Jan 21 11:26:43 crc kubenswrapper[4824]: I0121 11:26:43.784399 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-79bfcdcfbd-s2v5h" Jan 21 11:26:43 crc kubenswrapper[4824]: I0121 11:26:43.785777 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-vbmkp" event={"ID":"bb4ea3d2-63c7-4829-b6b8-8112476493fa","Type":"ContainerStarted","Data":"25b08d9ba24d1cbae3803365bca42ff0e044f0f812800389a684ef8eb1362dbe"} Jan 21 11:26:43 crc kubenswrapper[4824]: I0121 11:26:43.786282 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-55f844cf75-vbmkp" Jan 21 11:26:43 crc kubenswrapper[4824]: I0121 11:26:43.787419 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"12776f5e-c372-41bd-91fe-ba7d0915050a","Type":"ContainerStarted","Data":"c66cf375354a975633a38f62d2bcee1e315728b454ed407911dcd2b8fbd32686"} Jan 21 11:26:43 crc kubenswrapper[4824]: I0121 11:26:43.789307 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bbefda93-dc56-43c5-b131-56af96ca832b","Type":"ContainerStarted","Data":"af014f3d948b12e938ae99b89aabd762100d2a622ac15f0606ad12c79347a128"} Jan 21 11:26:43 crc kubenswrapper[4824]: I0121 11:26:43.790797 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"46f0fa4b-479a-42c5-b9b2-3f30a079be48","Type":"ContainerStarted","Data":"77d449bfcfde658bccf9c84cc9cff76978535cecd3aaeeb5e306a982392e5cf4"} Jan 21 11:26:43 crc kubenswrapper[4824]: I0121 11:26:43.801537 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-79bfcdcfbd-s2v5h" podStartSLOduration=2.801521821 podStartE2EDuration="2.801521821s" podCreationTimestamp="2026-01-21 11:26:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:26:43.797167158 +0000 UTC m=+946.090196451" watchObservedRunningTime="2026-01-21 11:26:43.801521821 +0000 UTC m=+946.094551113" Jan 21 11:26:43 crc kubenswrapper[4824]: I0121 11:26:43.824540 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=3.824524189 podStartE2EDuration="3.824524189s" podCreationTimestamp="2026-01-21 11:26:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:26:43.819076985 +0000 UTC m=+946.112106277" watchObservedRunningTime="2026-01-21 11:26:43.824524189 +0000 UTC m=+946.117553481" Jan 21 11:26:43 crc kubenswrapper[4824]: I0121 11:26:43.836637 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-659565f76f-jdg2q" Jan 21 11:26:43 crc kubenswrapper[4824]: I0121 11:26:43.855318 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-55f844cf75-vbmkp" podStartSLOduration=2.855302655 podStartE2EDuration="2.855302655s" podCreationTimestamp="2026-01-21 11:26:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:26:43.853986693 +0000 UTC m=+946.147015985" watchObservedRunningTime="2026-01-21 11:26:43.855302655 +0000 UTC m=+946.148331947" Jan 21 11:26:43 crc kubenswrapper[4824]: I0121 11:26:43.858329 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=3.858318923 podStartE2EDuration="3.858318923s" podCreationTimestamp="2026-01-21 11:26:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:26:43.841981959 +0000 UTC m=+946.135011251" watchObservedRunningTime="2026-01-21 11:26:43.858318923 +0000 UTC m=+946.151348226" Jan 21 11:26:44 crc kubenswrapper[4824]: I0121 11:26:44.148327 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-6vgh6" Jan 21 11:26:44 crc kubenswrapper[4824]: I0121 11:26:44.208497 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/08b8c363-a8c1-4dfe-8fd1-1af2d055a305-scripts\") pod \"08b8c363-a8c1-4dfe-8fd1-1af2d055a305\" (UID: \"08b8c363-a8c1-4dfe-8fd1-1af2d055a305\") " Jan 21 11:26:44 crc kubenswrapper[4824]: I0121 11:26:44.208639 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/08b8c363-a8c1-4dfe-8fd1-1af2d055a305-config-data\") pod \"08b8c363-a8c1-4dfe-8fd1-1af2d055a305\" (UID: \"08b8c363-a8c1-4dfe-8fd1-1af2d055a305\") " Jan 21 11:26:44 crc kubenswrapper[4824]: I0121 11:26:44.208728 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/08b8c363-a8c1-4dfe-8fd1-1af2d055a305-logs\") pod \"08b8c363-a8c1-4dfe-8fd1-1af2d055a305\" (UID: \"08b8c363-a8c1-4dfe-8fd1-1af2d055a305\") " Jan 21 11:26:44 crc kubenswrapper[4824]: I0121 11:26:44.208774 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gkb9f\" (UniqueName: \"kubernetes.io/projected/08b8c363-a8c1-4dfe-8fd1-1af2d055a305-kube-api-access-gkb9f\") pod \"08b8c363-a8c1-4dfe-8fd1-1af2d055a305\" (UID: \"08b8c363-a8c1-4dfe-8fd1-1af2d055a305\") " Jan 21 11:26:44 crc kubenswrapper[4824]: I0121 11:26:44.208803 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08b8c363-a8c1-4dfe-8fd1-1af2d055a305-combined-ca-bundle\") pod \"08b8c363-a8c1-4dfe-8fd1-1af2d055a305\" (UID: \"08b8c363-a8c1-4dfe-8fd1-1af2d055a305\") " Jan 21 11:26:44 crc kubenswrapper[4824]: I0121 11:26:44.209989 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/08b8c363-a8c1-4dfe-8fd1-1af2d055a305-logs" (OuterVolumeSpecName: "logs") pod "08b8c363-a8c1-4dfe-8fd1-1af2d055a305" (UID: "08b8c363-a8c1-4dfe-8fd1-1af2d055a305"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:26:44 crc kubenswrapper[4824]: I0121 11:26:44.213224 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/08b8c363-a8c1-4dfe-8fd1-1af2d055a305-scripts" (OuterVolumeSpecName: "scripts") pod "08b8c363-a8c1-4dfe-8fd1-1af2d055a305" (UID: "08b8c363-a8c1-4dfe-8fd1-1af2d055a305"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:26:44 crc kubenswrapper[4824]: I0121 11:26:44.214142 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/08b8c363-a8c1-4dfe-8fd1-1af2d055a305-kube-api-access-gkb9f" (OuterVolumeSpecName: "kube-api-access-gkb9f") pod "08b8c363-a8c1-4dfe-8fd1-1af2d055a305" (UID: "08b8c363-a8c1-4dfe-8fd1-1af2d055a305"). InnerVolumeSpecName "kube-api-access-gkb9f". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:26:44 crc kubenswrapper[4824]: I0121 11:26:44.273616 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/08b8c363-a8c1-4dfe-8fd1-1af2d055a305-config-data" (OuterVolumeSpecName: "config-data") pod "08b8c363-a8c1-4dfe-8fd1-1af2d055a305" (UID: "08b8c363-a8c1-4dfe-8fd1-1af2d055a305"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:26:44 crc kubenswrapper[4824]: I0121 11:26:44.290128 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/08b8c363-a8c1-4dfe-8fd1-1af2d055a305-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "08b8c363-a8c1-4dfe-8fd1-1af2d055a305" (UID: "08b8c363-a8c1-4dfe-8fd1-1af2d055a305"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:26:44 crc kubenswrapper[4824]: I0121 11:26:44.312970 4824 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/08b8c363-a8c1-4dfe-8fd1-1af2d055a305-logs\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:44 crc kubenswrapper[4824]: I0121 11:26:44.313064 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gkb9f\" (UniqueName: \"kubernetes.io/projected/08b8c363-a8c1-4dfe-8fd1-1af2d055a305-kube-api-access-gkb9f\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:44 crc kubenswrapper[4824]: I0121 11:26:44.313119 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08b8c363-a8c1-4dfe-8fd1-1af2d055a305-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:44 crc kubenswrapper[4824]: I0121 11:26:44.313164 4824 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/08b8c363-a8c1-4dfe-8fd1-1af2d055a305-scripts\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:44 crc kubenswrapper[4824]: I0121 11:26:44.313208 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/08b8c363-a8c1-4dfe-8fd1-1af2d055a305-config-data\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:44 crc kubenswrapper[4824]: I0121 11:26:44.532711 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-659565f76f-jdg2q"] Jan 21 11:26:44 crc kubenswrapper[4824]: W0121 11:26:44.535620 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb6f2ba03_746e_465c_bde2_c917129fb125.slice/crio-e0b06e6954ac89ea6fb74405de5e37daee67b3757a4ece6c64e7e321e47249be WatchSource:0}: Error finding container e0b06e6954ac89ea6fb74405de5e37daee67b3757a4ece6c64e7e321e47249be: Status 404 returned error can't find the container with id e0b06e6954ac89ea6fb74405de5e37daee67b3757a4ece6c64e7e321e47249be Jan 21 11:26:44 crc kubenswrapper[4824]: I0121 11:26:44.799623 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-6vgh6" event={"ID":"08b8c363-a8c1-4dfe-8fd1-1af2d055a305","Type":"ContainerDied","Data":"97721ccfe8abdf2324794ac67debd67a8fc655d3d7ab538bf2cb4b35e35f8386"} Jan 21 11:26:44 crc kubenswrapper[4824]: I0121 11:26:44.799672 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="97721ccfe8abdf2324794ac67debd67a8fc655d3d7ab538bf2cb4b35e35f8386" Jan 21 11:26:44 crc kubenswrapper[4824]: I0121 11:26:44.799638 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-6vgh6" Jan 21 11:26:44 crc kubenswrapper[4824]: I0121 11:26:44.801196 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-659565f76f-jdg2q" event={"ID":"b6f2ba03-746e-465c-bde2-c917129fb125","Type":"ContainerStarted","Data":"966a3b02cb0a14177696e58a750fbeec6fd6ca5d319be90ce60f05fc0315f36f"} Jan 21 11:26:44 crc kubenswrapper[4824]: I0121 11:26:44.801233 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-659565f76f-jdg2q" event={"ID":"b6f2ba03-746e-465c-bde2-c917129fb125","Type":"ContainerStarted","Data":"e0b06e6954ac89ea6fb74405de5e37daee67b3757a4ece6c64e7e321e47249be"} Jan 21 11:26:44 crc kubenswrapper[4824]: I0121 11:26:44.803552 4824 generic.go:334] "Generic (PLEG): container finished" podID="cadf2d38-24f9-43ab-8e29-803d97a0ff06" containerID="5dcf0246f4ab2f2f4f3230d6a47ec6ac1102fcb34b4edb4d89a73c99d792f5ff" exitCode=0 Jan 21 11:26:44 crc kubenswrapper[4824]: I0121 11:26:44.803602 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-jrsn8" event={"ID":"cadf2d38-24f9-43ab-8e29-803d97a0ff06","Type":"ContainerDied","Data":"5dcf0246f4ab2f2f4f3230d6a47ec6ac1102fcb34b4edb4d89a73c99d792f5ff"} Jan 21 11:26:44 crc kubenswrapper[4824]: I0121 11:26:44.805507 4824 generic.go:334] "Generic (PLEG): container finished" podID="a7e75011-d633-4ccc-951c-d019cb2100f9" containerID="2534da156868c09c8dd8dee0424c34fab5c7b1401695a79ef623bde9f1ef00f0" exitCode=0 Jan 21 11:26:44 crc kubenswrapper[4824]: I0121 11:26:44.805620 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-5l2gg" event={"ID":"a7e75011-d633-4ccc-951c-d019cb2100f9","Type":"ContainerDied","Data":"2534da156868c09c8dd8dee0424c34fab5c7b1401695a79ef623bde9f1ef00f0"} Jan 21 11:26:44 crc kubenswrapper[4824]: I0121 11:26:44.924709 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-8654bbf998-p8crm"] Jan 21 11:26:44 crc kubenswrapper[4824]: E0121 11:26:44.925076 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08b8c363-a8c1-4dfe-8fd1-1af2d055a305" containerName="placement-db-sync" Jan 21 11:26:44 crc kubenswrapper[4824]: I0121 11:26:44.925095 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="08b8c363-a8c1-4dfe-8fd1-1af2d055a305" containerName="placement-db-sync" Jan 21 11:26:44 crc kubenswrapper[4824]: I0121 11:26:44.925286 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="08b8c363-a8c1-4dfe-8fd1-1af2d055a305" containerName="placement-db-sync" Jan 21 11:26:44 crc kubenswrapper[4824]: I0121 11:26:44.926251 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-8654bbf998-p8crm" Jan 21 11:26:44 crc kubenswrapper[4824]: I0121 11:26:44.929739 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Jan 21 11:26:44 crc kubenswrapper[4824]: I0121 11:26:44.929986 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Jan 21 11:26:44 crc kubenswrapper[4824]: I0121 11:26:44.930130 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Jan 21 11:26:44 crc kubenswrapper[4824]: I0121 11:26:44.930203 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-n8hrk" Jan 21 11:26:44 crc kubenswrapper[4824]: I0121 11:26:44.930232 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Jan 21 11:26:44 crc kubenswrapper[4824]: I0121 11:26:44.933406 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-8654bbf998-p8crm"] Jan 21 11:26:45 crc kubenswrapper[4824]: I0121 11:26:45.027892 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c23477ba-5a41-44da-a8b7-6606595b9726-combined-ca-bundle\") pod \"placement-8654bbf998-p8crm\" (UID: \"c23477ba-5a41-44da-a8b7-6606595b9726\") " pod="openstack/placement-8654bbf998-p8crm" Jan 21 11:26:45 crc kubenswrapper[4824]: I0121 11:26:45.027993 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c23477ba-5a41-44da-a8b7-6606595b9726-scripts\") pod \"placement-8654bbf998-p8crm\" (UID: \"c23477ba-5a41-44da-a8b7-6606595b9726\") " pod="openstack/placement-8654bbf998-p8crm" Jan 21 11:26:45 crc kubenswrapper[4824]: I0121 11:26:45.028022 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c23477ba-5a41-44da-a8b7-6606595b9726-public-tls-certs\") pod \"placement-8654bbf998-p8crm\" (UID: \"c23477ba-5a41-44da-a8b7-6606595b9726\") " pod="openstack/placement-8654bbf998-p8crm" Jan 21 11:26:45 crc kubenswrapper[4824]: I0121 11:26:45.028058 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c23477ba-5a41-44da-a8b7-6606595b9726-config-data\") pod \"placement-8654bbf998-p8crm\" (UID: \"c23477ba-5a41-44da-a8b7-6606595b9726\") " pod="openstack/placement-8654bbf998-p8crm" Jan 21 11:26:45 crc kubenswrapper[4824]: I0121 11:26:45.028101 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c23477ba-5a41-44da-a8b7-6606595b9726-logs\") pod \"placement-8654bbf998-p8crm\" (UID: \"c23477ba-5a41-44da-a8b7-6606595b9726\") " pod="openstack/placement-8654bbf998-p8crm" Jan 21 11:26:45 crc kubenswrapper[4824]: I0121 11:26:45.028120 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dlbpz\" (UniqueName: \"kubernetes.io/projected/c23477ba-5a41-44da-a8b7-6606595b9726-kube-api-access-dlbpz\") pod \"placement-8654bbf998-p8crm\" (UID: \"c23477ba-5a41-44da-a8b7-6606595b9726\") " pod="openstack/placement-8654bbf998-p8crm" Jan 21 11:26:45 crc kubenswrapper[4824]: I0121 11:26:45.028140 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c23477ba-5a41-44da-a8b7-6606595b9726-internal-tls-certs\") pod \"placement-8654bbf998-p8crm\" (UID: \"c23477ba-5a41-44da-a8b7-6606595b9726\") " pod="openstack/placement-8654bbf998-p8crm" Jan 21 11:26:45 crc kubenswrapper[4824]: I0121 11:26:45.129523 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c23477ba-5a41-44da-a8b7-6606595b9726-scripts\") pod \"placement-8654bbf998-p8crm\" (UID: \"c23477ba-5a41-44da-a8b7-6606595b9726\") " pod="openstack/placement-8654bbf998-p8crm" Jan 21 11:26:45 crc kubenswrapper[4824]: I0121 11:26:45.129569 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c23477ba-5a41-44da-a8b7-6606595b9726-public-tls-certs\") pod \"placement-8654bbf998-p8crm\" (UID: \"c23477ba-5a41-44da-a8b7-6606595b9726\") " pod="openstack/placement-8654bbf998-p8crm" Jan 21 11:26:45 crc kubenswrapper[4824]: I0121 11:26:45.129612 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c23477ba-5a41-44da-a8b7-6606595b9726-config-data\") pod \"placement-8654bbf998-p8crm\" (UID: \"c23477ba-5a41-44da-a8b7-6606595b9726\") " pod="openstack/placement-8654bbf998-p8crm" Jan 21 11:26:45 crc kubenswrapper[4824]: I0121 11:26:45.129664 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c23477ba-5a41-44da-a8b7-6606595b9726-logs\") pod \"placement-8654bbf998-p8crm\" (UID: \"c23477ba-5a41-44da-a8b7-6606595b9726\") " pod="openstack/placement-8654bbf998-p8crm" Jan 21 11:26:45 crc kubenswrapper[4824]: I0121 11:26:45.129696 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dlbpz\" (UniqueName: \"kubernetes.io/projected/c23477ba-5a41-44da-a8b7-6606595b9726-kube-api-access-dlbpz\") pod \"placement-8654bbf998-p8crm\" (UID: \"c23477ba-5a41-44da-a8b7-6606595b9726\") " pod="openstack/placement-8654bbf998-p8crm" Jan 21 11:26:45 crc kubenswrapper[4824]: I0121 11:26:45.129730 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c23477ba-5a41-44da-a8b7-6606595b9726-internal-tls-certs\") pod \"placement-8654bbf998-p8crm\" (UID: \"c23477ba-5a41-44da-a8b7-6606595b9726\") " pod="openstack/placement-8654bbf998-p8crm" Jan 21 11:26:45 crc kubenswrapper[4824]: I0121 11:26:45.129794 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c23477ba-5a41-44da-a8b7-6606595b9726-combined-ca-bundle\") pod \"placement-8654bbf998-p8crm\" (UID: \"c23477ba-5a41-44da-a8b7-6606595b9726\") " pod="openstack/placement-8654bbf998-p8crm" Jan 21 11:26:45 crc kubenswrapper[4824]: I0121 11:26:45.130587 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c23477ba-5a41-44da-a8b7-6606595b9726-logs\") pod \"placement-8654bbf998-p8crm\" (UID: \"c23477ba-5a41-44da-a8b7-6606595b9726\") " pod="openstack/placement-8654bbf998-p8crm" Jan 21 11:26:45 crc kubenswrapper[4824]: I0121 11:26:45.133684 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c23477ba-5a41-44da-a8b7-6606595b9726-config-data\") pod \"placement-8654bbf998-p8crm\" (UID: \"c23477ba-5a41-44da-a8b7-6606595b9726\") " pod="openstack/placement-8654bbf998-p8crm" Jan 21 11:26:45 crc kubenswrapper[4824]: I0121 11:26:45.136202 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c23477ba-5a41-44da-a8b7-6606595b9726-public-tls-certs\") pod \"placement-8654bbf998-p8crm\" (UID: \"c23477ba-5a41-44da-a8b7-6606595b9726\") " pod="openstack/placement-8654bbf998-p8crm" Jan 21 11:26:45 crc kubenswrapper[4824]: I0121 11:26:45.139329 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c23477ba-5a41-44da-a8b7-6606595b9726-scripts\") pod \"placement-8654bbf998-p8crm\" (UID: \"c23477ba-5a41-44da-a8b7-6606595b9726\") " pod="openstack/placement-8654bbf998-p8crm" Jan 21 11:26:45 crc kubenswrapper[4824]: I0121 11:26:45.142389 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c23477ba-5a41-44da-a8b7-6606595b9726-internal-tls-certs\") pod \"placement-8654bbf998-p8crm\" (UID: \"c23477ba-5a41-44da-a8b7-6606595b9726\") " pod="openstack/placement-8654bbf998-p8crm" Jan 21 11:26:45 crc kubenswrapper[4824]: I0121 11:26:45.144400 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dlbpz\" (UniqueName: \"kubernetes.io/projected/c23477ba-5a41-44da-a8b7-6606595b9726-kube-api-access-dlbpz\") pod \"placement-8654bbf998-p8crm\" (UID: \"c23477ba-5a41-44da-a8b7-6606595b9726\") " pod="openstack/placement-8654bbf998-p8crm" Jan 21 11:26:45 crc kubenswrapper[4824]: I0121 11:26:45.157369 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c23477ba-5a41-44da-a8b7-6606595b9726-combined-ca-bundle\") pod \"placement-8654bbf998-p8crm\" (UID: \"c23477ba-5a41-44da-a8b7-6606595b9726\") " pod="openstack/placement-8654bbf998-p8crm" Jan 21 11:26:45 crc kubenswrapper[4824]: I0121 11:26:45.250627 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-8654bbf998-p8crm" Jan 21 11:26:45 crc kubenswrapper[4824]: I0121 11:26:45.719250 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-74f6bcbc87-sj79w" podUID="c4a81903-5e1f-4d9c-bacf-60639069db18" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.126:5353: i/o timeout" Jan 21 11:26:45 crc kubenswrapper[4824]: I0121 11:26:45.814872 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-659565f76f-jdg2q" event={"ID":"b6f2ba03-746e-465c-bde2-c917129fb125","Type":"ContainerStarted","Data":"4b3e1a4d84baaeed6c9761ff58fa7b697d120977b82c7b9024880b67d908aeed"} Jan 21 11:26:45 crc kubenswrapper[4824]: I0121 11:26:45.815790 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-659565f76f-jdg2q" Jan 21 11:26:45 crc kubenswrapper[4824]: I0121 11:26:45.833828 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-659565f76f-jdg2q" podStartSLOduration=2.833813488 podStartE2EDuration="2.833813488s" podCreationTimestamp="2026-01-21 11:26:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:26:45.827439735 +0000 UTC m=+948.120469027" watchObservedRunningTime="2026-01-21 11:26:45.833813488 +0000 UTC m=+948.126842780" Jan 21 11:26:46 crc kubenswrapper[4824]: I0121 11:26:46.065653 4824 patch_prober.go:28] interesting pod/machine-config-daemon-zqjg9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 21 11:26:46 crc kubenswrapper[4824]: I0121 11:26:46.065698 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 21 11:26:46 crc kubenswrapper[4824]: I0121 11:26:46.406551 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-5l2gg" Jan 21 11:26:46 crc kubenswrapper[4824]: I0121 11:26:46.553440 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8lj6r\" (UniqueName: \"kubernetes.io/projected/a7e75011-d633-4ccc-951c-d019cb2100f9-kube-api-access-8lj6r\") pod \"a7e75011-d633-4ccc-951c-d019cb2100f9\" (UID: \"a7e75011-d633-4ccc-951c-d019cb2100f9\") " Jan 21 11:26:46 crc kubenswrapper[4824]: I0121 11:26:46.553483 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a7e75011-d633-4ccc-951c-d019cb2100f9-config-data\") pod \"a7e75011-d633-4ccc-951c-d019cb2100f9\" (UID: \"a7e75011-d633-4ccc-951c-d019cb2100f9\") " Jan 21 11:26:46 crc kubenswrapper[4824]: I0121 11:26:46.553573 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a7e75011-d633-4ccc-951c-d019cb2100f9-scripts\") pod \"a7e75011-d633-4ccc-951c-d019cb2100f9\" (UID: \"a7e75011-d633-4ccc-951c-d019cb2100f9\") " Jan 21 11:26:46 crc kubenswrapper[4824]: I0121 11:26:46.553605 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a7e75011-d633-4ccc-951c-d019cb2100f9-fernet-keys\") pod \"a7e75011-d633-4ccc-951c-d019cb2100f9\" (UID: \"a7e75011-d633-4ccc-951c-d019cb2100f9\") " Jan 21 11:26:46 crc kubenswrapper[4824]: I0121 11:26:46.553668 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/a7e75011-d633-4ccc-951c-d019cb2100f9-credential-keys\") pod \"a7e75011-d633-4ccc-951c-d019cb2100f9\" (UID: \"a7e75011-d633-4ccc-951c-d019cb2100f9\") " Jan 21 11:26:46 crc kubenswrapper[4824]: I0121 11:26:46.553688 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7e75011-d633-4ccc-951c-d019cb2100f9-combined-ca-bundle\") pod \"a7e75011-d633-4ccc-951c-d019cb2100f9\" (UID: \"a7e75011-d633-4ccc-951c-d019cb2100f9\") " Jan 21 11:26:46 crc kubenswrapper[4824]: I0121 11:26:46.561259 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a7e75011-d633-4ccc-951c-d019cb2100f9-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "a7e75011-d633-4ccc-951c-d019cb2100f9" (UID: "a7e75011-d633-4ccc-951c-d019cb2100f9"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:26:46 crc kubenswrapper[4824]: I0121 11:26:46.561425 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a7e75011-d633-4ccc-951c-d019cb2100f9-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "a7e75011-d633-4ccc-951c-d019cb2100f9" (UID: "a7e75011-d633-4ccc-951c-d019cb2100f9"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:26:46 crc kubenswrapper[4824]: I0121 11:26:46.566203 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a7e75011-d633-4ccc-951c-d019cb2100f9-scripts" (OuterVolumeSpecName: "scripts") pod "a7e75011-d633-4ccc-951c-d019cb2100f9" (UID: "a7e75011-d633-4ccc-951c-d019cb2100f9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:26:46 crc kubenswrapper[4824]: I0121 11:26:46.568491 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a7e75011-d633-4ccc-951c-d019cb2100f9-kube-api-access-8lj6r" (OuterVolumeSpecName: "kube-api-access-8lj6r") pod "a7e75011-d633-4ccc-951c-d019cb2100f9" (UID: "a7e75011-d633-4ccc-951c-d019cb2100f9"). InnerVolumeSpecName "kube-api-access-8lj6r". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:26:46 crc kubenswrapper[4824]: I0121 11:26:46.576268 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a7e75011-d633-4ccc-951c-d019cb2100f9-config-data" (OuterVolumeSpecName: "config-data") pod "a7e75011-d633-4ccc-951c-d019cb2100f9" (UID: "a7e75011-d633-4ccc-951c-d019cb2100f9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:26:46 crc kubenswrapper[4824]: I0121 11:26:46.594180 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a7e75011-d633-4ccc-951c-d019cb2100f9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a7e75011-d633-4ccc-951c-d019cb2100f9" (UID: "a7e75011-d633-4ccc-951c-d019cb2100f9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:26:46 crc kubenswrapper[4824]: I0121 11:26:46.655225 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7e75011-d633-4ccc-951c-d019cb2100f9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:46 crc kubenswrapper[4824]: I0121 11:26:46.655253 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8lj6r\" (UniqueName: \"kubernetes.io/projected/a7e75011-d633-4ccc-951c-d019cb2100f9-kube-api-access-8lj6r\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:46 crc kubenswrapper[4824]: I0121 11:26:46.655265 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a7e75011-d633-4ccc-951c-d019cb2100f9-config-data\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:46 crc kubenswrapper[4824]: I0121 11:26:46.655273 4824 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a7e75011-d633-4ccc-951c-d019cb2100f9-scripts\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:46 crc kubenswrapper[4824]: I0121 11:26:46.655280 4824 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a7e75011-d633-4ccc-951c-d019cb2100f9-fernet-keys\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:46 crc kubenswrapper[4824]: I0121 11:26:46.655287 4824 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/a7e75011-d633-4ccc-951c-d019cb2100f9-credential-keys\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:46 crc kubenswrapper[4824]: I0121 11:26:46.822265 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-5l2gg" event={"ID":"a7e75011-d633-4ccc-951c-d019cb2100f9","Type":"ContainerDied","Data":"f73eda2acba113b106a0d01407072387e56c3b86b9c9c6135b58a1185903e93b"} Jan 21 11:26:46 crc kubenswrapper[4824]: I0121 11:26:46.822311 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f73eda2acba113b106a0d01407072387e56c3b86b9c9c6135b58a1185903e93b" Jan 21 11:26:46 crc kubenswrapper[4824]: I0121 11:26:46.822429 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-5l2gg" Jan 21 11:26:46 crc kubenswrapper[4824]: I0121 11:26:46.984773 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-6c5fc769c9-spmzr"] Jan 21 11:26:46 crc kubenswrapper[4824]: E0121 11:26:46.985408 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7e75011-d633-4ccc-951c-d019cb2100f9" containerName="keystone-bootstrap" Jan 21 11:26:46 crc kubenswrapper[4824]: I0121 11:26:46.985422 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7e75011-d633-4ccc-951c-d019cb2100f9" containerName="keystone-bootstrap" Jan 21 11:26:46 crc kubenswrapper[4824]: I0121 11:26:46.985585 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="a7e75011-d633-4ccc-951c-d019cb2100f9" containerName="keystone-bootstrap" Jan 21 11:26:46 crc kubenswrapper[4824]: I0121 11:26:46.986065 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-6c5fc769c9-spmzr" Jan 21 11:26:46 crc kubenswrapper[4824]: I0121 11:26:46.989312 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Jan 21 11:26:46 crc kubenswrapper[4824]: I0121 11:26:46.989476 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-qxhzv" Jan 21 11:26:46 crc kubenswrapper[4824]: I0121 11:26:46.989604 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Jan 21 11:26:46 crc kubenswrapper[4824]: I0121 11:26:46.989718 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Jan 21 11:26:46 crc kubenswrapper[4824]: I0121 11:26:46.989818 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Jan 21 11:26:46 crc kubenswrapper[4824]: I0121 11:26:46.990271 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Jan 21 11:26:46 crc kubenswrapper[4824]: I0121 11:26:46.998563 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-6c5fc769c9-spmzr"] Jan 21 11:26:47 crc kubenswrapper[4824]: I0121 11:26:47.165329 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7d943b5e-d05f-445a-888d-7ebe340468b9-public-tls-certs\") pod \"keystone-6c5fc769c9-spmzr\" (UID: \"7d943b5e-d05f-445a-888d-7ebe340468b9\") " pod="openstack/keystone-6c5fc769c9-spmzr" Jan 21 11:26:47 crc kubenswrapper[4824]: I0121 11:26:47.165369 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7d943b5e-d05f-445a-888d-7ebe340468b9-internal-tls-certs\") pod \"keystone-6c5fc769c9-spmzr\" (UID: \"7d943b5e-d05f-445a-888d-7ebe340468b9\") " pod="openstack/keystone-6c5fc769c9-spmzr" Jan 21 11:26:47 crc kubenswrapper[4824]: I0121 11:26:47.165486 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/7d943b5e-d05f-445a-888d-7ebe340468b9-fernet-keys\") pod \"keystone-6c5fc769c9-spmzr\" (UID: \"7d943b5e-d05f-445a-888d-7ebe340468b9\") " pod="openstack/keystone-6c5fc769c9-spmzr" Jan 21 11:26:47 crc kubenswrapper[4824]: I0121 11:26:47.165560 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7d943b5e-d05f-445a-888d-7ebe340468b9-scripts\") pod \"keystone-6c5fc769c9-spmzr\" (UID: \"7d943b5e-d05f-445a-888d-7ebe340468b9\") " pod="openstack/keystone-6c5fc769c9-spmzr" Jan 21 11:26:47 crc kubenswrapper[4824]: I0121 11:26:47.165594 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d943b5e-d05f-445a-888d-7ebe340468b9-combined-ca-bundle\") pod \"keystone-6c5fc769c9-spmzr\" (UID: \"7d943b5e-d05f-445a-888d-7ebe340468b9\") " pod="openstack/keystone-6c5fc769c9-spmzr" Jan 21 11:26:47 crc kubenswrapper[4824]: I0121 11:26:47.165653 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/7d943b5e-d05f-445a-888d-7ebe340468b9-credential-keys\") pod \"keystone-6c5fc769c9-spmzr\" (UID: \"7d943b5e-d05f-445a-888d-7ebe340468b9\") " pod="openstack/keystone-6c5fc769c9-spmzr" Jan 21 11:26:47 crc kubenswrapper[4824]: I0121 11:26:47.165679 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v487x\" (UniqueName: \"kubernetes.io/projected/7d943b5e-d05f-445a-888d-7ebe340468b9-kube-api-access-v487x\") pod \"keystone-6c5fc769c9-spmzr\" (UID: \"7d943b5e-d05f-445a-888d-7ebe340468b9\") " pod="openstack/keystone-6c5fc769c9-spmzr" Jan 21 11:26:47 crc kubenswrapper[4824]: I0121 11:26:47.165719 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7d943b5e-d05f-445a-888d-7ebe340468b9-config-data\") pod \"keystone-6c5fc769c9-spmzr\" (UID: \"7d943b5e-d05f-445a-888d-7ebe340468b9\") " pod="openstack/keystone-6c5fc769c9-spmzr" Jan 21 11:26:47 crc kubenswrapper[4824]: I0121 11:26:47.267414 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d943b5e-d05f-445a-888d-7ebe340468b9-combined-ca-bundle\") pod \"keystone-6c5fc769c9-spmzr\" (UID: \"7d943b5e-d05f-445a-888d-7ebe340468b9\") " pod="openstack/keystone-6c5fc769c9-spmzr" Jan 21 11:26:47 crc kubenswrapper[4824]: I0121 11:26:47.267499 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/7d943b5e-d05f-445a-888d-7ebe340468b9-credential-keys\") pod \"keystone-6c5fc769c9-spmzr\" (UID: \"7d943b5e-d05f-445a-888d-7ebe340468b9\") " pod="openstack/keystone-6c5fc769c9-spmzr" Jan 21 11:26:47 crc kubenswrapper[4824]: I0121 11:26:47.267538 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v487x\" (UniqueName: \"kubernetes.io/projected/7d943b5e-d05f-445a-888d-7ebe340468b9-kube-api-access-v487x\") pod \"keystone-6c5fc769c9-spmzr\" (UID: \"7d943b5e-d05f-445a-888d-7ebe340468b9\") " pod="openstack/keystone-6c5fc769c9-spmzr" Jan 21 11:26:47 crc kubenswrapper[4824]: I0121 11:26:47.267563 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7d943b5e-d05f-445a-888d-7ebe340468b9-config-data\") pod \"keystone-6c5fc769c9-spmzr\" (UID: \"7d943b5e-d05f-445a-888d-7ebe340468b9\") " pod="openstack/keystone-6c5fc769c9-spmzr" Jan 21 11:26:47 crc kubenswrapper[4824]: I0121 11:26:47.267633 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7d943b5e-d05f-445a-888d-7ebe340468b9-public-tls-certs\") pod \"keystone-6c5fc769c9-spmzr\" (UID: \"7d943b5e-d05f-445a-888d-7ebe340468b9\") " pod="openstack/keystone-6c5fc769c9-spmzr" Jan 21 11:26:47 crc kubenswrapper[4824]: I0121 11:26:47.267650 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7d943b5e-d05f-445a-888d-7ebe340468b9-internal-tls-certs\") pod \"keystone-6c5fc769c9-spmzr\" (UID: \"7d943b5e-d05f-445a-888d-7ebe340468b9\") " pod="openstack/keystone-6c5fc769c9-spmzr" Jan 21 11:26:47 crc kubenswrapper[4824]: I0121 11:26:47.267748 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/7d943b5e-d05f-445a-888d-7ebe340468b9-fernet-keys\") pod \"keystone-6c5fc769c9-spmzr\" (UID: \"7d943b5e-d05f-445a-888d-7ebe340468b9\") " pod="openstack/keystone-6c5fc769c9-spmzr" Jan 21 11:26:47 crc kubenswrapper[4824]: I0121 11:26:47.267807 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7d943b5e-d05f-445a-888d-7ebe340468b9-scripts\") pod \"keystone-6c5fc769c9-spmzr\" (UID: \"7d943b5e-d05f-445a-888d-7ebe340468b9\") " pod="openstack/keystone-6c5fc769c9-spmzr" Jan 21 11:26:47 crc kubenswrapper[4824]: I0121 11:26:47.272072 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/7d943b5e-d05f-445a-888d-7ebe340468b9-credential-keys\") pod \"keystone-6c5fc769c9-spmzr\" (UID: \"7d943b5e-d05f-445a-888d-7ebe340468b9\") " pod="openstack/keystone-6c5fc769c9-spmzr" Jan 21 11:26:47 crc kubenswrapper[4824]: I0121 11:26:47.273229 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7d943b5e-d05f-445a-888d-7ebe340468b9-public-tls-certs\") pod \"keystone-6c5fc769c9-spmzr\" (UID: \"7d943b5e-d05f-445a-888d-7ebe340468b9\") " pod="openstack/keystone-6c5fc769c9-spmzr" Jan 21 11:26:47 crc kubenswrapper[4824]: I0121 11:26:47.273669 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7d943b5e-d05f-445a-888d-7ebe340468b9-scripts\") pod \"keystone-6c5fc769c9-spmzr\" (UID: \"7d943b5e-d05f-445a-888d-7ebe340468b9\") " pod="openstack/keystone-6c5fc769c9-spmzr" Jan 21 11:26:47 crc kubenswrapper[4824]: I0121 11:26:47.273752 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7d943b5e-d05f-445a-888d-7ebe340468b9-config-data\") pod \"keystone-6c5fc769c9-spmzr\" (UID: \"7d943b5e-d05f-445a-888d-7ebe340468b9\") " pod="openstack/keystone-6c5fc769c9-spmzr" Jan 21 11:26:47 crc kubenswrapper[4824]: I0121 11:26:47.274029 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7d943b5e-d05f-445a-888d-7ebe340468b9-internal-tls-certs\") pod \"keystone-6c5fc769c9-spmzr\" (UID: \"7d943b5e-d05f-445a-888d-7ebe340468b9\") " pod="openstack/keystone-6c5fc769c9-spmzr" Jan 21 11:26:47 crc kubenswrapper[4824]: I0121 11:26:47.274171 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d943b5e-d05f-445a-888d-7ebe340468b9-combined-ca-bundle\") pod \"keystone-6c5fc769c9-spmzr\" (UID: \"7d943b5e-d05f-445a-888d-7ebe340468b9\") " pod="openstack/keystone-6c5fc769c9-spmzr" Jan 21 11:26:47 crc kubenswrapper[4824]: I0121 11:26:47.279255 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/7d943b5e-d05f-445a-888d-7ebe340468b9-fernet-keys\") pod \"keystone-6c5fc769c9-spmzr\" (UID: \"7d943b5e-d05f-445a-888d-7ebe340468b9\") " pod="openstack/keystone-6c5fc769c9-spmzr" Jan 21 11:26:47 crc kubenswrapper[4824]: I0121 11:26:47.281428 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v487x\" (UniqueName: \"kubernetes.io/projected/7d943b5e-d05f-445a-888d-7ebe340468b9-kube-api-access-v487x\") pod \"keystone-6c5fc769c9-spmzr\" (UID: \"7d943b5e-d05f-445a-888d-7ebe340468b9\") " pod="openstack/keystone-6c5fc769c9-spmzr" Jan 21 11:26:47 crc kubenswrapper[4824]: I0121 11:26:47.304929 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-6c5fc769c9-spmzr" Jan 21 11:26:47 crc kubenswrapper[4824]: I0121 11:26:47.844442 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-jrsn8" event={"ID":"cadf2d38-24f9-43ab-8e29-803d97a0ff06","Type":"ContainerDied","Data":"f8ce50a0aaa511cfcd5c91e987379f42c2650248768a77a65a4355ed457a06c8"} Jan 21 11:26:47 crc kubenswrapper[4824]: I0121 11:26:47.844627 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f8ce50a0aaa511cfcd5c91e987379f42c2650248768a77a65a4355ed457a06c8" Jan 21 11:26:47 crc kubenswrapper[4824]: I0121 11:26:47.879314 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-jrsn8" Jan 21 11:26:47 crc kubenswrapper[4824]: I0121 11:26:47.980836 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xrdwp\" (UniqueName: \"kubernetes.io/projected/cadf2d38-24f9-43ab-8e29-803d97a0ff06-kube-api-access-xrdwp\") pod \"cadf2d38-24f9-43ab-8e29-803d97a0ff06\" (UID: \"cadf2d38-24f9-43ab-8e29-803d97a0ff06\") " Jan 21 11:26:47 crc kubenswrapper[4824]: I0121 11:26:47.981030 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cadf2d38-24f9-43ab-8e29-803d97a0ff06-config-data\") pod \"cadf2d38-24f9-43ab-8e29-803d97a0ff06\" (UID: \"cadf2d38-24f9-43ab-8e29-803d97a0ff06\") " Jan 21 11:26:47 crc kubenswrapper[4824]: I0121 11:26:47.981075 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cadf2d38-24f9-43ab-8e29-803d97a0ff06-combined-ca-bundle\") pod \"cadf2d38-24f9-43ab-8e29-803d97a0ff06\" (UID: \"cadf2d38-24f9-43ab-8e29-803d97a0ff06\") " Jan 21 11:26:47 crc kubenswrapper[4824]: I0121 11:26:47.984410 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cadf2d38-24f9-43ab-8e29-803d97a0ff06-kube-api-access-xrdwp" (OuterVolumeSpecName: "kube-api-access-xrdwp") pod "cadf2d38-24f9-43ab-8e29-803d97a0ff06" (UID: "cadf2d38-24f9-43ab-8e29-803d97a0ff06"). InnerVolumeSpecName "kube-api-access-xrdwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:26:48 crc kubenswrapper[4824]: I0121 11:26:48.001110 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cadf2d38-24f9-43ab-8e29-803d97a0ff06-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cadf2d38-24f9-43ab-8e29-803d97a0ff06" (UID: "cadf2d38-24f9-43ab-8e29-803d97a0ff06"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:26:48 crc kubenswrapper[4824]: I0121 11:26:48.027610 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cadf2d38-24f9-43ab-8e29-803d97a0ff06-config-data" (OuterVolumeSpecName: "config-data") pod "cadf2d38-24f9-43ab-8e29-803d97a0ff06" (UID: "cadf2d38-24f9-43ab-8e29-803d97a0ff06"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:26:48 crc kubenswrapper[4824]: I0121 11:26:48.082522 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cadf2d38-24f9-43ab-8e29-803d97a0ff06-config-data\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:48 crc kubenswrapper[4824]: I0121 11:26:48.082543 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cadf2d38-24f9-43ab-8e29-803d97a0ff06-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:48 crc kubenswrapper[4824]: I0121 11:26:48.082552 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xrdwp\" (UniqueName: \"kubernetes.io/projected/cadf2d38-24f9-43ab-8e29-803d97a0ff06-kube-api-access-xrdwp\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:48 crc kubenswrapper[4824]: I0121 11:26:48.150075 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-8654bbf998-p8crm"] Jan 21 11:26:48 crc kubenswrapper[4824]: W0121 11:26:48.158563 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7d943b5e_d05f_445a_888d_7ebe340468b9.slice/crio-779d43c0e7120424033bc4414b128725901d8d0023671ffa45ed85a0227c9c69 WatchSource:0}: Error finding container 779d43c0e7120424033bc4414b128725901d8d0023671ffa45ed85a0227c9c69: Status 404 returned error can't find the container with id 779d43c0e7120424033bc4414b128725901d8d0023671ffa45ed85a0227c9c69 Jan 21 11:26:48 crc kubenswrapper[4824]: I0121 11:26:48.159922 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-6c5fc769c9-spmzr"] Jan 21 11:26:48 crc kubenswrapper[4824]: I0121 11:26:48.174450 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-dn2np"] Jan 21 11:26:48 crc kubenswrapper[4824]: E0121 11:26:48.174757 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cadf2d38-24f9-43ab-8e29-803d97a0ff06" containerName="heat-db-sync" Jan 21 11:26:48 crc kubenswrapper[4824]: I0121 11:26:48.174775 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="cadf2d38-24f9-43ab-8e29-803d97a0ff06" containerName="heat-db-sync" Jan 21 11:26:48 crc kubenswrapper[4824]: I0121 11:26:48.174965 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="cadf2d38-24f9-43ab-8e29-803d97a0ff06" containerName="heat-db-sync" Jan 21 11:26:48 crc kubenswrapper[4824]: I0121 11:26:48.176674 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dn2np" Jan 21 11:26:48 crc kubenswrapper[4824]: I0121 11:26:48.186275 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-dn2np"] Jan 21 11:26:48 crc kubenswrapper[4824]: I0121 11:26:48.285635 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d8c8cff-4340-4eb4-94b7-062968a14848-catalog-content\") pod \"certified-operators-dn2np\" (UID: \"1d8c8cff-4340-4eb4-94b7-062968a14848\") " pod="openshift-marketplace/certified-operators-dn2np" Jan 21 11:26:48 crc kubenswrapper[4824]: I0121 11:26:48.285900 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d8c8cff-4340-4eb4-94b7-062968a14848-utilities\") pod \"certified-operators-dn2np\" (UID: \"1d8c8cff-4340-4eb4-94b7-062968a14848\") " pod="openshift-marketplace/certified-operators-dn2np" Jan 21 11:26:48 crc kubenswrapper[4824]: I0121 11:26:48.286007 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f9j78\" (UniqueName: \"kubernetes.io/projected/1d8c8cff-4340-4eb4-94b7-062968a14848-kube-api-access-f9j78\") pod \"certified-operators-dn2np\" (UID: \"1d8c8cff-4340-4eb4-94b7-062968a14848\") " pod="openshift-marketplace/certified-operators-dn2np" Jan 21 11:26:48 crc kubenswrapper[4824]: I0121 11:26:48.387435 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d8c8cff-4340-4eb4-94b7-062968a14848-catalog-content\") pod \"certified-operators-dn2np\" (UID: \"1d8c8cff-4340-4eb4-94b7-062968a14848\") " pod="openshift-marketplace/certified-operators-dn2np" Jan 21 11:26:48 crc kubenswrapper[4824]: I0121 11:26:48.387524 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d8c8cff-4340-4eb4-94b7-062968a14848-utilities\") pod \"certified-operators-dn2np\" (UID: \"1d8c8cff-4340-4eb4-94b7-062968a14848\") " pod="openshift-marketplace/certified-operators-dn2np" Jan 21 11:26:48 crc kubenswrapper[4824]: I0121 11:26:48.387609 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f9j78\" (UniqueName: \"kubernetes.io/projected/1d8c8cff-4340-4eb4-94b7-062968a14848-kube-api-access-f9j78\") pod \"certified-operators-dn2np\" (UID: \"1d8c8cff-4340-4eb4-94b7-062968a14848\") " pod="openshift-marketplace/certified-operators-dn2np" Jan 21 11:26:48 crc kubenswrapper[4824]: I0121 11:26:48.388218 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d8c8cff-4340-4eb4-94b7-062968a14848-catalog-content\") pod \"certified-operators-dn2np\" (UID: \"1d8c8cff-4340-4eb4-94b7-062968a14848\") " pod="openshift-marketplace/certified-operators-dn2np" Jan 21 11:26:48 crc kubenswrapper[4824]: I0121 11:26:48.388497 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d8c8cff-4340-4eb4-94b7-062968a14848-utilities\") pod \"certified-operators-dn2np\" (UID: \"1d8c8cff-4340-4eb4-94b7-062968a14848\") " pod="openshift-marketplace/certified-operators-dn2np" Jan 21 11:26:48 crc kubenswrapper[4824]: I0121 11:26:48.404589 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f9j78\" (UniqueName: \"kubernetes.io/projected/1d8c8cff-4340-4eb4-94b7-062968a14848-kube-api-access-f9j78\") pod \"certified-operators-dn2np\" (UID: \"1d8c8cff-4340-4eb4-94b7-062968a14848\") " pod="openshift-marketplace/certified-operators-dn2np" Jan 21 11:26:48 crc kubenswrapper[4824]: I0121 11:26:48.502022 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dn2np" Jan 21 11:26:48 crc kubenswrapper[4824]: I0121 11:26:48.851239 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bbefda93-dc56-43c5-b131-56af96ca832b","Type":"ContainerStarted","Data":"be5899a6e3505ef77d83bc3a7dc27f5c3518d092ccfa11eacf115b078cae20ee"} Jan 21 11:26:48 crc kubenswrapper[4824]: I0121 11:26:48.852190 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-6c5fc769c9-spmzr" event={"ID":"7d943b5e-d05f-445a-888d-7ebe340468b9","Type":"ContainerStarted","Data":"97fb464b408f5ceff487705be319b8cd670592be37d52a0443879365b61921a9"} Jan 21 11:26:48 crc kubenswrapper[4824]: I0121 11:26:48.852218 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-6c5fc769c9-spmzr" event={"ID":"7d943b5e-d05f-445a-888d-7ebe340468b9","Type":"ContainerStarted","Data":"779d43c0e7120424033bc4414b128725901d8d0023671ffa45ed85a0227c9c69"} Jan 21 11:26:48 crc kubenswrapper[4824]: I0121 11:26:48.852980 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-6c5fc769c9-spmzr" Jan 21 11:26:48 crc kubenswrapper[4824]: I0121 11:26:48.854367 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-jrsn8" Jan 21 11:26:48 crc kubenswrapper[4824]: I0121 11:26:48.854362 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-8654bbf998-p8crm" event={"ID":"c23477ba-5a41-44da-a8b7-6606595b9726","Type":"ContainerStarted","Data":"a33d5c3f52560f39a45b643e1e7362cec3d1ffa8698489a0cd9d0e49beebc4f3"} Jan 21 11:26:48 crc kubenswrapper[4824]: I0121 11:26:48.854411 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-8654bbf998-p8crm" event={"ID":"c23477ba-5a41-44da-a8b7-6606595b9726","Type":"ContainerStarted","Data":"db3de21ed146f19802de452a644645642de1b010142ab29da18a4e42e355a6fe"} Jan 21 11:26:48 crc kubenswrapper[4824]: I0121 11:26:48.854555 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-8654bbf998-p8crm" Jan 21 11:26:48 crc kubenswrapper[4824]: I0121 11:26:48.854579 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-8654bbf998-p8crm" event={"ID":"c23477ba-5a41-44da-a8b7-6606595b9726","Type":"ContainerStarted","Data":"d922ef122df4ca799eb3f9e4b5e4b6c48e353cdfcff00495934a761838cd1cd6"} Jan 21 11:26:48 crc kubenswrapper[4824]: I0121 11:26:48.854592 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-8654bbf998-p8crm" Jan 21 11:26:48 crc kubenswrapper[4824]: I0121 11:26:48.868382 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-6c5fc769c9-spmzr" podStartSLOduration=2.868369402 podStartE2EDuration="2.868369402s" podCreationTimestamp="2026-01-21 11:26:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:26:48.864564345 +0000 UTC m=+951.157593637" watchObservedRunningTime="2026-01-21 11:26:48.868369402 +0000 UTC m=+951.161398695" Jan 21 11:26:48 crc kubenswrapper[4824]: I0121 11:26:48.883380 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-8654bbf998-p8crm" podStartSLOduration=4.883369656 podStartE2EDuration="4.883369656s" podCreationTimestamp="2026-01-21 11:26:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:26:48.879331099 +0000 UTC m=+951.172360391" watchObservedRunningTime="2026-01-21 11:26:48.883369656 +0000 UTC m=+951.176398949" Jan 21 11:26:48 crc kubenswrapper[4824]: I0121 11:26:48.966346 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-dn2np"] Jan 21 11:26:48 crc kubenswrapper[4824]: W0121 11:26:48.969857 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1d8c8cff_4340_4eb4_94b7_062968a14848.slice/crio-2dec804f6fee46d1464f08de4ca7da8580e7f171b751ac42a7b92ec863162d9b WatchSource:0}: Error finding container 2dec804f6fee46d1464f08de4ca7da8580e7f171b751ac42a7b92ec863162d9b: Status 404 returned error can't find the container with id 2dec804f6fee46d1464f08de4ca7da8580e7f171b751ac42a7b92ec863162d9b Jan 21 11:26:49 crc kubenswrapper[4824]: I0121 11:26:49.875471 4824 generic.go:334] "Generic (PLEG): container finished" podID="1d8c8cff-4340-4eb4-94b7-062968a14848" containerID="d64282b8c833bf7a0f6a7b257f155ac96972db2ea6ccea22f61d3f3c228fc02a" exitCode=0 Jan 21 11:26:49 crc kubenswrapper[4824]: I0121 11:26:49.876695 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dn2np" event={"ID":"1d8c8cff-4340-4eb4-94b7-062968a14848","Type":"ContainerDied","Data":"d64282b8c833bf7a0f6a7b257f155ac96972db2ea6ccea22f61d3f3c228fc02a"} Jan 21 11:26:49 crc kubenswrapper[4824]: I0121 11:26:49.876734 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dn2np" event={"ID":"1d8c8cff-4340-4eb4-94b7-062968a14848","Type":"ContainerStarted","Data":"2dec804f6fee46d1464f08de4ca7da8580e7f171b751ac42a7b92ec863162d9b"} Jan 21 11:26:51 crc kubenswrapper[4824]: I0121 11:26:51.051790 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Jan 21 11:26:51 crc kubenswrapper[4824]: I0121 11:26:51.052040 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Jan 21 11:26:51 crc kubenswrapper[4824]: I0121 11:26:51.063832 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Jan 21 11:26:51 crc kubenswrapper[4824]: I0121 11:26:51.063870 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Jan 21 11:26:51 crc kubenswrapper[4824]: I0121 11:26:51.073280 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Jan 21 11:26:51 crc kubenswrapper[4824]: I0121 11:26:51.079262 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Jan 21 11:26:51 crc kubenswrapper[4824]: I0121 11:26:51.092407 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Jan 21 11:26:51 crc kubenswrapper[4824]: I0121 11:26:51.094792 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Jan 21 11:26:51 crc kubenswrapper[4824]: I0121 11:26:51.513101 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-55f844cf75-vbmkp" Jan 21 11:26:51 crc kubenswrapper[4824]: I0121 11:26:51.565592 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-fh8q6"] Jan 21 11:26:51 crc kubenswrapper[4824]: I0121 11:26:51.565823 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-785d8bcb8c-fh8q6" podUID="8214df67-a984-4a5b-bcbb-6b16ab42f5fe" containerName="dnsmasq-dns" containerID="cri-o://56442b913a4db71125dc75408dc50b059e8519640a2cc312a0c8277e0b1aefab" gracePeriod=10 Jan 21 11:26:51 crc kubenswrapper[4824]: I0121 11:26:51.893009 4824 generic.go:334] "Generic (PLEG): container finished" podID="8214df67-a984-4a5b-bcbb-6b16ab42f5fe" containerID="56442b913a4db71125dc75408dc50b059e8519640a2cc312a0c8277e0b1aefab" exitCode=0 Jan 21 11:26:51 crc kubenswrapper[4824]: I0121 11:26:51.893255 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-fh8q6" event={"ID":"8214df67-a984-4a5b-bcbb-6b16ab42f5fe","Type":"ContainerDied","Data":"56442b913a4db71125dc75408dc50b059e8519640a2cc312a0c8277e0b1aefab"} Jan 21 11:26:51 crc kubenswrapper[4824]: I0121 11:26:51.893813 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Jan 21 11:26:51 crc kubenswrapper[4824]: I0121 11:26:51.893838 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Jan 21 11:26:51 crc kubenswrapper[4824]: I0121 11:26:51.893848 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Jan 21 11:26:51 crc kubenswrapper[4824]: I0121 11:26:51.893856 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Jan 21 11:26:52 crc kubenswrapper[4824]: I0121 11:26:52.012344 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-785d8bcb8c-fh8q6" podUID="8214df67-a984-4a5b-bcbb-6b16ab42f5fe" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.140:5353: connect: connection refused" Jan 21 11:26:53 crc kubenswrapper[4824]: I0121 11:26:53.430854 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-785d8bcb8c-fh8q6" Jan 21 11:26:53 crc kubenswrapper[4824]: I0121 11:26:53.538158 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Jan 21 11:26:53 crc kubenswrapper[4824]: I0121 11:26:53.596323 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8214df67-a984-4a5b-bcbb-6b16ab42f5fe-ovsdbserver-sb\") pod \"8214df67-a984-4a5b-bcbb-6b16ab42f5fe\" (UID: \"8214df67-a984-4a5b-bcbb-6b16ab42f5fe\") " Jan 21 11:26:53 crc kubenswrapper[4824]: I0121 11:26:53.596457 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8214df67-a984-4a5b-bcbb-6b16ab42f5fe-config\") pod \"8214df67-a984-4a5b-bcbb-6b16ab42f5fe\" (UID: \"8214df67-a984-4a5b-bcbb-6b16ab42f5fe\") " Jan 21 11:26:53 crc kubenswrapper[4824]: I0121 11:26:53.596503 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x8x9b\" (UniqueName: \"kubernetes.io/projected/8214df67-a984-4a5b-bcbb-6b16ab42f5fe-kube-api-access-x8x9b\") pod \"8214df67-a984-4a5b-bcbb-6b16ab42f5fe\" (UID: \"8214df67-a984-4a5b-bcbb-6b16ab42f5fe\") " Jan 21 11:26:53 crc kubenswrapper[4824]: I0121 11:26:53.596529 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8214df67-a984-4a5b-bcbb-6b16ab42f5fe-ovsdbserver-nb\") pod \"8214df67-a984-4a5b-bcbb-6b16ab42f5fe\" (UID: \"8214df67-a984-4a5b-bcbb-6b16ab42f5fe\") " Jan 21 11:26:53 crc kubenswrapper[4824]: I0121 11:26:53.596587 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8214df67-a984-4a5b-bcbb-6b16ab42f5fe-dns-swift-storage-0\") pod \"8214df67-a984-4a5b-bcbb-6b16ab42f5fe\" (UID: \"8214df67-a984-4a5b-bcbb-6b16ab42f5fe\") " Jan 21 11:26:53 crc kubenswrapper[4824]: I0121 11:26:53.596619 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8214df67-a984-4a5b-bcbb-6b16ab42f5fe-dns-svc\") pod \"8214df67-a984-4a5b-bcbb-6b16ab42f5fe\" (UID: \"8214df67-a984-4a5b-bcbb-6b16ab42f5fe\") " Jan 21 11:26:53 crc kubenswrapper[4824]: I0121 11:26:53.606581 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Jan 21 11:26:53 crc kubenswrapper[4824]: I0121 11:26:53.614630 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8214df67-a984-4a5b-bcbb-6b16ab42f5fe-kube-api-access-x8x9b" (OuterVolumeSpecName: "kube-api-access-x8x9b") pod "8214df67-a984-4a5b-bcbb-6b16ab42f5fe" (UID: "8214df67-a984-4a5b-bcbb-6b16ab42f5fe"). InnerVolumeSpecName "kube-api-access-x8x9b". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:26:53 crc kubenswrapper[4824]: I0121 11:26:53.634247 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8214df67-a984-4a5b-bcbb-6b16ab42f5fe-config" (OuterVolumeSpecName: "config") pod "8214df67-a984-4a5b-bcbb-6b16ab42f5fe" (UID: "8214df67-a984-4a5b-bcbb-6b16ab42f5fe"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:26:53 crc kubenswrapper[4824]: I0121 11:26:53.634284 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8214df67-a984-4a5b-bcbb-6b16ab42f5fe-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "8214df67-a984-4a5b-bcbb-6b16ab42f5fe" (UID: "8214df67-a984-4a5b-bcbb-6b16ab42f5fe"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:26:53 crc kubenswrapper[4824]: I0121 11:26:53.638617 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8214df67-a984-4a5b-bcbb-6b16ab42f5fe-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "8214df67-a984-4a5b-bcbb-6b16ab42f5fe" (UID: "8214df67-a984-4a5b-bcbb-6b16ab42f5fe"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:26:53 crc kubenswrapper[4824]: I0121 11:26:53.642602 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8214df67-a984-4a5b-bcbb-6b16ab42f5fe-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "8214df67-a984-4a5b-bcbb-6b16ab42f5fe" (UID: "8214df67-a984-4a5b-bcbb-6b16ab42f5fe"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:26:53 crc kubenswrapper[4824]: I0121 11:26:53.650005 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8214df67-a984-4a5b-bcbb-6b16ab42f5fe-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "8214df67-a984-4a5b-bcbb-6b16ab42f5fe" (UID: "8214df67-a984-4a5b-bcbb-6b16ab42f5fe"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:26:53 crc kubenswrapper[4824]: I0121 11:26:53.666256 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Jan 21 11:26:53 crc kubenswrapper[4824]: I0121 11:26:53.698469 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8214df67-a984-4a5b-bcbb-6b16ab42f5fe-config\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:53 crc kubenswrapper[4824]: I0121 11:26:53.698496 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x8x9b\" (UniqueName: \"kubernetes.io/projected/8214df67-a984-4a5b-bcbb-6b16ab42f5fe-kube-api-access-x8x9b\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:53 crc kubenswrapper[4824]: I0121 11:26:53.698508 4824 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8214df67-a984-4a5b-bcbb-6b16ab42f5fe-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:53 crc kubenswrapper[4824]: I0121 11:26:53.698516 4824 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8214df67-a984-4a5b-bcbb-6b16ab42f5fe-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:53 crc kubenswrapper[4824]: I0121 11:26:53.698525 4824 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8214df67-a984-4a5b-bcbb-6b16ab42f5fe-dns-svc\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:53 crc kubenswrapper[4824]: I0121 11:26:53.698532 4824 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8214df67-a984-4a5b-bcbb-6b16ab42f5fe-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:53 crc kubenswrapper[4824]: I0121 11:26:53.799551 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Jan 21 11:26:53 crc kubenswrapper[4824]: I0121 11:26:53.909266 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-785d8bcb8c-fh8q6" Jan 21 11:26:53 crc kubenswrapper[4824]: I0121 11:26:53.909591 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-fh8q6" event={"ID":"8214df67-a984-4a5b-bcbb-6b16ab42f5fe","Type":"ContainerDied","Data":"a7ed7bee402dc3db5565fe550b060e37c1771f55a94528a7a39dbd2741c1ec27"} Jan 21 11:26:53 crc kubenswrapper[4824]: I0121 11:26:53.909632 4824 scope.go:117] "RemoveContainer" containerID="56442b913a4db71125dc75408dc50b059e8519640a2cc312a0c8277e0b1aefab" Jan 21 11:26:53 crc kubenswrapper[4824]: I0121 11:26:53.954348 4824 scope.go:117] "RemoveContainer" containerID="7af315bc544e0ed568b3db042720a7c28e2f59b1fb260d2208c816705728d7f5" Jan 21 11:26:53 crc kubenswrapper[4824]: I0121 11:26:53.978680 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-fh8q6"] Jan 21 11:26:53 crc kubenswrapper[4824]: I0121 11:26:53.984401 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-fh8q6"] Jan 21 11:26:54 crc kubenswrapper[4824]: I0121 11:26:54.062074 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8214df67-a984-4a5b-bcbb-6b16ab42f5fe" path="/var/lib/kubelet/pods/8214df67-a984-4a5b-bcbb-6b16ab42f5fe/volumes" Jan 21 11:26:54 crc kubenswrapper[4824]: I0121 11:26:54.926375 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bbefda93-dc56-43c5-b131-56af96ca832b","Type":"ContainerStarted","Data":"c07784e789097e065abafdd72f5db898c8ca368962bfcfa018e7cbce776cea3b"} Jan 21 11:26:54 crc kubenswrapper[4824]: I0121 11:26:54.926508 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="bbefda93-dc56-43c5-b131-56af96ca832b" containerName="ceilometer-central-agent" containerID="cri-o://c4e76cd725e81ba223348c70595837e4caaba881aaa9eba8437cfc5280bd859c" gracePeriod=30 Jan 21 11:26:54 crc kubenswrapper[4824]: I0121 11:26:54.926621 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="bbefda93-dc56-43c5-b131-56af96ca832b" containerName="ceilometer-notification-agent" containerID="cri-o://af014f3d948b12e938ae99b89aabd762100d2a622ac15f0606ad12c79347a128" gracePeriod=30 Jan 21 11:26:54 crc kubenswrapper[4824]: I0121 11:26:54.926666 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="bbefda93-dc56-43c5-b131-56af96ca832b" containerName="proxy-httpd" containerID="cri-o://c07784e789097e065abafdd72f5db898c8ca368962bfcfa018e7cbce776cea3b" gracePeriod=30 Jan 21 11:26:54 crc kubenswrapper[4824]: I0121 11:26:54.926689 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Jan 21 11:26:54 crc kubenswrapper[4824]: I0121 11:26:54.926769 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="bbefda93-dc56-43c5-b131-56af96ca832b" containerName="sg-core" containerID="cri-o://be5899a6e3505ef77d83bc3a7dc27f5c3518d092ccfa11eacf115b078cae20ee" gracePeriod=30 Jan 21 11:26:54 crc kubenswrapper[4824]: I0121 11:26:54.930845 4824 generic.go:334] "Generic (PLEG): container finished" podID="1d8c8cff-4340-4eb4-94b7-062968a14848" containerID="d4406aff85ae4b7edaab2e90513b6f7cdb4c88bf1bcc43b39ae793abac0eb943" exitCode=0 Jan 21 11:26:54 crc kubenswrapper[4824]: I0121 11:26:54.931949 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dn2np" event={"ID":"1d8c8cff-4340-4eb4-94b7-062968a14848","Type":"ContainerDied","Data":"d4406aff85ae4b7edaab2e90513b6f7cdb4c88bf1bcc43b39ae793abac0eb943"} Jan 21 11:26:54 crc kubenswrapper[4824]: I0121 11:26:54.948226 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.473406197 podStartE2EDuration="33.948216281s" podCreationTimestamp="2026-01-21 11:26:21 +0000 UTC" firstStartedPulling="2026-01-21 11:26:22.402277514 +0000 UTC m=+924.695306806" lastFinishedPulling="2026-01-21 11:26:53.877087598 +0000 UTC m=+956.170116890" observedRunningTime="2026-01-21 11:26:54.945212356 +0000 UTC m=+957.238241648" watchObservedRunningTime="2026-01-21 11:26:54.948216281 +0000 UTC m=+957.241245573" Jan 21 11:26:55 crc kubenswrapper[4824]: I0121 11:26:55.938596 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-rlp2t" event={"ID":"0a6c0628-4522-4bb7-8a82-cc2e019eca2a","Type":"ContainerStarted","Data":"b5f108624f476c15b99bcdcb7c95850ef2819568712069d3c2536bdcb6daa49b"} Jan 21 11:26:55 crc kubenswrapper[4824]: I0121 11:26:55.952367 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-rlp2t" podStartSLOduration=2.6327185650000002 podStartE2EDuration="33.952354966s" podCreationTimestamp="2026-01-21 11:26:22 +0000 UTC" firstStartedPulling="2026-01-21 11:26:23.218138454 +0000 UTC m=+925.511167746" lastFinishedPulling="2026-01-21 11:26:54.537774854 +0000 UTC m=+956.830804147" observedRunningTime="2026-01-21 11:26:55.951670775 +0000 UTC m=+958.244700067" watchObservedRunningTime="2026-01-21 11:26:55.952354966 +0000 UTC m=+958.245384257" Jan 21 11:26:55 crc kubenswrapper[4824]: I0121 11:26:55.961988 4824 generic.go:334] "Generic (PLEG): container finished" podID="bbefda93-dc56-43c5-b131-56af96ca832b" containerID="c07784e789097e065abafdd72f5db898c8ca368962bfcfa018e7cbce776cea3b" exitCode=0 Jan 21 11:26:55 crc kubenswrapper[4824]: I0121 11:26:55.962019 4824 generic.go:334] "Generic (PLEG): container finished" podID="bbefda93-dc56-43c5-b131-56af96ca832b" containerID="be5899a6e3505ef77d83bc3a7dc27f5c3518d092ccfa11eacf115b078cae20ee" exitCode=2 Jan 21 11:26:55 crc kubenswrapper[4824]: I0121 11:26:55.962028 4824 generic.go:334] "Generic (PLEG): container finished" podID="bbefda93-dc56-43c5-b131-56af96ca832b" containerID="c4e76cd725e81ba223348c70595837e4caaba881aaa9eba8437cfc5280bd859c" exitCode=0 Jan 21 11:26:55 crc kubenswrapper[4824]: I0121 11:26:55.962055 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bbefda93-dc56-43c5-b131-56af96ca832b","Type":"ContainerDied","Data":"c07784e789097e065abafdd72f5db898c8ca368962bfcfa018e7cbce776cea3b"} Jan 21 11:26:55 crc kubenswrapper[4824]: I0121 11:26:55.962104 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bbefda93-dc56-43c5-b131-56af96ca832b","Type":"ContainerDied","Data":"be5899a6e3505ef77d83bc3a7dc27f5c3518d092ccfa11eacf115b078cae20ee"} Jan 21 11:26:55 crc kubenswrapper[4824]: I0121 11:26:55.962115 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bbefda93-dc56-43c5-b131-56af96ca832b","Type":"ContainerDied","Data":"c4e76cd725e81ba223348c70595837e4caaba881aaa9eba8437cfc5280bd859c"} Jan 21 11:26:55 crc kubenswrapper[4824]: I0121 11:26:55.964446 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dn2np" event={"ID":"1d8c8cff-4340-4eb4-94b7-062968a14848","Type":"ContainerStarted","Data":"42cfb3a03e211fbd5fa55f6c859e8ec840fddea6ec887daa75d38ab58da624f4"} Jan 21 11:26:55 crc kubenswrapper[4824]: I0121 11:26:55.982306 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-dn2np" podStartSLOduration=2.405619777 podStartE2EDuration="7.982293647s" podCreationTimestamp="2026-01-21 11:26:48 +0000 UTC" firstStartedPulling="2026-01-21 11:26:49.877638805 +0000 UTC m=+952.170668097" lastFinishedPulling="2026-01-21 11:26:55.454312674 +0000 UTC m=+957.747341967" observedRunningTime="2026-01-21 11:26:55.980171524 +0000 UTC m=+958.273200817" watchObservedRunningTime="2026-01-21 11:26:55.982293647 +0000 UTC m=+958.275322940" Jan 21 11:26:56 crc kubenswrapper[4824]: I0121 11:26:56.972762 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-ngq69" event={"ID":"262733b9-4179-49e3-aee9-b62197cc89ba","Type":"ContainerStarted","Data":"4c9c6f12e02388718e6b9fff3020bdecb9cf28f3ef2f988918e2aeb78dac7d68"} Jan 21 11:26:56 crc kubenswrapper[4824]: I0121 11:26:56.986255 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-ngq69" podStartSLOduration=1.8551470490000002 podStartE2EDuration="34.986243206s" podCreationTimestamp="2026-01-21 11:26:22 +0000 UTC" firstStartedPulling="2026-01-21 11:26:23.395650882 +0000 UTC m=+925.688680175" lastFinishedPulling="2026-01-21 11:26:56.52674704 +0000 UTC m=+958.819776332" observedRunningTime="2026-01-21 11:26:56.983586665 +0000 UTC m=+959.276615946" watchObservedRunningTime="2026-01-21 11:26:56.986243206 +0000 UTC m=+959.279272498" Jan 21 11:26:57 crc kubenswrapper[4824]: I0121 11:26:57.979659 4824 generic.go:334] "Generic (PLEG): container finished" podID="262733b9-4179-49e3-aee9-b62197cc89ba" containerID="4c9c6f12e02388718e6b9fff3020bdecb9cf28f3ef2f988918e2aeb78dac7d68" exitCode=0 Jan 21 11:26:57 crc kubenswrapper[4824]: I0121 11:26:57.979753 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-ngq69" event={"ID":"262733b9-4179-49e3-aee9-b62197cc89ba","Type":"ContainerDied","Data":"4c9c6f12e02388718e6b9fff3020bdecb9cf28f3ef2f988918e2aeb78dac7d68"} Jan 21 11:26:57 crc kubenswrapper[4824]: I0121 11:26:57.980973 4824 generic.go:334] "Generic (PLEG): container finished" podID="0a6c0628-4522-4bb7-8a82-cc2e019eca2a" containerID="b5f108624f476c15b99bcdcb7c95850ef2819568712069d3c2536bdcb6daa49b" exitCode=0 Jan 21 11:26:57 crc kubenswrapper[4824]: I0121 11:26:57.981012 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-rlp2t" event={"ID":"0a6c0628-4522-4bb7-8a82-cc2e019eca2a","Type":"ContainerDied","Data":"b5f108624f476c15b99bcdcb7c95850ef2819568712069d3c2536bdcb6daa49b"} Jan 21 11:26:58 crc kubenswrapper[4824]: I0121 11:26:58.308506 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 21 11:26:58 crc kubenswrapper[4824]: I0121 11:26:58.489621 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bbefda93-dc56-43c5-b131-56af96ca832b-sg-core-conf-yaml\") pod \"bbefda93-dc56-43c5-b131-56af96ca832b\" (UID: \"bbefda93-dc56-43c5-b131-56af96ca832b\") " Jan 21 11:26:58 crc kubenswrapper[4824]: I0121 11:26:58.490008 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bbefda93-dc56-43c5-b131-56af96ca832b-log-httpd\") pod \"bbefda93-dc56-43c5-b131-56af96ca832b\" (UID: \"bbefda93-dc56-43c5-b131-56af96ca832b\") " Jan 21 11:26:58 crc kubenswrapper[4824]: I0121 11:26:58.490040 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sgmtk\" (UniqueName: \"kubernetes.io/projected/bbefda93-dc56-43c5-b131-56af96ca832b-kube-api-access-sgmtk\") pod \"bbefda93-dc56-43c5-b131-56af96ca832b\" (UID: \"bbefda93-dc56-43c5-b131-56af96ca832b\") " Jan 21 11:26:58 crc kubenswrapper[4824]: I0121 11:26:58.490099 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bbefda93-dc56-43c5-b131-56af96ca832b-combined-ca-bundle\") pod \"bbefda93-dc56-43c5-b131-56af96ca832b\" (UID: \"bbefda93-dc56-43c5-b131-56af96ca832b\") " Jan 21 11:26:58 crc kubenswrapper[4824]: I0121 11:26:58.490171 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bbefda93-dc56-43c5-b131-56af96ca832b-config-data\") pod \"bbefda93-dc56-43c5-b131-56af96ca832b\" (UID: \"bbefda93-dc56-43c5-b131-56af96ca832b\") " Jan 21 11:26:58 crc kubenswrapper[4824]: I0121 11:26:58.490262 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bbefda93-dc56-43c5-b131-56af96ca832b-scripts\") pod \"bbefda93-dc56-43c5-b131-56af96ca832b\" (UID: \"bbefda93-dc56-43c5-b131-56af96ca832b\") " Jan 21 11:26:58 crc kubenswrapper[4824]: I0121 11:26:58.490288 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bbefda93-dc56-43c5-b131-56af96ca832b-run-httpd\") pod \"bbefda93-dc56-43c5-b131-56af96ca832b\" (UID: \"bbefda93-dc56-43c5-b131-56af96ca832b\") " Jan 21 11:26:58 crc kubenswrapper[4824]: I0121 11:26:58.490493 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bbefda93-dc56-43c5-b131-56af96ca832b-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "bbefda93-dc56-43c5-b131-56af96ca832b" (UID: "bbefda93-dc56-43c5-b131-56af96ca832b"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:26:58 crc kubenswrapper[4824]: I0121 11:26:58.490902 4824 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bbefda93-dc56-43c5-b131-56af96ca832b-log-httpd\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:58 crc kubenswrapper[4824]: I0121 11:26:58.490897 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bbefda93-dc56-43c5-b131-56af96ca832b-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "bbefda93-dc56-43c5-b131-56af96ca832b" (UID: "bbefda93-dc56-43c5-b131-56af96ca832b"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:26:58 crc kubenswrapper[4824]: I0121 11:26:58.494304 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bbefda93-dc56-43c5-b131-56af96ca832b-scripts" (OuterVolumeSpecName: "scripts") pod "bbefda93-dc56-43c5-b131-56af96ca832b" (UID: "bbefda93-dc56-43c5-b131-56af96ca832b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:26:58 crc kubenswrapper[4824]: I0121 11:26:58.494395 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bbefda93-dc56-43c5-b131-56af96ca832b-kube-api-access-sgmtk" (OuterVolumeSpecName: "kube-api-access-sgmtk") pod "bbefda93-dc56-43c5-b131-56af96ca832b" (UID: "bbefda93-dc56-43c5-b131-56af96ca832b"). InnerVolumeSpecName "kube-api-access-sgmtk". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:26:58 crc kubenswrapper[4824]: I0121 11:26:58.502505 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-dn2np" Jan 21 11:26:58 crc kubenswrapper[4824]: I0121 11:26:58.502560 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-dn2np" Jan 21 11:26:58 crc kubenswrapper[4824]: I0121 11:26:58.509616 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bbefda93-dc56-43c5-b131-56af96ca832b-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "bbefda93-dc56-43c5-b131-56af96ca832b" (UID: "bbefda93-dc56-43c5-b131-56af96ca832b"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:26:58 crc kubenswrapper[4824]: I0121 11:26:58.533679 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-dn2np" Jan 21 11:26:58 crc kubenswrapper[4824]: I0121 11:26:58.536144 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bbefda93-dc56-43c5-b131-56af96ca832b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bbefda93-dc56-43c5-b131-56af96ca832b" (UID: "bbefda93-dc56-43c5-b131-56af96ca832b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:26:58 crc kubenswrapper[4824]: I0121 11:26:58.552572 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bbefda93-dc56-43c5-b131-56af96ca832b-config-data" (OuterVolumeSpecName: "config-data") pod "bbefda93-dc56-43c5-b131-56af96ca832b" (UID: "bbefda93-dc56-43c5-b131-56af96ca832b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:26:58 crc kubenswrapper[4824]: I0121 11:26:58.592867 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sgmtk\" (UniqueName: \"kubernetes.io/projected/bbefda93-dc56-43c5-b131-56af96ca832b-kube-api-access-sgmtk\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:58 crc kubenswrapper[4824]: I0121 11:26:58.592902 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bbefda93-dc56-43c5-b131-56af96ca832b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:58 crc kubenswrapper[4824]: I0121 11:26:58.592912 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bbefda93-dc56-43c5-b131-56af96ca832b-config-data\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:58 crc kubenswrapper[4824]: I0121 11:26:58.592920 4824 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bbefda93-dc56-43c5-b131-56af96ca832b-scripts\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:58 crc kubenswrapper[4824]: I0121 11:26:58.592928 4824 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bbefda93-dc56-43c5-b131-56af96ca832b-run-httpd\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:58 crc kubenswrapper[4824]: I0121 11:26:58.592935 4824 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bbefda93-dc56-43c5-b131-56af96ca832b-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:58 crc kubenswrapper[4824]: I0121 11:26:58.989804 4824 generic.go:334] "Generic (PLEG): container finished" podID="bbefda93-dc56-43c5-b131-56af96ca832b" containerID="af014f3d948b12e938ae99b89aabd762100d2a622ac15f0606ad12c79347a128" exitCode=0 Jan 21 11:26:58 crc kubenswrapper[4824]: I0121 11:26:58.989873 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 21 11:26:58 crc kubenswrapper[4824]: I0121 11:26:58.989917 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bbefda93-dc56-43c5-b131-56af96ca832b","Type":"ContainerDied","Data":"af014f3d948b12e938ae99b89aabd762100d2a622ac15f0606ad12c79347a128"} Jan 21 11:26:58 crc kubenswrapper[4824]: I0121 11:26:58.989976 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bbefda93-dc56-43c5-b131-56af96ca832b","Type":"ContainerDied","Data":"9fa11b86f012b31aa61fb0528e706362f80104db47ab25f797420eb04f92a3fa"} Jan 21 11:26:58 crc kubenswrapper[4824]: I0121 11:26:58.990001 4824 scope.go:117] "RemoveContainer" containerID="c07784e789097e065abafdd72f5db898c8ca368962bfcfa018e7cbce776cea3b" Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.008047 4824 scope.go:117] "RemoveContainer" containerID="be5899a6e3505ef77d83bc3a7dc27f5c3518d092ccfa11eacf115b078cae20ee" Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.021348 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.024647 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.034266 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Jan 21 11:26:59 crc kubenswrapper[4824]: E0121 11:26:59.034578 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bbefda93-dc56-43c5-b131-56af96ca832b" containerName="sg-core" Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.034597 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="bbefda93-dc56-43c5-b131-56af96ca832b" containerName="sg-core" Jan 21 11:26:59 crc kubenswrapper[4824]: E0121 11:26:59.034618 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8214df67-a984-4a5b-bcbb-6b16ab42f5fe" containerName="dnsmasq-dns" Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.034624 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="8214df67-a984-4a5b-bcbb-6b16ab42f5fe" containerName="dnsmasq-dns" Jan 21 11:26:59 crc kubenswrapper[4824]: E0121 11:26:59.034637 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bbefda93-dc56-43c5-b131-56af96ca832b" containerName="ceilometer-central-agent" Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.034644 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="bbefda93-dc56-43c5-b131-56af96ca832b" containerName="ceilometer-central-agent" Jan 21 11:26:59 crc kubenswrapper[4824]: E0121 11:26:59.034656 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8214df67-a984-4a5b-bcbb-6b16ab42f5fe" containerName="init" Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.034661 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="8214df67-a984-4a5b-bcbb-6b16ab42f5fe" containerName="init" Jan 21 11:26:59 crc kubenswrapper[4824]: E0121 11:26:59.034674 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bbefda93-dc56-43c5-b131-56af96ca832b" containerName="ceilometer-notification-agent" Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.034679 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="bbefda93-dc56-43c5-b131-56af96ca832b" containerName="ceilometer-notification-agent" Jan 21 11:26:59 crc kubenswrapper[4824]: E0121 11:26:59.034690 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bbefda93-dc56-43c5-b131-56af96ca832b" containerName="proxy-httpd" Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.034695 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="bbefda93-dc56-43c5-b131-56af96ca832b" containerName="proxy-httpd" Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.034850 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="bbefda93-dc56-43c5-b131-56af96ca832b" containerName="ceilometer-central-agent" Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.034874 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="bbefda93-dc56-43c5-b131-56af96ca832b" containerName="ceilometer-notification-agent" Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.034893 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="bbefda93-dc56-43c5-b131-56af96ca832b" containerName="sg-core" Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.034905 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="bbefda93-dc56-43c5-b131-56af96ca832b" containerName="proxy-httpd" Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.034919 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="8214df67-a984-4a5b-bcbb-6b16ab42f5fe" containerName="dnsmasq-dns" Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.036325 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.038799 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.038975 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.046343 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.047575 4824 scope.go:117] "RemoveContainer" containerID="af014f3d948b12e938ae99b89aabd762100d2a622ac15f0606ad12c79347a128" Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.088695 4824 scope.go:117] "RemoveContainer" containerID="c4e76cd725e81ba223348c70595837e4caaba881aaa9eba8437cfc5280bd859c" Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.103581 4824 scope.go:117] "RemoveContainer" containerID="c07784e789097e065abafdd72f5db898c8ca368962bfcfa018e7cbce776cea3b" Jan 21 11:26:59 crc kubenswrapper[4824]: E0121 11:26:59.104245 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c07784e789097e065abafdd72f5db898c8ca368962bfcfa018e7cbce776cea3b\": container with ID starting with c07784e789097e065abafdd72f5db898c8ca368962bfcfa018e7cbce776cea3b not found: ID does not exist" containerID="c07784e789097e065abafdd72f5db898c8ca368962bfcfa018e7cbce776cea3b" Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.104283 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c07784e789097e065abafdd72f5db898c8ca368962bfcfa018e7cbce776cea3b"} err="failed to get container status \"c07784e789097e065abafdd72f5db898c8ca368962bfcfa018e7cbce776cea3b\": rpc error: code = NotFound desc = could not find container \"c07784e789097e065abafdd72f5db898c8ca368962bfcfa018e7cbce776cea3b\": container with ID starting with c07784e789097e065abafdd72f5db898c8ca368962bfcfa018e7cbce776cea3b not found: ID does not exist" Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.104328 4824 scope.go:117] "RemoveContainer" containerID="be5899a6e3505ef77d83bc3a7dc27f5c3518d092ccfa11eacf115b078cae20ee" Jan 21 11:26:59 crc kubenswrapper[4824]: E0121 11:26:59.104638 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"be5899a6e3505ef77d83bc3a7dc27f5c3518d092ccfa11eacf115b078cae20ee\": container with ID starting with be5899a6e3505ef77d83bc3a7dc27f5c3518d092ccfa11eacf115b078cae20ee not found: ID does not exist" containerID="be5899a6e3505ef77d83bc3a7dc27f5c3518d092ccfa11eacf115b078cae20ee" Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.104667 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"be5899a6e3505ef77d83bc3a7dc27f5c3518d092ccfa11eacf115b078cae20ee"} err="failed to get container status \"be5899a6e3505ef77d83bc3a7dc27f5c3518d092ccfa11eacf115b078cae20ee\": rpc error: code = NotFound desc = could not find container \"be5899a6e3505ef77d83bc3a7dc27f5c3518d092ccfa11eacf115b078cae20ee\": container with ID starting with be5899a6e3505ef77d83bc3a7dc27f5c3518d092ccfa11eacf115b078cae20ee not found: ID does not exist" Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.104686 4824 scope.go:117] "RemoveContainer" containerID="af014f3d948b12e938ae99b89aabd762100d2a622ac15f0606ad12c79347a128" Jan 21 11:26:59 crc kubenswrapper[4824]: E0121 11:26:59.104980 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"af014f3d948b12e938ae99b89aabd762100d2a622ac15f0606ad12c79347a128\": container with ID starting with af014f3d948b12e938ae99b89aabd762100d2a622ac15f0606ad12c79347a128 not found: ID does not exist" containerID="af014f3d948b12e938ae99b89aabd762100d2a622ac15f0606ad12c79347a128" Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.105014 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"af014f3d948b12e938ae99b89aabd762100d2a622ac15f0606ad12c79347a128"} err="failed to get container status \"af014f3d948b12e938ae99b89aabd762100d2a622ac15f0606ad12c79347a128\": rpc error: code = NotFound desc = could not find container \"af014f3d948b12e938ae99b89aabd762100d2a622ac15f0606ad12c79347a128\": container with ID starting with af014f3d948b12e938ae99b89aabd762100d2a622ac15f0606ad12c79347a128 not found: ID does not exist" Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.105038 4824 scope.go:117] "RemoveContainer" containerID="c4e76cd725e81ba223348c70595837e4caaba881aaa9eba8437cfc5280bd859c" Jan 21 11:26:59 crc kubenswrapper[4824]: E0121 11:26:59.105308 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c4e76cd725e81ba223348c70595837e4caaba881aaa9eba8437cfc5280bd859c\": container with ID starting with c4e76cd725e81ba223348c70595837e4caaba881aaa9eba8437cfc5280bd859c not found: ID does not exist" containerID="c4e76cd725e81ba223348c70595837e4caaba881aaa9eba8437cfc5280bd859c" Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.105351 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c4e76cd725e81ba223348c70595837e4caaba881aaa9eba8437cfc5280bd859c"} err="failed to get container status \"c4e76cd725e81ba223348c70595837e4caaba881aaa9eba8437cfc5280bd859c\": rpc error: code = NotFound desc = could not find container \"c4e76cd725e81ba223348c70595837e4caaba881aaa9eba8437cfc5280bd859c\": container with ID starting with c4e76cd725e81ba223348c70595837e4caaba881aaa9eba8437cfc5280bd859c not found: ID does not exist" Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.200361 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bae23da5-948a-4883-b88d-d6ed5d6ccba7-scripts\") pod \"ceilometer-0\" (UID: \"bae23da5-948a-4883-b88d-d6ed5d6ccba7\") " pod="openstack/ceilometer-0" Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.200609 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nmjb9\" (UniqueName: \"kubernetes.io/projected/bae23da5-948a-4883-b88d-d6ed5d6ccba7-kube-api-access-nmjb9\") pod \"ceilometer-0\" (UID: \"bae23da5-948a-4883-b88d-d6ed5d6ccba7\") " pod="openstack/ceilometer-0" Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.200652 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bae23da5-948a-4883-b88d-d6ed5d6ccba7-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"bae23da5-948a-4883-b88d-d6ed5d6ccba7\") " pod="openstack/ceilometer-0" Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.200671 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bae23da5-948a-4883-b88d-d6ed5d6ccba7-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"bae23da5-948a-4883-b88d-d6ed5d6ccba7\") " pod="openstack/ceilometer-0" Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.200690 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bae23da5-948a-4883-b88d-d6ed5d6ccba7-log-httpd\") pod \"ceilometer-0\" (UID: \"bae23da5-948a-4883-b88d-d6ed5d6ccba7\") " pod="openstack/ceilometer-0" Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.200791 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bae23da5-948a-4883-b88d-d6ed5d6ccba7-run-httpd\") pod \"ceilometer-0\" (UID: \"bae23da5-948a-4883-b88d-d6ed5d6ccba7\") " pod="openstack/ceilometer-0" Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.200811 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bae23da5-948a-4883-b88d-d6ed5d6ccba7-config-data\") pod \"ceilometer-0\" (UID: \"bae23da5-948a-4883-b88d-d6ed5d6ccba7\") " pod="openstack/ceilometer-0" Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.302217 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bae23da5-948a-4883-b88d-d6ed5d6ccba7-run-httpd\") pod \"ceilometer-0\" (UID: \"bae23da5-948a-4883-b88d-d6ed5d6ccba7\") " pod="openstack/ceilometer-0" Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.302258 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bae23da5-948a-4883-b88d-d6ed5d6ccba7-config-data\") pod \"ceilometer-0\" (UID: \"bae23da5-948a-4883-b88d-d6ed5d6ccba7\") " pod="openstack/ceilometer-0" Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.302343 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bae23da5-948a-4883-b88d-d6ed5d6ccba7-scripts\") pod \"ceilometer-0\" (UID: \"bae23da5-948a-4883-b88d-d6ed5d6ccba7\") " pod="openstack/ceilometer-0" Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.302377 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nmjb9\" (UniqueName: \"kubernetes.io/projected/bae23da5-948a-4883-b88d-d6ed5d6ccba7-kube-api-access-nmjb9\") pod \"ceilometer-0\" (UID: \"bae23da5-948a-4883-b88d-d6ed5d6ccba7\") " pod="openstack/ceilometer-0" Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.302416 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bae23da5-948a-4883-b88d-d6ed5d6ccba7-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"bae23da5-948a-4883-b88d-d6ed5d6ccba7\") " pod="openstack/ceilometer-0" Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.302436 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bae23da5-948a-4883-b88d-d6ed5d6ccba7-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"bae23da5-948a-4883-b88d-d6ed5d6ccba7\") " pod="openstack/ceilometer-0" Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.302456 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bae23da5-948a-4883-b88d-d6ed5d6ccba7-log-httpd\") pod \"ceilometer-0\" (UID: \"bae23da5-948a-4883-b88d-d6ed5d6ccba7\") " pod="openstack/ceilometer-0" Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.302825 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bae23da5-948a-4883-b88d-d6ed5d6ccba7-log-httpd\") pod \"ceilometer-0\" (UID: \"bae23da5-948a-4883-b88d-d6ed5d6ccba7\") " pod="openstack/ceilometer-0" Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.303040 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bae23da5-948a-4883-b88d-d6ed5d6ccba7-run-httpd\") pod \"ceilometer-0\" (UID: \"bae23da5-948a-4883-b88d-d6ed5d6ccba7\") " pod="openstack/ceilometer-0" Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.308038 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bae23da5-948a-4883-b88d-d6ed5d6ccba7-config-data\") pod \"ceilometer-0\" (UID: \"bae23da5-948a-4883-b88d-d6ed5d6ccba7\") " pod="openstack/ceilometer-0" Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.308434 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bae23da5-948a-4883-b88d-d6ed5d6ccba7-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"bae23da5-948a-4883-b88d-d6ed5d6ccba7\") " pod="openstack/ceilometer-0" Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.314134 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bae23da5-948a-4883-b88d-d6ed5d6ccba7-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"bae23da5-948a-4883-b88d-d6ed5d6ccba7\") " pod="openstack/ceilometer-0" Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.315351 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bae23da5-948a-4883-b88d-d6ed5d6ccba7-scripts\") pod \"ceilometer-0\" (UID: \"bae23da5-948a-4883-b88d-d6ed5d6ccba7\") " pod="openstack/ceilometer-0" Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.317879 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nmjb9\" (UniqueName: \"kubernetes.io/projected/bae23da5-948a-4883-b88d-d6ed5d6ccba7-kube-api-access-nmjb9\") pod \"ceilometer-0\" (UID: \"bae23da5-948a-4883-b88d-d6ed5d6ccba7\") " pod="openstack/ceilometer-0" Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.375402 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-ngq69" Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.381163 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-rlp2t" Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.385433 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.505518 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0a6c0628-4522-4bb7-8a82-cc2e019eca2a-scripts\") pod \"0a6c0628-4522-4bb7-8a82-cc2e019eca2a\" (UID: \"0a6c0628-4522-4bb7-8a82-cc2e019eca2a\") " Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.505739 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a6c0628-4522-4bb7-8a82-cc2e019eca2a-combined-ca-bundle\") pod \"0a6c0628-4522-4bb7-8a82-cc2e019eca2a\" (UID: \"0a6c0628-4522-4bb7-8a82-cc2e019eca2a\") " Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.505801 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0a6c0628-4522-4bb7-8a82-cc2e019eca2a-etc-machine-id\") pod \"0a6c0628-4522-4bb7-8a82-cc2e019eca2a\" (UID: \"0a6c0628-4522-4bb7-8a82-cc2e019eca2a\") " Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.505832 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/0a6c0628-4522-4bb7-8a82-cc2e019eca2a-db-sync-config-data\") pod \"0a6c0628-4522-4bb7-8a82-cc2e019eca2a\" (UID: \"0a6c0628-4522-4bb7-8a82-cc2e019eca2a\") " Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.505864 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/262733b9-4179-49e3-aee9-b62197cc89ba-db-sync-config-data\") pod \"262733b9-4179-49e3-aee9-b62197cc89ba\" (UID: \"262733b9-4179-49e3-aee9-b62197cc89ba\") " Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.505887 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/262733b9-4179-49e3-aee9-b62197cc89ba-combined-ca-bundle\") pod \"262733b9-4179-49e3-aee9-b62197cc89ba\" (UID: \"262733b9-4179-49e3-aee9-b62197cc89ba\") " Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.505916 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a6c0628-4522-4bb7-8a82-cc2e019eca2a-config-data\") pod \"0a6c0628-4522-4bb7-8a82-cc2e019eca2a\" (UID: \"0a6c0628-4522-4bb7-8a82-cc2e019eca2a\") " Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.505972 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-flpqh\" (UniqueName: \"kubernetes.io/projected/262733b9-4179-49e3-aee9-b62197cc89ba-kube-api-access-flpqh\") pod \"262733b9-4179-49e3-aee9-b62197cc89ba\" (UID: \"262733b9-4179-49e3-aee9-b62197cc89ba\") " Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.506033 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-msmnx\" (UniqueName: \"kubernetes.io/projected/0a6c0628-4522-4bb7-8a82-cc2e019eca2a-kube-api-access-msmnx\") pod \"0a6c0628-4522-4bb7-8a82-cc2e019eca2a\" (UID: \"0a6c0628-4522-4bb7-8a82-cc2e019eca2a\") " Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.506952 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0a6c0628-4522-4bb7-8a82-cc2e019eca2a-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "0a6c0628-4522-4bb7-8a82-cc2e019eca2a" (UID: "0a6c0628-4522-4bb7-8a82-cc2e019eca2a"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.509791 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a6c0628-4522-4bb7-8a82-cc2e019eca2a-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "0a6c0628-4522-4bb7-8a82-cc2e019eca2a" (UID: "0a6c0628-4522-4bb7-8a82-cc2e019eca2a"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.510198 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a6c0628-4522-4bb7-8a82-cc2e019eca2a-scripts" (OuterVolumeSpecName: "scripts") pod "0a6c0628-4522-4bb7-8a82-cc2e019eca2a" (UID: "0a6c0628-4522-4bb7-8a82-cc2e019eca2a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.510359 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0a6c0628-4522-4bb7-8a82-cc2e019eca2a-kube-api-access-msmnx" (OuterVolumeSpecName: "kube-api-access-msmnx") pod "0a6c0628-4522-4bb7-8a82-cc2e019eca2a" (UID: "0a6c0628-4522-4bb7-8a82-cc2e019eca2a"). InnerVolumeSpecName "kube-api-access-msmnx". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.510780 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/262733b9-4179-49e3-aee9-b62197cc89ba-kube-api-access-flpqh" (OuterVolumeSpecName: "kube-api-access-flpqh") pod "262733b9-4179-49e3-aee9-b62197cc89ba" (UID: "262733b9-4179-49e3-aee9-b62197cc89ba"). InnerVolumeSpecName "kube-api-access-flpqh". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.510846 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/262733b9-4179-49e3-aee9-b62197cc89ba-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "262733b9-4179-49e3-aee9-b62197cc89ba" (UID: "262733b9-4179-49e3-aee9-b62197cc89ba"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.524943 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a6c0628-4522-4bb7-8a82-cc2e019eca2a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0a6c0628-4522-4bb7-8a82-cc2e019eca2a" (UID: "0a6c0628-4522-4bb7-8a82-cc2e019eca2a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.534166 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/262733b9-4179-49e3-aee9-b62197cc89ba-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "262733b9-4179-49e3-aee9-b62197cc89ba" (UID: "262733b9-4179-49e3-aee9-b62197cc89ba"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.542048 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a6c0628-4522-4bb7-8a82-cc2e019eca2a-config-data" (OuterVolumeSpecName: "config-data") pod "0a6c0628-4522-4bb7-8a82-cc2e019eca2a" (UID: "0a6c0628-4522-4bb7-8a82-cc2e019eca2a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.607883 4824 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0a6c0628-4522-4bb7-8a82-cc2e019eca2a-scripts\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.607910 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a6c0628-4522-4bb7-8a82-cc2e019eca2a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.607922 4824 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0a6c0628-4522-4bb7-8a82-cc2e019eca2a-etc-machine-id\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.607931 4824 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/0a6c0628-4522-4bb7-8a82-cc2e019eca2a-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.607939 4824 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/262733b9-4179-49e3-aee9-b62197cc89ba-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.607947 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/262733b9-4179-49e3-aee9-b62197cc89ba-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.607967 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a6c0628-4522-4bb7-8a82-cc2e019eca2a-config-data\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.607975 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-flpqh\" (UniqueName: \"kubernetes.io/projected/262733b9-4179-49e3-aee9-b62197cc89ba-kube-api-access-flpqh\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.607984 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-msmnx\" (UniqueName: \"kubernetes.io/projected/0a6c0628-4522-4bb7-8a82-cc2e019eca2a-kube-api-access-msmnx\") on node \"crc\" DevicePath \"\"" Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.749396 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Jan 21 11:26:59 crc kubenswrapper[4824]: W0121 11:26:59.750870 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbae23da5_948a_4883_b88d_d6ed5d6ccba7.slice/crio-b8cc173acd912ad690d3e7aa92c24c20b537f220a873ef96f5148fc3b668366a WatchSource:0}: Error finding container b8cc173acd912ad690d3e7aa92c24c20b537f220a873ef96f5148fc3b668366a: Status 404 returned error can't find the container with id b8cc173acd912ad690d3e7aa92c24c20b537f220a873ef96f5148fc3b668366a Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.997088 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-rlp2t" event={"ID":"0a6c0628-4522-4bb7-8a82-cc2e019eca2a","Type":"ContainerDied","Data":"877238fc09eb99e400629467da99c2ba4a6e53c0afc16677b1308d3c52849ae9"} Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.997118 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-rlp2t" Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.997137 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="877238fc09eb99e400629467da99c2ba4a6e53c0afc16677b1308d3c52849ae9" Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.999826 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-ngq69" event={"ID":"262733b9-4179-49e3-aee9-b62197cc89ba","Type":"ContainerDied","Data":"2bab663bc55719655bbe7aad12a3a6a58c0374bdaecaed83d199e55c2baa7ddf"} Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.999850 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2bab663bc55719655bbe7aad12a3a6a58c0374bdaecaed83d199e55c2baa7ddf" Jan 21 11:26:59 crc kubenswrapper[4824]: I0121 11:26:59.999888 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-ngq69" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.002319 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bae23da5-948a-4883-b88d-d6ed5d6ccba7","Type":"ContainerStarted","Data":"b8cc173acd912ad690d3e7aa92c24c20b537f220a873ef96f5148fc3b668366a"} Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.042352 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-dn2np" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.058644 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bbefda93-dc56-43c5-b131-56af96ca832b" path="/var/lib/kubelet/pods/bbefda93-dc56-43c5-b131-56af96ca832b/volumes" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.082019 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-dn2np"] Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.236144 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Jan 21 11:27:00 crc kubenswrapper[4824]: E0121 11:27:00.236464 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="262733b9-4179-49e3-aee9-b62197cc89ba" containerName="barbican-db-sync" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.236478 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="262733b9-4179-49e3-aee9-b62197cc89ba" containerName="barbican-db-sync" Jan 21 11:27:00 crc kubenswrapper[4824]: E0121 11:27:00.236486 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a6c0628-4522-4bb7-8a82-cc2e019eca2a" containerName="cinder-db-sync" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.236492 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a6c0628-4522-4bb7-8a82-cc2e019eca2a" containerName="cinder-db-sync" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.236666 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="262733b9-4179-49e3-aee9-b62197cc89ba" containerName="barbican-db-sync" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.236699 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="0a6c0628-4522-4bb7-8a82-cc2e019eca2a" containerName="cinder-db-sync" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.238431 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.242892 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-2xjd5" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.242985 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.243092 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.250524 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.261103 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.295670 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-b895b5785-g4p8w"] Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.325133 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b895b5785-g4p8w" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.368639 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-b895b5785-g4p8w"] Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.428495 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a5721831-a6aa-4d42-b62a-041dbe5c23ad-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"a5721831-a6aa-4d42-b62a-041dbe5c23ad\") " pod="openstack/cinder-scheduler-0" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.428542 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wgb7x\" (UniqueName: \"kubernetes.io/projected/f68506fe-ac80-41e7-87c0-c7b5ac8b5362-kube-api-access-wgb7x\") pod \"dnsmasq-dns-b895b5785-g4p8w\" (UID: \"f68506fe-ac80-41e7-87c0-c7b5ac8b5362\") " pod="openstack/dnsmasq-dns-b895b5785-g4p8w" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.428593 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a5721831-a6aa-4d42-b62a-041dbe5c23ad-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"a5721831-a6aa-4d42-b62a-041dbe5c23ad\") " pod="openstack/cinder-scheduler-0" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.428703 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f68506fe-ac80-41e7-87c0-c7b5ac8b5362-dns-svc\") pod \"dnsmasq-dns-b895b5785-g4p8w\" (UID: \"f68506fe-ac80-41e7-87c0-c7b5ac8b5362\") " pod="openstack/dnsmasq-dns-b895b5785-g4p8w" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.428792 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a5721831-a6aa-4d42-b62a-041dbe5c23ad-scripts\") pod \"cinder-scheduler-0\" (UID: \"a5721831-a6aa-4d42-b62a-041dbe5c23ad\") " pod="openstack/cinder-scheduler-0" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.428874 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5721831-a6aa-4d42-b62a-041dbe5c23ad-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"a5721831-a6aa-4d42-b62a-041dbe5c23ad\") " pod="openstack/cinder-scheduler-0" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.428899 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a5721831-a6aa-4d42-b62a-041dbe5c23ad-config-data\") pod \"cinder-scheduler-0\" (UID: \"a5721831-a6aa-4d42-b62a-041dbe5c23ad\") " pod="openstack/cinder-scheduler-0" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.428940 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f68506fe-ac80-41e7-87c0-c7b5ac8b5362-ovsdbserver-sb\") pod \"dnsmasq-dns-b895b5785-g4p8w\" (UID: \"f68506fe-ac80-41e7-87c0-c7b5ac8b5362\") " pod="openstack/dnsmasq-dns-b895b5785-g4p8w" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.428988 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f68506fe-ac80-41e7-87c0-c7b5ac8b5362-config\") pod \"dnsmasq-dns-b895b5785-g4p8w\" (UID: \"f68506fe-ac80-41e7-87c0-c7b5ac8b5362\") " pod="openstack/dnsmasq-dns-b895b5785-g4p8w" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.429012 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f68506fe-ac80-41e7-87c0-c7b5ac8b5362-dns-swift-storage-0\") pod \"dnsmasq-dns-b895b5785-g4p8w\" (UID: \"f68506fe-ac80-41e7-87c0-c7b5ac8b5362\") " pod="openstack/dnsmasq-dns-b895b5785-g4p8w" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.429038 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cmmwh\" (UniqueName: \"kubernetes.io/projected/a5721831-a6aa-4d42-b62a-041dbe5c23ad-kube-api-access-cmmwh\") pod \"cinder-scheduler-0\" (UID: \"a5721831-a6aa-4d42-b62a-041dbe5c23ad\") " pod="openstack/cinder-scheduler-0" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.429082 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f68506fe-ac80-41e7-87c0-c7b5ac8b5362-ovsdbserver-nb\") pod \"dnsmasq-dns-b895b5785-g4p8w\" (UID: \"f68506fe-ac80-41e7-87c0-c7b5ac8b5362\") " pod="openstack/dnsmasq-dns-b895b5785-g4p8w" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.441295 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.442546 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.445381 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.452641 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.533684 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f68506fe-ac80-41e7-87c0-c7b5ac8b5362-dns-svc\") pod \"dnsmasq-dns-b895b5785-g4p8w\" (UID: \"f68506fe-ac80-41e7-87c0-c7b5ac8b5362\") " pod="openstack/dnsmasq-dns-b895b5785-g4p8w" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.533971 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a5721831-a6aa-4d42-b62a-041dbe5c23ad-scripts\") pod \"cinder-scheduler-0\" (UID: \"a5721831-a6aa-4d42-b62a-041dbe5c23ad\") " pod="openstack/cinder-scheduler-0" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.534028 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5721831-a6aa-4d42-b62a-041dbe5c23ad-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"a5721831-a6aa-4d42-b62a-041dbe5c23ad\") " pod="openstack/cinder-scheduler-0" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.534050 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a5721831-a6aa-4d42-b62a-041dbe5c23ad-config-data\") pod \"cinder-scheduler-0\" (UID: \"a5721831-a6aa-4d42-b62a-041dbe5c23ad\") " pod="openstack/cinder-scheduler-0" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.534075 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f68506fe-ac80-41e7-87c0-c7b5ac8b5362-ovsdbserver-sb\") pod \"dnsmasq-dns-b895b5785-g4p8w\" (UID: \"f68506fe-ac80-41e7-87c0-c7b5ac8b5362\") " pod="openstack/dnsmasq-dns-b895b5785-g4p8w" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.534100 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f68506fe-ac80-41e7-87c0-c7b5ac8b5362-config\") pod \"dnsmasq-dns-b895b5785-g4p8w\" (UID: \"f68506fe-ac80-41e7-87c0-c7b5ac8b5362\") " pod="openstack/dnsmasq-dns-b895b5785-g4p8w" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.534120 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f68506fe-ac80-41e7-87c0-c7b5ac8b5362-dns-swift-storage-0\") pod \"dnsmasq-dns-b895b5785-g4p8w\" (UID: \"f68506fe-ac80-41e7-87c0-c7b5ac8b5362\") " pod="openstack/dnsmasq-dns-b895b5785-g4p8w" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.534141 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cmmwh\" (UniqueName: \"kubernetes.io/projected/a5721831-a6aa-4d42-b62a-041dbe5c23ad-kube-api-access-cmmwh\") pod \"cinder-scheduler-0\" (UID: \"a5721831-a6aa-4d42-b62a-041dbe5c23ad\") " pod="openstack/cinder-scheduler-0" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.534173 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f68506fe-ac80-41e7-87c0-c7b5ac8b5362-ovsdbserver-nb\") pod \"dnsmasq-dns-b895b5785-g4p8w\" (UID: \"f68506fe-ac80-41e7-87c0-c7b5ac8b5362\") " pod="openstack/dnsmasq-dns-b895b5785-g4p8w" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.534239 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a5721831-a6aa-4d42-b62a-041dbe5c23ad-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"a5721831-a6aa-4d42-b62a-041dbe5c23ad\") " pod="openstack/cinder-scheduler-0" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.534266 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wgb7x\" (UniqueName: \"kubernetes.io/projected/f68506fe-ac80-41e7-87c0-c7b5ac8b5362-kube-api-access-wgb7x\") pod \"dnsmasq-dns-b895b5785-g4p8w\" (UID: \"f68506fe-ac80-41e7-87c0-c7b5ac8b5362\") " pod="openstack/dnsmasq-dns-b895b5785-g4p8w" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.534305 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a5721831-a6aa-4d42-b62a-041dbe5c23ad-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"a5721831-a6aa-4d42-b62a-041dbe5c23ad\") " pod="openstack/cinder-scheduler-0" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.534495 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f68506fe-ac80-41e7-87c0-c7b5ac8b5362-dns-svc\") pod \"dnsmasq-dns-b895b5785-g4p8w\" (UID: \"f68506fe-ac80-41e7-87c0-c7b5ac8b5362\") " pod="openstack/dnsmasq-dns-b895b5785-g4p8w" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.534744 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a5721831-a6aa-4d42-b62a-041dbe5c23ad-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"a5721831-a6aa-4d42-b62a-041dbe5c23ad\") " pod="openstack/cinder-scheduler-0" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.535360 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f68506fe-ac80-41e7-87c0-c7b5ac8b5362-ovsdbserver-nb\") pod \"dnsmasq-dns-b895b5785-g4p8w\" (UID: \"f68506fe-ac80-41e7-87c0-c7b5ac8b5362\") " pod="openstack/dnsmasq-dns-b895b5785-g4p8w" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.535947 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f68506fe-ac80-41e7-87c0-c7b5ac8b5362-dns-swift-storage-0\") pod \"dnsmasq-dns-b895b5785-g4p8w\" (UID: \"f68506fe-ac80-41e7-87c0-c7b5ac8b5362\") " pod="openstack/dnsmasq-dns-b895b5785-g4p8w" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.536462 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f68506fe-ac80-41e7-87c0-c7b5ac8b5362-ovsdbserver-sb\") pod \"dnsmasq-dns-b895b5785-g4p8w\" (UID: \"f68506fe-ac80-41e7-87c0-c7b5ac8b5362\") " pod="openstack/dnsmasq-dns-b895b5785-g4p8w" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.537198 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f68506fe-ac80-41e7-87c0-c7b5ac8b5362-config\") pod \"dnsmasq-dns-b895b5785-g4p8w\" (UID: \"f68506fe-ac80-41e7-87c0-c7b5ac8b5362\") " pod="openstack/dnsmasq-dns-b895b5785-g4p8w" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.543461 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5721831-a6aa-4d42-b62a-041dbe5c23ad-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"a5721831-a6aa-4d42-b62a-041dbe5c23ad\") " pod="openstack/cinder-scheduler-0" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.543647 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a5721831-a6aa-4d42-b62a-041dbe5c23ad-config-data\") pod \"cinder-scheduler-0\" (UID: \"a5721831-a6aa-4d42-b62a-041dbe5c23ad\") " pod="openstack/cinder-scheduler-0" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.546563 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a5721831-a6aa-4d42-b62a-041dbe5c23ad-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"a5721831-a6aa-4d42-b62a-041dbe5c23ad\") " pod="openstack/cinder-scheduler-0" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.554581 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a5721831-a6aa-4d42-b62a-041dbe5c23ad-scripts\") pod \"cinder-scheduler-0\" (UID: \"a5721831-a6aa-4d42-b62a-041dbe5c23ad\") " pod="openstack/cinder-scheduler-0" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.568406 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cmmwh\" (UniqueName: \"kubernetes.io/projected/a5721831-a6aa-4d42-b62a-041dbe5c23ad-kube-api-access-cmmwh\") pod \"cinder-scheduler-0\" (UID: \"a5721831-a6aa-4d42-b62a-041dbe5c23ad\") " pod="openstack/cinder-scheduler-0" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.568469 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-7cf777545f-f8drb"] Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.569790 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-7cf777545f-f8drb" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.570439 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.571487 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-7cf777545f-f8drb"] Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.576613 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.576786 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-8q76f" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.576916 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.594274 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wgb7x\" (UniqueName: \"kubernetes.io/projected/f68506fe-ac80-41e7-87c0-c7b5ac8b5362-kube-api-access-wgb7x\") pod \"dnsmasq-dns-b895b5785-g4p8w\" (UID: \"f68506fe-ac80-41e7-87c0-c7b5ac8b5362\") " pod="openstack/dnsmasq-dns-b895b5785-g4p8w" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.596032 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-79b7cfb8cd-ktpp5"] Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.599411 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-79b7cfb8cd-ktpp5" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.606906 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.613257 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-79b7cfb8cd-ktpp5"] Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.635492 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/62025fca-0943-48d2-b139-c55b0c7df358-etc-machine-id\") pod \"cinder-api-0\" (UID: \"62025fca-0943-48d2-b139-c55b0c7df358\") " pod="openstack/cinder-api-0" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.635564 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/62025fca-0943-48d2-b139-c55b0c7df358-logs\") pod \"cinder-api-0\" (UID: \"62025fca-0943-48d2-b139-c55b0c7df358\") " pod="openstack/cinder-api-0" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.635592 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/62025fca-0943-48d2-b139-c55b0c7df358-scripts\") pod \"cinder-api-0\" (UID: \"62025fca-0943-48d2-b139-c55b0c7df358\") " pod="openstack/cinder-api-0" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.635636 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s2qhr\" (UniqueName: \"kubernetes.io/projected/62025fca-0943-48d2-b139-c55b0c7df358-kube-api-access-s2qhr\") pod \"cinder-api-0\" (UID: \"62025fca-0943-48d2-b139-c55b0c7df358\") " pod="openstack/cinder-api-0" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.635681 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62025fca-0943-48d2-b139-c55b0c7df358-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"62025fca-0943-48d2-b139-c55b0c7df358\") " pod="openstack/cinder-api-0" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.635711 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62025fca-0943-48d2-b139-c55b0c7df358-config-data\") pod \"cinder-api-0\" (UID: \"62025fca-0943-48d2-b139-c55b0c7df358\") " pod="openstack/cinder-api-0" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.635759 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/62025fca-0943-48d2-b139-c55b0c7df358-config-data-custom\") pod \"cinder-api-0\" (UID: \"62025fca-0943-48d2-b139-c55b0c7df358\") " pod="openstack/cinder-api-0" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.663496 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-b895b5785-g4p8w"] Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.664182 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b895b5785-g4p8w" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.676472 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-q2xgx"] Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.678859 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-q2xgx" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.692528 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-q2xgx"] Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.737670 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62025fca-0943-48d2-b139-c55b0c7df358-config-data\") pod \"cinder-api-0\" (UID: \"62025fca-0943-48d2-b139-c55b0c7df358\") " pod="openstack/cinder-api-0" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.737737 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/283e6982-7444-48c7-b7ac-ed4210329a08-config-data\") pod \"barbican-worker-7cf777545f-f8drb\" (UID: \"283e6982-7444-48c7-b7ac-ed4210329a08\") " pod="openstack/barbican-worker-7cf777545f-f8drb" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.737759 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/283e6982-7444-48c7-b7ac-ed4210329a08-combined-ca-bundle\") pod \"barbican-worker-7cf777545f-f8drb\" (UID: \"283e6982-7444-48c7-b7ac-ed4210329a08\") " pod="openstack/barbican-worker-7cf777545f-f8drb" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.737792 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/62025fca-0943-48d2-b139-c55b0c7df358-config-data-custom\") pod \"cinder-api-0\" (UID: \"62025fca-0943-48d2-b139-c55b0c7df358\") " pod="openstack/cinder-api-0" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.737806 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8b84edbe-4485-4027-bfc4-cebbb1b10543-config-data\") pod \"barbican-keystone-listener-79b7cfb8cd-ktpp5\" (UID: \"8b84edbe-4485-4027-bfc4-cebbb1b10543\") " pod="openstack/barbican-keystone-listener-79b7cfb8cd-ktpp5" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.737906 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8b84edbe-4485-4027-bfc4-cebbb1b10543-logs\") pod \"barbican-keystone-listener-79b7cfb8cd-ktpp5\" (UID: \"8b84edbe-4485-4027-bfc4-cebbb1b10543\") " pod="openstack/barbican-keystone-listener-79b7cfb8cd-ktpp5" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.737973 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/62025fca-0943-48d2-b139-c55b0c7df358-etc-machine-id\") pod \"cinder-api-0\" (UID: \"62025fca-0943-48d2-b139-c55b0c7df358\") " pod="openstack/cinder-api-0" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.738014 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wdbsk\" (UniqueName: \"kubernetes.io/projected/283e6982-7444-48c7-b7ac-ed4210329a08-kube-api-access-wdbsk\") pod \"barbican-worker-7cf777545f-f8drb\" (UID: \"283e6982-7444-48c7-b7ac-ed4210329a08\") " pod="openstack/barbican-worker-7cf777545f-f8drb" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.738048 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/62025fca-0943-48d2-b139-c55b0c7df358-logs\") pod \"cinder-api-0\" (UID: \"62025fca-0943-48d2-b139-c55b0c7df358\") " pod="openstack/cinder-api-0" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.738067 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/283e6982-7444-48c7-b7ac-ed4210329a08-config-data-custom\") pod \"barbican-worker-7cf777545f-f8drb\" (UID: \"283e6982-7444-48c7-b7ac-ed4210329a08\") " pod="openstack/barbican-worker-7cf777545f-f8drb" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.738088 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/62025fca-0943-48d2-b139-c55b0c7df358-scripts\") pod \"cinder-api-0\" (UID: \"62025fca-0943-48d2-b139-c55b0c7df358\") " pod="openstack/cinder-api-0" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.738129 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2qhr\" (UniqueName: \"kubernetes.io/projected/62025fca-0943-48d2-b139-c55b0c7df358-kube-api-access-s2qhr\") pod \"cinder-api-0\" (UID: \"62025fca-0943-48d2-b139-c55b0c7df358\") " pod="openstack/cinder-api-0" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.738148 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8b84edbe-4485-4027-bfc4-cebbb1b10543-config-data-custom\") pod \"barbican-keystone-listener-79b7cfb8cd-ktpp5\" (UID: \"8b84edbe-4485-4027-bfc4-cebbb1b10543\") " pod="openstack/barbican-keystone-listener-79b7cfb8cd-ktpp5" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.738166 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8b84edbe-4485-4027-bfc4-cebbb1b10543-combined-ca-bundle\") pod \"barbican-keystone-listener-79b7cfb8cd-ktpp5\" (UID: \"8b84edbe-4485-4027-bfc4-cebbb1b10543\") " pod="openstack/barbican-keystone-listener-79b7cfb8cd-ktpp5" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.738180 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/283e6982-7444-48c7-b7ac-ed4210329a08-logs\") pod \"barbican-worker-7cf777545f-f8drb\" (UID: \"283e6982-7444-48c7-b7ac-ed4210329a08\") " pod="openstack/barbican-worker-7cf777545f-f8drb" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.738205 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62025fca-0943-48d2-b139-c55b0c7df358-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"62025fca-0943-48d2-b139-c55b0c7df358\") " pod="openstack/cinder-api-0" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.738232 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-spbcb\" (UniqueName: \"kubernetes.io/projected/8b84edbe-4485-4027-bfc4-cebbb1b10543-kube-api-access-spbcb\") pod \"barbican-keystone-listener-79b7cfb8cd-ktpp5\" (UID: \"8b84edbe-4485-4027-bfc4-cebbb1b10543\") " pod="openstack/barbican-keystone-listener-79b7cfb8cd-ktpp5" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.739496 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/62025fca-0943-48d2-b139-c55b0c7df358-logs\") pod \"cinder-api-0\" (UID: \"62025fca-0943-48d2-b139-c55b0c7df358\") " pod="openstack/cinder-api-0" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.746800 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/62025fca-0943-48d2-b139-c55b0c7df358-scripts\") pod \"cinder-api-0\" (UID: \"62025fca-0943-48d2-b139-c55b0c7df358\") " pod="openstack/cinder-api-0" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.748058 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/62025fca-0943-48d2-b139-c55b0c7df358-etc-machine-id\") pod \"cinder-api-0\" (UID: \"62025fca-0943-48d2-b139-c55b0c7df358\") " pod="openstack/cinder-api-0" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.749500 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/62025fca-0943-48d2-b139-c55b0c7df358-config-data-custom\") pod \"cinder-api-0\" (UID: \"62025fca-0943-48d2-b139-c55b0c7df358\") " pod="openstack/cinder-api-0" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.749685 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62025fca-0943-48d2-b139-c55b0c7df358-config-data\") pod \"cinder-api-0\" (UID: \"62025fca-0943-48d2-b139-c55b0c7df358\") " pod="openstack/cinder-api-0" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.758329 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62025fca-0943-48d2-b139-c55b0c7df358-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"62025fca-0943-48d2-b139-c55b0c7df358\") " pod="openstack/cinder-api-0" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.763796 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-6b9b444cf4-q85qq"] Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.765168 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6b9b444cf4-q85qq" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.766135 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2qhr\" (UniqueName: \"kubernetes.io/projected/62025fca-0943-48d2-b139-c55b0c7df358-kube-api-access-s2qhr\") pod \"cinder-api-0\" (UID: \"62025fca-0943-48d2-b139-c55b0c7df358\") " pod="openstack/cinder-api-0" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.770883 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.774791 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.778302 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-6b9b444cf4-q85qq"] Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.840855 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/661c81f6-6854-416e-aea7-1344a635f2db-config\") pod \"dnsmasq-dns-5c9776ccc5-q2xgx\" (UID: \"661c81f6-6854-416e-aea7-1344a635f2db\") " pod="openstack/dnsmasq-dns-5c9776ccc5-q2xgx" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.840941 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/283e6982-7444-48c7-b7ac-ed4210329a08-config-data-custom\") pod \"barbican-worker-7cf777545f-f8drb\" (UID: \"283e6982-7444-48c7-b7ac-ed4210329a08\") " pod="openstack/barbican-worker-7cf777545f-f8drb" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.840983 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/661c81f6-6854-416e-aea7-1344a635f2db-dns-svc\") pod \"dnsmasq-dns-5c9776ccc5-q2xgx\" (UID: \"661c81f6-6854-416e-aea7-1344a635f2db\") " pod="openstack/dnsmasq-dns-5c9776ccc5-q2xgx" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.841012 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/661c81f6-6854-416e-aea7-1344a635f2db-ovsdbserver-nb\") pod \"dnsmasq-dns-5c9776ccc5-q2xgx\" (UID: \"661c81f6-6854-416e-aea7-1344a635f2db\") " pod="openstack/dnsmasq-dns-5c9776ccc5-q2xgx" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.841034 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/661c81f6-6854-416e-aea7-1344a635f2db-ovsdbserver-sb\") pod \"dnsmasq-dns-5c9776ccc5-q2xgx\" (UID: \"661c81f6-6854-416e-aea7-1344a635f2db\") " pod="openstack/dnsmasq-dns-5c9776ccc5-q2xgx" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.841057 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8b84edbe-4485-4027-bfc4-cebbb1b10543-config-data-custom\") pod \"barbican-keystone-listener-79b7cfb8cd-ktpp5\" (UID: \"8b84edbe-4485-4027-bfc4-cebbb1b10543\") " pod="openstack/barbican-keystone-listener-79b7cfb8cd-ktpp5" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.841072 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8b84edbe-4485-4027-bfc4-cebbb1b10543-combined-ca-bundle\") pod \"barbican-keystone-listener-79b7cfb8cd-ktpp5\" (UID: \"8b84edbe-4485-4027-bfc4-cebbb1b10543\") " pod="openstack/barbican-keystone-listener-79b7cfb8cd-ktpp5" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.841089 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/283e6982-7444-48c7-b7ac-ed4210329a08-logs\") pod \"barbican-worker-7cf777545f-f8drb\" (UID: \"283e6982-7444-48c7-b7ac-ed4210329a08\") " pod="openstack/barbican-worker-7cf777545f-f8drb" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.841118 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-spbcb\" (UniqueName: \"kubernetes.io/projected/8b84edbe-4485-4027-bfc4-cebbb1b10543-kube-api-access-spbcb\") pod \"barbican-keystone-listener-79b7cfb8cd-ktpp5\" (UID: \"8b84edbe-4485-4027-bfc4-cebbb1b10543\") " pod="openstack/barbican-keystone-listener-79b7cfb8cd-ktpp5" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.841142 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/283e6982-7444-48c7-b7ac-ed4210329a08-config-data\") pod \"barbican-worker-7cf777545f-f8drb\" (UID: \"283e6982-7444-48c7-b7ac-ed4210329a08\") " pod="openstack/barbican-worker-7cf777545f-f8drb" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.841157 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/283e6982-7444-48c7-b7ac-ed4210329a08-combined-ca-bundle\") pod \"barbican-worker-7cf777545f-f8drb\" (UID: \"283e6982-7444-48c7-b7ac-ed4210329a08\") " pod="openstack/barbican-worker-7cf777545f-f8drb" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.841177 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8b84edbe-4485-4027-bfc4-cebbb1b10543-config-data\") pod \"barbican-keystone-listener-79b7cfb8cd-ktpp5\" (UID: \"8b84edbe-4485-4027-bfc4-cebbb1b10543\") " pod="openstack/barbican-keystone-listener-79b7cfb8cd-ktpp5" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.841206 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/661c81f6-6854-416e-aea7-1344a635f2db-dns-swift-storage-0\") pod \"dnsmasq-dns-5c9776ccc5-q2xgx\" (UID: \"661c81f6-6854-416e-aea7-1344a635f2db\") " pod="openstack/dnsmasq-dns-5c9776ccc5-q2xgx" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.841233 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s4ttl\" (UniqueName: \"kubernetes.io/projected/661c81f6-6854-416e-aea7-1344a635f2db-kube-api-access-s4ttl\") pod \"dnsmasq-dns-5c9776ccc5-q2xgx\" (UID: \"661c81f6-6854-416e-aea7-1344a635f2db\") " pod="openstack/dnsmasq-dns-5c9776ccc5-q2xgx" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.841261 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8b84edbe-4485-4027-bfc4-cebbb1b10543-logs\") pod \"barbican-keystone-listener-79b7cfb8cd-ktpp5\" (UID: \"8b84edbe-4485-4027-bfc4-cebbb1b10543\") " pod="openstack/barbican-keystone-listener-79b7cfb8cd-ktpp5" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.841306 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wdbsk\" (UniqueName: \"kubernetes.io/projected/283e6982-7444-48c7-b7ac-ed4210329a08-kube-api-access-wdbsk\") pod \"barbican-worker-7cf777545f-f8drb\" (UID: \"283e6982-7444-48c7-b7ac-ed4210329a08\") " pod="openstack/barbican-worker-7cf777545f-f8drb" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.841775 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/283e6982-7444-48c7-b7ac-ed4210329a08-logs\") pod \"barbican-worker-7cf777545f-f8drb\" (UID: \"283e6982-7444-48c7-b7ac-ed4210329a08\") " pod="openstack/barbican-worker-7cf777545f-f8drb" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.842078 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8b84edbe-4485-4027-bfc4-cebbb1b10543-logs\") pod \"barbican-keystone-listener-79b7cfb8cd-ktpp5\" (UID: \"8b84edbe-4485-4027-bfc4-cebbb1b10543\") " pod="openstack/barbican-keystone-listener-79b7cfb8cd-ktpp5" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.846461 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/283e6982-7444-48c7-b7ac-ed4210329a08-config-data\") pod \"barbican-worker-7cf777545f-f8drb\" (UID: \"283e6982-7444-48c7-b7ac-ed4210329a08\") " pod="openstack/barbican-worker-7cf777545f-f8drb" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.847499 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/283e6982-7444-48c7-b7ac-ed4210329a08-config-data-custom\") pod \"barbican-worker-7cf777545f-f8drb\" (UID: \"283e6982-7444-48c7-b7ac-ed4210329a08\") " pod="openstack/barbican-worker-7cf777545f-f8drb" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.848617 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/283e6982-7444-48c7-b7ac-ed4210329a08-combined-ca-bundle\") pod \"barbican-worker-7cf777545f-f8drb\" (UID: \"283e6982-7444-48c7-b7ac-ed4210329a08\") " pod="openstack/barbican-worker-7cf777545f-f8drb" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.848793 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8b84edbe-4485-4027-bfc4-cebbb1b10543-combined-ca-bundle\") pod \"barbican-keystone-listener-79b7cfb8cd-ktpp5\" (UID: \"8b84edbe-4485-4027-bfc4-cebbb1b10543\") " pod="openstack/barbican-keystone-listener-79b7cfb8cd-ktpp5" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.849419 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8b84edbe-4485-4027-bfc4-cebbb1b10543-config-data-custom\") pod \"barbican-keystone-listener-79b7cfb8cd-ktpp5\" (UID: \"8b84edbe-4485-4027-bfc4-cebbb1b10543\") " pod="openstack/barbican-keystone-listener-79b7cfb8cd-ktpp5" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.849878 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8b84edbe-4485-4027-bfc4-cebbb1b10543-config-data\") pod \"barbican-keystone-listener-79b7cfb8cd-ktpp5\" (UID: \"8b84edbe-4485-4027-bfc4-cebbb1b10543\") " pod="openstack/barbican-keystone-listener-79b7cfb8cd-ktpp5" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.857140 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-spbcb\" (UniqueName: \"kubernetes.io/projected/8b84edbe-4485-4027-bfc4-cebbb1b10543-kube-api-access-spbcb\") pod \"barbican-keystone-listener-79b7cfb8cd-ktpp5\" (UID: \"8b84edbe-4485-4027-bfc4-cebbb1b10543\") " pod="openstack/barbican-keystone-listener-79b7cfb8cd-ktpp5" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.857940 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wdbsk\" (UniqueName: \"kubernetes.io/projected/283e6982-7444-48c7-b7ac-ed4210329a08-kube-api-access-wdbsk\") pod \"barbican-worker-7cf777545f-f8drb\" (UID: \"283e6982-7444-48c7-b7ac-ed4210329a08\") " pod="openstack/barbican-worker-7cf777545f-f8drb" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.898499 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-7cf777545f-f8drb" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.926707 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-79b7cfb8cd-ktpp5" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.942080 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4ed8e1b1-57c3-4b01-bbd7-9dc5125f4f94-config-data\") pod \"barbican-api-6b9b444cf4-q85qq\" (UID: \"4ed8e1b1-57c3-4b01-bbd7-9dc5125f4f94\") " pod="openstack/barbican-api-6b9b444cf4-q85qq" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.942127 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ed8e1b1-57c3-4b01-bbd7-9dc5125f4f94-combined-ca-bundle\") pod \"barbican-api-6b9b444cf4-q85qq\" (UID: \"4ed8e1b1-57c3-4b01-bbd7-9dc5125f4f94\") " pod="openstack/barbican-api-6b9b444cf4-q85qq" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.942163 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/661c81f6-6854-416e-aea7-1344a635f2db-config\") pod \"dnsmasq-dns-5c9776ccc5-q2xgx\" (UID: \"661c81f6-6854-416e-aea7-1344a635f2db\") " pod="openstack/dnsmasq-dns-5c9776ccc5-q2xgx" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.942187 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/661c81f6-6854-416e-aea7-1344a635f2db-dns-svc\") pod \"dnsmasq-dns-5c9776ccc5-q2xgx\" (UID: \"661c81f6-6854-416e-aea7-1344a635f2db\") " pod="openstack/dnsmasq-dns-5c9776ccc5-q2xgx" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.942894 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/661c81f6-6854-416e-aea7-1344a635f2db-dns-svc\") pod \"dnsmasq-dns-5c9776ccc5-q2xgx\" (UID: \"661c81f6-6854-416e-aea7-1344a635f2db\") " pod="openstack/dnsmasq-dns-5c9776ccc5-q2xgx" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.942945 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/661c81f6-6854-416e-aea7-1344a635f2db-ovsdbserver-nb\") pod \"dnsmasq-dns-5c9776ccc5-q2xgx\" (UID: \"661c81f6-6854-416e-aea7-1344a635f2db\") " pod="openstack/dnsmasq-dns-5c9776ccc5-q2xgx" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.942991 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/661c81f6-6854-416e-aea7-1344a635f2db-config\") pod \"dnsmasq-dns-5c9776ccc5-q2xgx\" (UID: \"661c81f6-6854-416e-aea7-1344a635f2db\") " pod="openstack/dnsmasq-dns-5c9776ccc5-q2xgx" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.943551 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/661c81f6-6854-416e-aea7-1344a635f2db-ovsdbserver-sb\") pod \"dnsmasq-dns-5c9776ccc5-q2xgx\" (UID: \"661c81f6-6854-416e-aea7-1344a635f2db\") " pod="openstack/dnsmasq-dns-5c9776ccc5-q2xgx" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.943001 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/661c81f6-6854-416e-aea7-1344a635f2db-ovsdbserver-sb\") pod \"dnsmasq-dns-5c9776ccc5-q2xgx\" (UID: \"661c81f6-6854-416e-aea7-1344a635f2db\") " pod="openstack/dnsmasq-dns-5c9776ccc5-q2xgx" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.943616 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/661c81f6-6854-416e-aea7-1344a635f2db-ovsdbserver-nb\") pod \"dnsmasq-dns-5c9776ccc5-q2xgx\" (UID: \"661c81f6-6854-416e-aea7-1344a635f2db\") " pod="openstack/dnsmasq-dns-5c9776ccc5-q2xgx" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.943660 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vnlw6\" (UniqueName: \"kubernetes.io/projected/4ed8e1b1-57c3-4b01-bbd7-9dc5125f4f94-kube-api-access-vnlw6\") pod \"barbican-api-6b9b444cf4-q85qq\" (UID: \"4ed8e1b1-57c3-4b01-bbd7-9dc5125f4f94\") " pod="openstack/barbican-api-6b9b444cf4-q85qq" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.943777 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4ed8e1b1-57c3-4b01-bbd7-9dc5125f4f94-logs\") pod \"barbican-api-6b9b444cf4-q85qq\" (UID: \"4ed8e1b1-57c3-4b01-bbd7-9dc5125f4f94\") " pod="openstack/barbican-api-6b9b444cf4-q85qq" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.943800 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/661c81f6-6854-416e-aea7-1344a635f2db-dns-swift-storage-0\") pod \"dnsmasq-dns-5c9776ccc5-q2xgx\" (UID: \"661c81f6-6854-416e-aea7-1344a635f2db\") " pod="openstack/dnsmasq-dns-5c9776ccc5-q2xgx" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.943824 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4ed8e1b1-57c3-4b01-bbd7-9dc5125f4f94-config-data-custom\") pod \"barbican-api-6b9b444cf4-q85qq\" (UID: \"4ed8e1b1-57c3-4b01-bbd7-9dc5125f4f94\") " pod="openstack/barbican-api-6b9b444cf4-q85qq" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.943844 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s4ttl\" (UniqueName: \"kubernetes.io/projected/661c81f6-6854-416e-aea7-1344a635f2db-kube-api-access-s4ttl\") pod \"dnsmasq-dns-5c9776ccc5-q2xgx\" (UID: \"661c81f6-6854-416e-aea7-1344a635f2db\") " pod="openstack/dnsmasq-dns-5c9776ccc5-q2xgx" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.944363 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/661c81f6-6854-416e-aea7-1344a635f2db-dns-swift-storage-0\") pod \"dnsmasq-dns-5c9776ccc5-q2xgx\" (UID: \"661c81f6-6854-416e-aea7-1344a635f2db\") " pod="openstack/dnsmasq-dns-5c9776ccc5-q2xgx" Jan 21 11:27:00 crc kubenswrapper[4824]: I0121 11:27:00.960852 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s4ttl\" (UniqueName: \"kubernetes.io/projected/661c81f6-6854-416e-aea7-1344a635f2db-kube-api-access-s4ttl\") pod \"dnsmasq-dns-5c9776ccc5-q2xgx\" (UID: \"661c81f6-6854-416e-aea7-1344a635f2db\") " pod="openstack/dnsmasq-dns-5c9776ccc5-q2xgx" Jan 21 11:27:01 crc kubenswrapper[4824]: I0121 11:27:01.006764 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-q2xgx" Jan 21 11:27:01 crc kubenswrapper[4824]: I0121 11:27:01.013937 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bae23da5-948a-4883-b88d-d6ed5d6ccba7","Type":"ContainerStarted","Data":"7b2e891014d0053a1d16ec02f1fc280a3a0cbb96983be7316f8030e04e57c2be"} Jan 21 11:27:01 crc kubenswrapper[4824]: I0121 11:27:01.044946 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4ed8e1b1-57c3-4b01-bbd7-9dc5125f4f94-config-data\") pod \"barbican-api-6b9b444cf4-q85qq\" (UID: \"4ed8e1b1-57c3-4b01-bbd7-9dc5125f4f94\") " pod="openstack/barbican-api-6b9b444cf4-q85qq" Jan 21 11:27:01 crc kubenswrapper[4824]: I0121 11:27:01.045018 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ed8e1b1-57c3-4b01-bbd7-9dc5125f4f94-combined-ca-bundle\") pod \"barbican-api-6b9b444cf4-q85qq\" (UID: \"4ed8e1b1-57c3-4b01-bbd7-9dc5125f4f94\") " pod="openstack/barbican-api-6b9b444cf4-q85qq" Jan 21 11:27:01 crc kubenswrapper[4824]: I0121 11:27:01.045122 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vnlw6\" (UniqueName: \"kubernetes.io/projected/4ed8e1b1-57c3-4b01-bbd7-9dc5125f4f94-kube-api-access-vnlw6\") pod \"barbican-api-6b9b444cf4-q85qq\" (UID: \"4ed8e1b1-57c3-4b01-bbd7-9dc5125f4f94\") " pod="openstack/barbican-api-6b9b444cf4-q85qq" Jan 21 11:27:01 crc kubenswrapper[4824]: I0121 11:27:01.045184 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4ed8e1b1-57c3-4b01-bbd7-9dc5125f4f94-logs\") pod \"barbican-api-6b9b444cf4-q85qq\" (UID: \"4ed8e1b1-57c3-4b01-bbd7-9dc5125f4f94\") " pod="openstack/barbican-api-6b9b444cf4-q85qq" Jan 21 11:27:01 crc kubenswrapper[4824]: I0121 11:27:01.045211 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4ed8e1b1-57c3-4b01-bbd7-9dc5125f4f94-config-data-custom\") pod \"barbican-api-6b9b444cf4-q85qq\" (UID: \"4ed8e1b1-57c3-4b01-bbd7-9dc5125f4f94\") " pod="openstack/barbican-api-6b9b444cf4-q85qq" Jan 21 11:27:01 crc kubenswrapper[4824]: I0121 11:27:01.046182 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4ed8e1b1-57c3-4b01-bbd7-9dc5125f4f94-logs\") pod \"barbican-api-6b9b444cf4-q85qq\" (UID: \"4ed8e1b1-57c3-4b01-bbd7-9dc5125f4f94\") " pod="openstack/barbican-api-6b9b444cf4-q85qq" Jan 21 11:27:01 crc kubenswrapper[4824]: I0121 11:27:01.049606 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4ed8e1b1-57c3-4b01-bbd7-9dc5125f4f94-config-data\") pod \"barbican-api-6b9b444cf4-q85qq\" (UID: \"4ed8e1b1-57c3-4b01-bbd7-9dc5125f4f94\") " pod="openstack/barbican-api-6b9b444cf4-q85qq" Jan 21 11:27:01 crc kubenswrapper[4824]: I0121 11:27:01.053151 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ed8e1b1-57c3-4b01-bbd7-9dc5125f4f94-combined-ca-bundle\") pod \"barbican-api-6b9b444cf4-q85qq\" (UID: \"4ed8e1b1-57c3-4b01-bbd7-9dc5125f4f94\") " pod="openstack/barbican-api-6b9b444cf4-q85qq" Jan 21 11:27:01 crc kubenswrapper[4824]: I0121 11:27:01.054160 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4ed8e1b1-57c3-4b01-bbd7-9dc5125f4f94-config-data-custom\") pod \"barbican-api-6b9b444cf4-q85qq\" (UID: \"4ed8e1b1-57c3-4b01-bbd7-9dc5125f4f94\") " pod="openstack/barbican-api-6b9b444cf4-q85qq" Jan 21 11:27:01 crc kubenswrapper[4824]: I0121 11:27:01.062419 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vnlw6\" (UniqueName: \"kubernetes.io/projected/4ed8e1b1-57c3-4b01-bbd7-9dc5125f4f94-kube-api-access-vnlw6\") pod \"barbican-api-6b9b444cf4-q85qq\" (UID: \"4ed8e1b1-57c3-4b01-bbd7-9dc5125f4f94\") " pod="openstack/barbican-api-6b9b444cf4-q85qq" Jan 21 11:27:01 crc kubenswrapper[4824]: I0121 11:27:01.102462 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6b9b444cf4-q85qq" Jan 21 11:27:01 crc kubenswrapper[4824]: I0121 11:27:01.118071 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-b895b5785-g4p8w"] Jan 21 11:27:01 crc kubenswrapper[4824]: I0121 11:27:01.154689 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Jan 21 11:27:01 crc kubenswrapper[4824]: I0121 11:27:01.234406 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Jan 21 11:27:01 crc kubenswrapper[4824]: W0121 11:27:01.302027 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda5721831_a6aa_4d42_b62a_041dbe5c23ad.slice/crio-ccfadddd012e5287b137e7bb05a0f3a68e04a10bc60fc278e3e35eab10438b65 WatchSource:0}: Error finding container ccfadddd012e5287b137e7bb05a0f3a68e04a10bc60fc278e3e35eab10438b65: Status 404 returned error can't find the container with id ccfadddd012e5287b137e7bb05a0f3a68e04a10bc60fc278e3e35eab10438b65 Jan 21 11:27:01 crc kubenswrapper[4824]: I0121 11:27:01.338859 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-7cf777545f-f8drb"] Jan 21 11:27:01 crc kubenswrapper[4824]: I0121 11:27:01.438535 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-79b7cfb8cd-ktpp5"] Jan 21 11:27:01 crc kubenswrapper[4824]: W0121 11:27:01.440899 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8b84edbe_4485_4027_bfc4_cebbb1b10543.slice/crio-55fc42714f1e8c2485a2f2ab2991b189f14890bb62beec8d56c4a223261fc69c WatchSource:0}: Error finding container 55fc42714f1e8c2485a2f2ab2991b189f14890bb62beec8d56c4a223261fc69c: Status 404 returned error can't find the container with id 55fc42714f1e8c2485a2f2ab2991b189f14890bb62beec8d56c4a223261fc69c Jan 21 11:27:01 crc kubenswrapper[4824]: I0121 11:27:01.490768 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-q2xgx"] Jan 21 11:27:01 crc kubenswrapper[4824]: W0121 11:27:01.502357 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod661c81f6_6854_416e_aea7_1344a635f2db.slice/crio-e5f93c719de6d721a85fa8dea128d5a96c33c0da5f5e176c82969aadb4af41ac WatchSource:0}: Error finding container e5f93c719de6d721a85fa8dea128d5a96c33c0da5f5e176c82969aadb4af41ac: Status 404 returned error can't find the container with id e5f93c719de6d721a85fa8dea128d5a96c33c0da5f5e176c82969aadb4af41ac Jan 21 11:27:01 crc kubenswrapper[4824]: I0121 11:27:01.715221 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-6b9b444cf4-q85qq"] Jan 21 11:27:01 crc kubenswrapper[4824]: W0121 11:27:01.777363 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4ed8e1b1_57c3_4b01_bbd7_9dc5125f4f94.slice/crio-62531d41bed5d935b440d701f3d3a82064300927572a77364b43daf9ada74a10 WatchSource:0}: Error finding container 62531d41bed5d935b440d701f3d3a82064300927572a77364b43daf9ada74a10: Status 404 returned error can't find the container with id 62531d41bed5d935b440d701f3d3a82064300927572a77364b43daf9ada74a10 Jan 21 11:27:02 crc kubenswrapper[4824]: I0121 11:27:02.024329 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6b9b444cf4-q85qq" event={"ID":"4ed8e1b1-57c3-4b01-bbd7-9dc5125f4f94","Type":"ContainerStarted","Data":"41e3cbe62a3cfe1f1349c3e05b9d5169c6ca9dd78e1f915854df40b91a504d55"} Jan 21 11:27:02 crc kubenswrapper[4824]: I0121 11:27:02.024574 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6b9b444cf4-q85qq" event={"ID":"4ed8e1b1-57c3-4b01-bbd7-9dc5125f4f94","Type":"ContainerStarted","Data":"62531d41bed5d935b440d701f3d3a82064300927572a77364b43daf9ada74a10"} Jan 21 11:27:02 crc kubenswrapper[4824]: I0121 11:27:02.026163 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bae23da5-948a-4883-b88d-d6ed5d6ccba7","Type":"ContainerStarted","Data":"3b7c3d7bba0374d53d0ac03b9cb2de415ba41cc2525693fb788df9d2d6551fe4"} Jan 21 11:27:02 crc kubenswrapper[4824]: I0121 11:27:02.028256 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"a5721831-a6aa-4d42-b62a-041dbe5c23ad","Type":"ContainerStarted","Data":"ccfadddd012e5287b137e7bb05a0f3a68e04a10bc60fc278e3e35eab10438b65"} Jan 21 11:27:02 crc kubenswrapper[4824]: I0121 11:27:02.029395 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-7cf777545f-f8drb" event={"ID":"283e6982-7444-48c7-b7ac-ed4210329a08","Type":"ContainerStarted","Data":"0cdcd0fb4815f9aa1c23324306a72652651347364577112954dc427dfd596ae4"} Jan 21 11:27:02 crc kubenswrapper[4824]: I0121 11:27:02.030859 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-79b7cfb8cd-ktpp5" event={"ID":"8b84edbe-4485-4027-bfc4-cebbb1b10543","Type":"ContainerStarted","Data":"55fc42714f1e8c2485a2f2ab2991b189f14890bb62beec8d56c4a223261fc69c"} Jan 21 11:27:02 crc kubenswrapper[4824]: I0121 11:27:02.032289 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"62025fca-0943-48d2-b139-c55b0c7df358","Type":"ContainerStarted","Data":"0f1a05d13f073268720b76257e1f39bfa980627e6f98480d79c357a688a740d9"} Jan 21 11:27:02 crc kubenswrapper[4824]: I0121 11:27:02.032315 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"62025fca-0943-48d2-b139-c55b0c7df358","Type":"ContainerStarted","Data":"54ae046171ebb1a5c5a1082e7114ac886653fcd28b5ecbb331be434df22b2e10"} Jan 21 11:27:02 crc kubenswrapper[4824]: I0121 11:27:02.039690 4824 generic.go:334] "Generic (PLEG): container finished" podID="661c81f6-6854-416e-aea7-1344a635f2db" containerID="18cc2896c05b5182d3e266af8c0e8d6f107da5daa9b81a2e4d5ef8e0cf371e65" exitCode=0 Jan 21 11:27:02 crc kubenswrapper[4824]: I0121 11:27:02.039760 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-q2xgx" event={"ID":"661c81f6-6854-416e-aea7-1344a635f2db","Type":"ContainerDied","Data":"18cc2896c05b5182d3e266af8c0e8d6f107da5daa9b81a2e4d5ef8e0cf371e65"} Jan 21 11:27:02 crc kubenswrapper[4824]: I0121 11:27:02.039786 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-q2xgx" event={"ID":"661c81f6-6854-416e-aea7-1344a635f2db","Type":"ContainerStarted","Data":"e5f93c719de6d721a85fa8dea128d5a96c33c0da5f5e176c82969aadb4af41ac"} Jan 21 11:27:02 crc kubenswrapper[4824]: I0121 11:27:02.042039 4824 generic.go:334] "Generic (PLEG): container finished" podID="f68506fe-ac80-41e7-87c0-c7b5ac8b5362" containerID="ce260d4bf12b843ead9c0a8144fa07e77b168be5d263f9833e8d46c101318a5d" exitCode=0 Jan 21 11:27:02 crc kubenswrapper[4824]: I0121 11:27:02.042101 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b895b5785-g4p8w" event={"ID":"f68506fe-ac80-41e7-87c0-c7b5ac8b5362","Type":"ContainerDied","Data":"ce260d4bf12b843ead9c0a8144fa07e77b168be5d263f9833e8d46c101318a5d"} Jan 21 11:27:02 crc kubenswrapper[4824]: I0121 11:27:02.042129 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b895b5785-g4p8w" event={"ID":"f68506fe-ac80-41e7-87c0-c7b5ac8b5362","Type":"ContainerStarted","Data":"9f0d2bd9547e5b9af28f1727793676788462d4c9d941a44af9285262e1676154"} Jan 21 11:27:02 crc kubenswrapper[4824]: I0121 11:27:02.042333 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-dn2np" podUID="1d8c8cff-4340-4eb4-94b7-062968a14848" containerName="registry-server" containerID="cri-o://42cfb3a03e211fbd5fa55f6c859e8ec840fddea6ec887daa75d38ab58da624f4" gracePeriod=2 Jan 21 11:27:02 crc kubenswrapper[4824]: I0121 11:27:02.329649 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Jan 21 11:27:02 crc kubenswrapper[4824]: I0121 11:27:02.353001 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b895b5785-g4p8w" Jan 21 11:27:02 crc kubenswrapper[4824]: I0121 11:27:02.482932 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f68506fe-ac80-41e7-87c0-c7b5ac8b5362-dns-swift-storage-0\") pod \"f68506fe-ac80-41e7-87c0-c7b5ac8b5362\" (UID: \"f68506fe-ac80-41e7-87c0-c7b5ac8b5362\") " Jan 21 11:27:02 crc kubenswrapper[4824]: I0121 11:27:02.483009 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f68506fe-ac80-41e7-87c0-c7b5ac8b5362-config\") pod \"f68506fe-ac80-41e7-87c0-c7b5ac8b5362\" (UID: \"f68506fe-ac80-41e7-87c0-c7b5ac8b5362\") " Jan 21 11:27:02 crc kubenswrapper[4824]: I0121 11:27:02.483043 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f68506fe-ac80-41e7-87c0-c7b5ac8b5362-ovsdbserver-sb\") pod \"f68506fe-ac80-41e7-87c0-c7b5ac8b5362\" (UID: \"f68506fe-ac80-41e7-87c0-c7b5ac8b5362\") " Jan 21 11:27:02 crc kubenswrapper[4824]: I0121 11:27:02.483114 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wgb7x\" (UniqueName: \"kubernetes.io/projected/f68506fe-ac80-41e7-87c0-c7b5ac8b5362-kube-api-access-wgb7x\") pod \"f68506fe-ac80-41e7-87c0-c7b5ac8b5362\" (UID: \"f68506fe-ac80-41e7-87c0-c7b5ac8b5362\") " Jan 21 11:27:02 crc kubenswrapper[4824]: I0121 11:27:02.483194 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f68506fe-ac80-41e7-87c0-c7b5ac8b5362-dns-svc\") pod \"f68506fe-ac80-41e7-87c0-c7b5ac8b5362\" (UID: \"f68506fe-ac80-41e7-87c0-c7b5ac8b5362\") " Jan 21 11:27:02 crc kubenswrapper[4824]: I0121 11:27:02.483239 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f68506fe-ac80-41e7-87c0-c7b5ac8b5362-ovsdbserver-nb\") pod \"f68506fe-ac80-41e7-87c0-c7b5ac8b5362\" (UID: \"f68506fe-ac80-41e7-87c0-c7b5ac8b5362\") " Jan 21 11:27:02 crc kubenswrapper[4824]: I0121 11:27:02.497828 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f68506fe-ac80-41e7-87c0-c7b5ac8b5362-kube-api-access-wgb7x" (OuterVolumeSpecName: "kube-api-access-wgb7x") pod "f68506fe-ac80-41e7-87c0-c7b5ac8b5362" (UID: "f68506fe-ac80-41e7-87c0-c7b5ac8b5362"). InnerVolumeSpecName "kube-api-access-wgb7x". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:27:02 crc kubenswrapper[4824]: I0121 11:27:02.509757 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dn2np" Jan 21 11:27:02 crc kubenswrapper[4824]: I0121 11:27:02.513536 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f68506fe-ac80-41e7-87c0-c7b5ac8b5362-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "f68506fe-ac80-41e7-87c0-c7b5ac8b5362" (UID: "f68506fe-ac80-41e7-87c0-c7b5ac8b5362"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:27:02 crc kubenswrapper[4824]: I0121 11:27:02.521988 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f68506fe-ac80-41e7-87c0-c7b5ac8b5362-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "f68506fe-ac80-41e7-87c0-c7b5ac8b5362" (UID: "f68506fe-ac80-41e7-87c0-c7b5ac8b5362"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:27:02 crc kubenswrapper[4824]: I0121 11:27:02.527201 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f68506fe-ac80-41e7-87c0-c7b5ac8b5362-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "f68506fe-ac80-41e7-87c0-c7b5ac8b5362" (UID: "f68506fe-ac80-41e7-87c0-c7b5ac8b5362"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:27:02 crc kubenswrapper[4824]: I0121 11:27:02.533661 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f68506fe-ac80-41e7-87c0-c7b5ac8b5362-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "f68506fe-ac80-41e7-87c0-c7b5ac8b5362" (UID: "f68506fe-ac80-41e7-87c0-c7b5ac8b5362"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:27:02 crc kubenswrapper[4824]: I0121 11:27:02.534984 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f68506fe-ac80-41e7-87c0-c7b5ac8b5362-config" (OuterVolumeSpecName: "config") pod "f68506fe-ac80-41e7-87c0-c7b5ac8b5362" (UID: "f68506fe-ac80-41e7-87c0-c7b5ac8b5362"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:27:02 crc kubenswrapper[4824]: I0121 11:27:02.602385 4824 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f68506fe-ac80-41e7-87c0-c7b5ac8b5362-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:02 crc kubenswrapper[4824]: I0121 11:27:02.602412 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f68506fe-ac80-41e7-87c0-c7b5ac8b5362-config\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:02 crc kubenswrapper[4824]: I0121 11:27:02.602420 4824 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f68506fe-ac80-41e7-87c0-c7b5ac8b5362-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:02 crc kubenswrapper[4824]: I0121 11:27:02.602431 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wgb7x\" (UniqueName: \"kubernetes.io/projected/f68506fe-ac80-41e7-87c0-c7b5ac8b5362-kube-api-access-wgb7x\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:02 crc kubenswrapper[4824]: I0121 11:27:02.602439 4824 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f68506fe-ac80-41e7-87c0-c7b5ac8b5362-dns-svc\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:02 crc kubenswrapper[4824]: I0121 11:27:02.602447 4824 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f68506fe-ac80-41e7-87c0-c7b5ac8b5362-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:02 crc kubenswrapper[4824]: I0121 11:27:02.706211 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f9j78\" (UniqueName: \"kubernetes.io/projected/1d8c8cff-4340-4eb4-94b7-062968a14848-kube-api-access-f9j78\") pod \"1d8c8cff-4340-4eb4-94b7-062968a14848\" (UID: \"1d8c8cff-4340-4eb4-94b7-062968a14848\") " Jan 21 11:27:02 crc kubenswrapper[4824]: I0121 11:27:02.706317 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d8c8cff-4340-4eb4-94b7-062968a14848-utilities\") pod \"1d8c8cff-4340-4eb4-94b7-062968a14848\" (UID: \"1d8c8cff-4340-4eb4-94b7-062968a14848\") " Jan 21 11:27:02 crc kubenswrapper[4824]: I0121 11:27:02.706431 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d8c8cff-4340-4eb4-94b7-062968a14848-catalog-content\") pod \"1d8c8cff-4340-4eb4-94b7-062968a14848\" (UID: \"1d8c8cff-4340-4eb4-94b7-062968a14848\") " Jan 21 11:27:02 crc kubenswrapper[4824]: I0121 11:27:02.708466 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d8c8cff-4340-4eb4-94b7-062968a14848-utilities" (OuterVolumeSpecName: "utilities") pod "1d8c8cff-4340-4eb4-94b7-062968a14848" (UID: "1d8c8cff-4340-4eb4-94b7-062968a14848"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:27:02 crc kubenswrapper[4824]: I0121 11:27:02.726788 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d8c8cff-4340-4eb4-94b7-062968a14848-kube-api-access-f9j78" (OuterVolumeSpecName: "kube-api-access-f9j78") pod "1d8c8cff-4340-4eb4-94b7-062968a14848" (UID: "1d8c8cff-4340-4eb4-94b7-062968a14848"). InnerVolumeSpecName "kube-api-access-f9j78". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:27:02 crc kubenswrapper[4824]: I0121 11:27:02.757129 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d8c8cff-4340-4eb4-94b7-062968a14848-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d8c8cff-4340-4eb4-94b7-062968a14848" (UID: "1d8c8cff-4340-4eb4-94b7-062968a14848"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:27:02 crc kubenswrapper[4824]: I0121 11:27:02.807896 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f9j78\" (UniqueName: \"kubernetes.io/projected/1d8c8cff-4340-4eb4-94b7-062968a14848-kube-api-access-f9j78\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:02 crc kubenswrapper[4824]: I0121 11:27:02.807921 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d8c8cff-4340-4eb4-94b7-062968a14848-utilities\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:02 crc kubenswrapper[4824]: I0121 11:27:02.807930 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d8c8cff-4340-4eb4-94b7-062968a14848-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:03 crc kubenswrapper[4824]: I0121 11:27:03.051632 4824 generic.go:334] "Generic (PLEG): container finished" podID="1d8c8cff-4340-4eb4-94b7-062968a14848" containerID="42cfb3a03e211fbd5fa55f6c859e8ec840fddea6ec887daa75d38ab58da624f4" exitCode=0 Jan 21 11:27:03 crc kubenswrapper[4824]: I0121 11:27:03.051891 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dn2np" event={"ID":"1d8c8cff-4340-4eb4-94b7-062968a14848","Type":"ContainerDied","Data":"42cfb3a03e211fbd5fa55f6c859e8ec840fddea6ec887daa75d38ab58da624f4"} Jan 21 11:27:03 crc kubenswrapper[4824]: I0121 11:27:03.051917 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dn2np" event={"ID":"1d8c8cff-4340-4eb4-94b7-062968a14848","Type":"ContainerDied","Data":"2dec804f6fee46d1464f08de4ca7da8580e7f171b751ac42a7b92ec863162d9b"} Jan 21 11:27:03 crc kubenswrapper[4824]: I0121 11:27:03.051933 4824 scope.go:117] "RemoveContainer" containerID="42cfb3a03e211fbd5fa55f6c859e8ec840fddea6ec887daa75d38ab58da624f4" Jan 21 11:27:03 crc kubenswrapper[4824]: I0121 11:27:03.052064 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dn2np" Jan 21 11:27:03 crc kubenswrapper[4824]: I0121 11:27:03.059406 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b895b5785-g4p8w" event={"ID":"f68506fe-ac80-41e7-87c0-c7b5ac8b5362","Type":"ContainerDied","Data":"9f0d2bd9547e5b9af28f1727793676788462d4c9d941a44af9285262e1676154"} Jan 21 11:27:03 crc kubenswrapper[4824]: I0121 11:27:03.059488 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b895b5785-g4p8w" Jan 21 11:27:03 crc kubenswrapper[4824]: I0121 11:27:03.062574 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6b9b444cf4-q85qq" event={"ID":"4ed8e1b1-57c3-4b01-bbd7-9dc5125f4f94","Type":"ContainerStarted","Data":"820df2cc441bc3b9b3b9761588c4025b1f941d3a17f1376789c1aa427ba78ecd"} Jan 21 11:27:03 crc kubenswrapper[4824]: I0121 11:27:03.063309 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-6b9b444cf4-q85qq" Jan 21 11:27:03 crc kubenswrapper[4824]: I0121 11:27:03.063337 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-6b9b444cf4-q85qq" Jan 21 11:27:03 crc kubenswrapper[4824]: I0121 11:27:03.066051 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bae23da5-948a-4883-b88d-d6ed5d6ccba7","Type":"ContainerStarted","Data":"1568d27a18b9471d6e4be5cfb1adad4950254a106a787756913dd9ddb73cf5f9"} Jan 21 11:27:03 crc kubenswrapper[4824]: I0121 11:27:03.067488 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"a5721831-a6aa-4d42-b62a-041dbe5c23ad","Type":"ContainerStarted","Data":"945d74317b4fe36380b7b590e48fb68b917191cdb410f4c0e01be7c34a318d74"} Jan 21 11:27:03 crc kubenswrapper[4824]: I0121 11:27:03.069261 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"62025fca-0943-48d2-b139-c55b0c7df358","Type":"ContainerStarted","Data":"f742f2e3d618281030fb88995a308c64ce6b08d4c15341a260e0188485483434"} Jan 21 11:27:03 crc kubenswrapper[4824]: I0121 11:27:03.069332 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="62025fca-0943-48d2-b139-c55b0c7df358" containerName="cinder-api-log" containerID="cri-o://0f1a05d13f073268720b76257e1f39bfa980627e6f98480d79c357a688a740d9" gracePeriod=30 Jan 21 11:27:03 crc kubenswrapper[4824]: I0121 11:27:03.069391 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Jan 21 11:27:03 crc kubenswrapper[4824]: I0121 11:27:03.069402 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="62025fca-0943-48d2-b139-c55b0c7df358" containerName="cinder-api" containerID="cri-o://f742f2e3d618281030fb88995a308c64ce6b08d4c15341a260e0188485483434" gracePeriod=30 Jan 21 11:27:03 crc kubenswrapper[4824]: I0121 11:27:03.079221 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-q2xgx" event={"ID":"661c81f6-6854-416e-aea7-1344a635f2db","Type":"ContainerStarted","Data":"4b21f7d92e0f61e8df8d11e4819f3e436ec4a36c57f8769d50008872b4efc836"} Jan 21 11:27:03 crc kubenswrapper[4824]: I0121 11:27:03.079560 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5c9776ccc5-q2xgx" Jan 21 11:27:03 crc kubenswrapper[4824]: I0121 11:27:03.082171 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-6b9b444cf4-q85qq" podStartSLOduration=3.082160537 podStartE2EDuration="3.082160537s" podCreationTimestamp="2026-01-21 11:27:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:27:03.079356107 +0000 UTC m=+965.372385399" watchObservedRunningTime="2026-01-21 11:27:03.082160537 +0000 UTC m=+965.375189829" Jan 21 11:27:03 crc kubenswrapper[4824]: I0121 11:27:03.096792 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=3.096779301 podStartE2EDuration="3.096779301s" podCreationTimestamp="2026-01-21 11:27:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:27:03.090422289 +0000 UTC m=+965.383451582" watchObservedRunningTime="2026-01-21 11:27:03.096779301 +0000 UTC m=+965.389808593" Jan 21 11:27:03 crc kubenswrapper[4824]: I0121 11:27:03.118107 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5c9776ccc5-q2xgx" podStartSLOduration=3.118089778 podStartE2EDuration="3.118089778s" podCreationTimestamp="2026-01-21 11:27:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:27:03.114475089 +0000 UTC m=+965.407504382" watchObservedRunningTime="2026-01-21 11:27:03.118089778 +0000 UTC m=+965.411119069" Jan 21 11:27:03 crc kubenswrapper[4824]: I0121 11:27:03.153787 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-dn2np"] Jan 21 11:27:03 crc kubenswrapper[4824]: I0121 11:27:03.161812 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-dn2np"] Jan 21 11:27:03 crc kubenswrapper[4824]: I0121 11:27:03.194413 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-b895b5785-g4p8w"] Jan 21 11:27:03 crc kubenswrapper[4824]: I0121 11:27:03.199856 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-b895b5785-g4p8w"] Jan 21 11:27:03 crc kubenswrapper[4824]: I0121 11:27:03.245712 4824 scope.go:117] "RemoveContainer" containerID="d4406aff85ae4b7edaab2e90513b6f7cdb4c88bf1bcc43b39ae793abac0eb943" Jan 21 11:27:03 crc kubenswrapper[4824]: I0121 11:27:03.493194 4824 scope.go:117] "RemoveContainer" containerID="d64282b8c833bf7a0f6a7b257f155ac96972db2ea6ccea22f61d3f3c228fc02a" Jan 21 11:27:03 crc kubenswrapper[4824]: I0121 11:27:03.534163 4824 scope.go:117] "RemoveContainer" containerID="42cfb3a03e211fbd5fa55f6c859e8ec840fddea6ec887daa75d38ab58da624f4" Jan 21 11:27:03 crc kubenswrapper[4824]: E0121 11:27:03.538785 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"42cfb3a03e211fbd5fa55f6c859e8ec840fddea6ec887daa75d38ab58da624f4\": container with ID starting with 42cfb3a03e211fbd5fa55f6c859e8ec840fddea6ec887daa75d38ab58da624f4 not found: ID does not exist" containerID="42cfb3a03e211fbd5fa55f6c859e8ec840fddea6ec887daa75d38ab58da624f4" Jan 21 11:27:03 crc kubenswrapper[4824]: I0121 11:27:03.538819 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"42cfb3a03e211fbd5fa55f6c859e8ec840fddea6ec887daa75d38ab58da624f4"} err="failed to get container status \"42cfb3a03e211fbd5fa55f6c859e8ec840fddea6ec887daa75d38ab58da624f4\": rpc error: code = NotFound desc = could not find container \"42cfb3a03e211fbd5fa55f6c859e8ec840fddea6ec887daa75d38ab58da624f4\": container with ID starting with 42cfb3a03e211fbd5fa55f6c859e8ec840fddea6ec887daa75d38ab58da624f4 not found: ID does not exist" Jan 21 11:27:03 crc kubenswrapper[4824]: I0121 11:27:03.538842 4824 scope.go:117] "RemoveContainer" containerID="d4406aff85ae4b7edaab2e90513b6f7cdb4c88bf1bcc43b39ae793abac0eb943" Jan 21 11:27:03 crc kubenswrapper[4824]: E0121 11:27:03.539192 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d4406aff85ae4b7edaab2e90513b6f7cdb4c88bf1bcc43b39ae793abac0eb943\": container with ID starting with d4406aff85ae4b7edaab2e90513b6f7cdb4c88bf1bcc43b39ae793abac0eb943 not found: ID does not exist" containerID="d4406aff85ae4b7edaab2e90513b6f7cdb4c88bf1bcc43b39ae793abac0eb943" Jan 21 11:27:03 crc kubenswrapper[4824]: I0121 11:27:03.539233 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d4406aff85ae4b7edaab2e90513b6f7cdb4c88bf1bcc43b39ae793abac0eb943"} err="failed to get container status \"d4406aff85ae4b7edaab2e90513b6f7cdb4c88bf1bcc43b39ae793abac0eb943\": rpc error: code = NotFound desc = could not find container \"d4406aff85ae4b7edaab2e90513b6f7cdb4c88bf1bcc43b39ae793abac0eb943\": container with ID starting with d4406aff85ae4b7edaab2e90513b6f7cdb4c88bf1bcc43b39ae793abac0eb943 not found: ID does not exist" Jan 21 11:27:03 crc kubenswrapper[4824]: I0121 11:27:03.539261 4824 scope.go:117] "RemoveContainer" containerID="d64282b8c833bf7a0f6a7b257f155ac96972db2ea6ccea22f61d3f3c228fc02a" Jan 21 11:27:03 crc kubenswrapper[4824]: E0121 11:27:03.539520 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d64282b8c833bf7a0f6a7b257f155ac96972db2ea6ccea22f61d3f3c228fc02a\": container with ID starting with d64282b8c833bf7a0f6a7b257f155ac96972db2ea6ccea22f61d3f3c228fc02a not found: ID does not exist" containerID="d64282b8c833bf7a0f6a7b257f155ac96972db2ea6ccea22f61d3f3c228fc02a" Jan 21 11:27:03 crc kubenswrapper[4824]: I0121 11:27:03.539560 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d64282b8c833bf7a0f6a7b257f155ac96972db2ea6ccea22f61d3f3c228fc02a"} err="failed to get container status \"d64282b8c833bf7a0f6a7b257f155ac96972db2ea6ccea22f61d3f3c228fc02a\": rpc error: code = NotFound desc = could not find container \"d64282b8c833bf7a0f6a7b257f155ac96972db2ea6ccea22f61d3f3c228fc02a\": container with ID starting with d64282b8c833bf7a0f6a7b257f155ac96972db2ea6ccea22f61d3f3c228fc02a not found: ID does not exist" Jan 21 11:27:03 crc kubenswrapper[4824]: I0121 11:27:03.539582 4824 scope.go:117] "RemoveContainer" containerID="ce260d4bf12b843ead9c0a8144fa07e77b168be5d263f9833e8d46c101318a5d" Jan 21 11:27:03 crc kubenswrapper[4824]: I0121 11:27:03.759582 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Jan 21 11:27:03 crc kubenswrapper[4824]: I0121 11:27:03.932766 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/62025fca-0943-48d2-b139-c55b0c7df358-etc-machine-id\") pod \"62025fca-0943-48d2-b139-c55b0c7df358\" (UID: \"62025fca-0943-48d2-b139-c55b0c7df358\") " Jan 21 11:27:03 crc kubenswrapper[4824]: I0121 11:27:03.933027 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/62025fca-0943-48d2-b139-c55b0c7df358-config-data-custom\") pod \"62025fca-0943-48d2-b139-c55b0c7df358\" (UID: \"62025fca-0943-48d2-b139-c55b0c7df358\") " Jan 21 11:27:03 crc kubenswrapper[4824]: I0121 11:27:03.933828 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62025fca-0943-48d2-b139-c55b0c7df358-config-data\") pod \"62025fca-0943-48d2-b139-c55b0c7df358\" (UID: \"62025fca-0943-48d2-b139-c55b0c7df358\") " Jan 21 11:27:03 crc kubenswrapper[4824]: I0121 11:27:03.933142 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/62025fca-0943-48d2-b139-c55b0c7df358-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "62025fca-0943-48d2-b139-c55b0c7df358" (UID: "62025fca-0943-48d2-b139-c55b0c7df358"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 21 11:27:03 crc kubenswrapper[4824]: I0121 11:27:03.934053 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62025fca-0943-48d2-b139-c55b0c7df358-combined-ca-bundle\") pod \"62025fca-0943-48d2-b139-c55b0c7df358\" (UID: \"62025fca-0943-48d2-b139-c55b0c7df358\") " Jan 21 11:27:03 crc kubenswrapper[4824]: I0121 11:27:03.934359 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/62025fca-0943-48d2-b139-c55b0c7df358-logs\") pod \"62025fca-0943-48d2-b139-c55b0c7df358\" (UID: \"62025fca-0943-48d2-b139-c55b0c7df358\") " Jan 21 11:27:03 crc kubenswrapper[4824]: I0121 11:27:03.934443 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/62025fca-0943-48d2-b139-c55b0c7df358-scripts\") pod \"62025fca-0943-48d2-b139-c55b0c7df358\" (UID: \"62025fca-0943-48d2-b139-c55b0c7df358\") " Jan 21 11:27:03 crc kubenswrapper[4824]: I0121 11:27:03.934576 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s2qhr\" (UniqueName: \"kubernetes.io/projected/62025fca-0943-48d2-b139-c55b0c7df358-kube-api-access-s2qhr\") pod \"62025fca-0943-48d2-b139-c55b0c7df358\" (UID: \"62025fca-0943-48d2-b139-c55b0c7df358\") " Jan 21 11:27:03 crc kubenswrapper[4824]: I0121 11:27:03.935205 4824 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/62025fca-0943-48d2-b139-c55b0c7df358-etc-machine-id\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:03 crc kubenswrapper[4824]: I0121 11:27:03.934600 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/62025fca-0943-48d2-b139-c55b0c7df358-logs" (OuterVolumeSpecName: "logs") pod "62025fca-0943-48d2-b139-c55b0c7df358" (UID: "62025fca-0943-48d2-b139-c55b0c7df358"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:27:03 crc kubenswrapper[4824]: I0121 11:27:03.938142 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/62025fca-0943-48d2-b139-c55b0c7df358-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "62025fca-0943-48d2-b139-c55b0c7df358" (UID: "62025fca-0943-48d2-b139-c55b0c7df358"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:27:03 crc kubenswrapper[4824]: I0121 11:27:03.939496 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/62025fca-0943-48d2-b139-c55b0c7df358-scripts" (OuterVolumeSpecName: "scripts") pod "62025fca-0943-48d2-b139-c55b0c7df358" (UID: "62025fca-0943-48d2-b139-c55b0c7df358"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:27:03 crc kubenswrapper[4824]: I0121 11:27:03.940010 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/62025fca-0943-48d2-b139-c55b0c7df358-kube-api-access-s2qhr" (OuterVolumeSpecName: "kube-api-access-s2qhr") pod "62025fca-0943-48d2-b139-c55b0c7df358" (UID: "62025fca-0943-48d2-b139-c55b0c7df358"). InnerVolumeSpecName "kube-api-access-s2qhr". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:27:03 crc kubenswrapper[4824]: I0121 11:27:03.953411 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/62025fca-0943-48d2-b139-c55b0c7df358-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "62025fca-0943-48d2-b139-c55b0c7df358" (UID: "62025fca-0943-48d2-b139-c55b0c7df358"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:27:03 crc kubenswrapper[4824]: I0121 11:27:03.981780 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/62025fca-0943-48d2-b139-c55b0c7df358-config-data" (OuterVolumeSpecName: "config-data") pod "62025fca-0943-48d2-b139-c55b0c7df358" (UID: "62025fca-0943-48d2-b139-c55b0c7df358"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:27:04 crc kubenswrapper[4824]: I0121 11:27:04.037058 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62025fca-0943-48d2-b139-c55b0c7df358-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:04 crc kubenswrapper[4824]: I0121 11:27:04.037087 4824 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/62025fca-0943-48d2-b139-c55b0c7df358-logs\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:04 crc kubenswrapper[4824]: I0121 11:27:04.037096 4824 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/62025fca-0943-48d2-b139-c55b0c7df358-scripts\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:04 crc kubenswrapper[4824]: I0121 11:27:04.037104 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s2qhr\" (UniqueName: \"kubernetes.io/projected/62025fca-0943-48d2-b139-c55b0c7df358-kube-api-access-s2qhr\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:04 crc kubenswrapper[4824]: I0121 11:27:04.037114 4824 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/62025fca-0943-48d2-b139-c55b0c7df358-config-data-custom\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:04 crc kubenswrapper[4824]: I0121 11:27:04.037121 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62025fca-0943-48d2-b139-c55b0c7df358-config-data\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:04 crc kubenswrapper[4824]: I0121 11:27:04.056470 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d8c8cff-4340-4eb4-94b7-062968a14848" path="/var/lib/kubelet/pods/1d8c8cff-4340-4eb4-94b7-062968a14848/volumes" Jan 21 11:27:04 crc kubenswrapper[4824]: I0121 11:27:04.057233 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f68506fe-ac80-41e7-87c0-c7b5ac8b5362" path="/var/lib/kubelet/pods/f68506fe-ac80-41e7-87c0-c7b5ac8b5362/volumes" Jan 21 11:27:04 crc kubenswrapper[4824]: I0121 11:27:04.086435 4824 generic.go:334] "Generic (PLEG): container finished" podID="62025fca-0943-48d2-b139-c55b0c7df358" containerID="f742f2e3d618281030fb88995a308c64ce6b08d4c15341a260e0188485483434" exitCode=0 Jan 21 11:27:04 crc kubenswrapper[4824]: I0121 11:27:04.086509 4824 generic.go:334] "Generic (PLEG): container finished" podID="62025fca-0943-48d2-b139-c55b0c7df358" containerID="0f1a05d13f073268720b76257e1f39bfa980627e6f98480d79c357a688a740d9" exitCode=143 Jan 21 11:27:04 crc kubenswrapper[4824]: I0121 11:27:04.086590 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"62025fca-0943-48d2-b139-c55b0c7df358","Type":"ContainerDied","Data":"f742f2e3d618281030fb88995a308c64ce6b08d4c15341a260e0188485483434"} Jan 21 11:27:04 crc kubenswrapper[4824]: I0121 11:27:04.086666 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"62025fca-0943-48d2-b139-c55b0c7df358","Type":"ContainerDied","Data":"0f1a05d13f073268720b76257e1f39bfa980627e6f98480d79c357a688a740d9"} Jan 21 11:27:04 crc kubenswrapper[4824]: I0121 11:27:04.086733 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"62025fca-0943-48d2-b139-c55b0c7df358","Type":"ContainerDied","Data":"54ae046171ebb1a5c5a1082e7114ac886653fcd28b5ecbb331be434df22b2e10"} Jan 21 11:27:04 crc kubenswrapper[4824]: I0121 11:27:04.086799 4824 scope.go:117] "RemoveContainer" containerID="f742f2e3d618281030fb88995a308c64ce6b08d4c15341a260e0188485483434" Jan 21 11:27:04 crc kubenswrapper[4824]: I0121 11:27:04.086949 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Jan 21 11:27:04 crc kubenswrapper[4824]: I0121 11:27:04.100359 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bae23da5-948a-4883-b88d-d6ed5d6ccba7","Type":"ContainerStarted","Data":"8eb8e5c1518e1715f46644f48ab6d186888980058e318f472bef1f90eb8e3c37"} Jan 21 11:27:04 crc kubenswrapper[4824]: I0121 11:27:04.100588 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Jan 21 11:27:04 crc kubenswrapper[4824]: I0121 11:27:04.103838 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"a5721831-a6aa-4d42-b62a-041dbe5c23ad","Type":"ContainerStarted","Data":"b1d6ea0f0147f78d0d5492b8698d35f699b27a92f4975bcb0d9a7a8cc6032f2a"} Jan 21 11:27:04 crc kubenswrapper[4824]: I0121 11:27:04.109373 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-7cf777545f-f8drb" event={"ID":"283e6982-7444-48c7-b7ac-ed4210329a08","Type":"ContainerStarted","Data":"0140e48b86398e0eb1d1abbcdeb37e528fd4aaa0df2804bbddad333abe356436"} Jan 21 11:27:04 crc kubenswrapper[4824]: I0121 11:27:04.109478 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-7cf777545f-f8drb" event={"ID":"283e6982-7444-48c7-b7ac-ed4210329a08","Type":"ContainerStarted","Data":"62f66de215fca8a12184ba58b7db09842a6daa1ccadd8e7b306a2bf8dcedded7"} Jan 21 11:27:04 crc kubenswrapper[4824]: I0121 11:27:04.121772 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.285811915 podStartE2EDuration="5.121757253s" podCreationTimestamp="2026-01-21 11:26:59 +0000 UTC" firstStartedPulling="2026-01-21 11:26:59.753143791 +0000 UTC m=+962.046173083" lastFinishedPulling="2026-01-21 11:27:03.589089129 +0000 UTC m=+965.882118421" observedRunningTime="2026-01-21 11:27:04.114808446 +0000 UTC m=+966.407837738" watchObservedRunningTime="2026-01-21 11:27:04.121757253 +0000 UTC m=+966.414786545" Jan 21 11:27:04 crc kubenswrapper[4824]: I0121 11:27:04.127556 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-79b7cfb8cd-ktpp5" event={"ID":"8b84edbe-4485-4027-bfc4-cebbb1b10543","Type":"ContainerStarted","Data":"ab256689d77ac6ba349d939c9dfa6b029fe4bea2c7392364417c8b51ecfca90f"} Jan 21 11:27:04 crc kubenswrapper[4824]: I0121 11:27:04.127665 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-79b7cfb8cd-ktpp5" event={"ID":"8b84edbe-4485-4027-bfc4-cebbb1b10543","Type":"ContainerStarted","Data":"450ab9f6c9cbe4618fe598d7905f883020025f5d04727d40f1f9d48dedbdd2ae"} Jan 21 11:27:04 crc kubenswrapper[4824]: I0121 11:27:04.132550 4824 scope.go:117] "RemoveContainer" containerID="0f1a05d13f073268720b76257e1f39bfa980627e6f98480d79c357a688a740d9" Jan 21 11:27:04 crc kubenswrapper[4824]: I0121 11:27:04.147593 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.462352276 podStartE2EDuration="4.14756866s" podCreationTimestamp="2026-01-21 11:27:00 +0000 UTC" firstStartedPulling="2026-01-21 11:27:01.308068844 +0000 UTC m=+963.601098137" lastFinishedPulling="2026-01-21 11:27:01.993285229 +0000 UTC m=+964.286314521" observedRunningTime="2026-01-21 11:27:04.132864544 +0000 UTC m=+966.425893836" watchObservedRunningTime="2026-01-21 11:27:04.14756866 +0000 UTC m=+966.440597952" Jan 21 11:27:04 crc kubenswrapper[4824]: I0121 11:27:04.157870 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Jan 21 11:27:04 crc kubenswrapper[4824]: I0121 11:27:04.160523 4824 scope.go:117] "RemoveContainer" containerID="f742f2e3d618281030fb88995a308c64ce6b08d4c15341a260e0188485483434" Jan 21 11:27:04 crc kubenswrapper[4824]: E0121 11:27:04.165454 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f742f2e3d618281030fb88995a308c64ce6b08d4c15341a260e0188485483434\": container with ID starting with f742f2e3d618281030fb88995a308c64ce6b08d4c15341a260e0188485483434 not found: ID does not exist" containerID="f742f2e3d618281030fb88995a308c64ce6b08d4c15341a260e0188485483434" Jan 21 11:27:04 crc kubenswrapper[4824]: I0121 11:27:04.165491 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f742f2e3d618281030fb88995a308c64ce6b08d4c15341a260e0188485483434"} err="failed to get container status \"f742f2e3d618281030fb88995a308c64ce6b08d4c15341a260e0188485483434\": rpc error: code = NotFound desc = could not find container \"f742f2e3d618281030fb88995a308c64ce6b08d4c15341a260e0188485483434\": container with ID starting with f742f2e3d618281030fb88995a308c64ce6b08d4c15341a260e0188485483434 not found: ID does not exist" Jan 21 11:27:04 crc kubenswrapper[4824]: I0121 11:27:04.165512 4824 scope.go:117] "RemoveContainer" containerID="0f1a05d13f073268720b76257e1f39bfa980627e6f98480d79c357a688a740d9" Jan 21 11:27:04 crc kubenswrapper[4824]: E0121 11:27:04.165856 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0f1a05d13f073268720b76257e1f39bfa980627e6f98480d79c357a688a740d9\": container with ID starting with 0f1a05d13f073268720b76257e1f39bfa980627e6f98480d79c357a688a740d9 not found: ID does not exist" containerID="0f1a05d13f073268720b76257e1f39bfa980627e6f98480d79c357a688a740d9" Jan 21 11:27:04 crc kubenswrapper[4824]: I0121 11:27:04.165881 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0f1a05d13f073268720b76257e1f39bfa980627e6f98480d79c357a688a740d9"} err="failed to get container status \"0f1a05d13f073268720b76257e1f39bfa980627e6f98480d79c357a688a740d9\": rpc error: code = NotFound desc = could not find container \"0f1a05d13f073268720b76257e1f39bfa980627e6f98480d79c357a688a740d9\": container with ID starting with 0f1a05d13f073268720b76257e1f39bfa980627e6f98480d79c357a688a740d9 not found: ID does not exist" Jan 21 11:27:04 crc kubenswrapper[4824]: I0121 11:27:04.165895 4824 scope.go:117] "RemoveContainer" containerID="f742f2e3d618281030fb88995a308c64ce6b08d4c15341a260e0188485483434" Jan 21 11:27:04 crc kubenswrapper[4824]: I0121 11:27:04.166155 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f742f2e3d618281030fb88995a308c64ce6b08d4c15341a260e0188485483434"} err="failed to get container status \"f742f2e3d618281030fb88995a308c64ce6b08d4c15341a260e0188485483434\": rpc error: code = NotFound desc = could not find container \"f742f2e3d618281030fb88995a308c64ce6b08d4c15341a260e0188485483434\": container with ID starting with f742f2e3d618281030fb88995a308c64ce6b08d4c15341a260e0188485483434 not found: ID does not exist" Jan 21 11:27:04 crc kubenswrapper[4824]: I0121 11:27:04.166188 4824 scope.go:117] "RemoveContainer" containerID="0f1a05d13f073268720b76257e1f39bfa980627e6f98480d79c357a688a740d9" Jan 21 11:27:04 crc kubenswrapper[4824]: I0121 11:27:04.166399 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0f1a05d13f073268720b76257e1f39bfa980627e6f98480d79c357a688a740d9"} err="failed to get container status \"0f1a05d13f073268720b76257e1f39bfa980627e6f98480d79c357a688a740d9\": rpc error: code = NotFound desc = could not find container \"0f1a05d13f073268720b76257e1f39bfa980627e6f98480d79c357a688a740d9\": container with ID starting with 0f1a05d13f073268720b76257e1f39bfa980627e6f98480d79c357a688a740d9 not found: ID does not exist" Jan 21 11:27:04 crc kubenswrapper[4824]: I0121 11:27:04.167381 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Jan 21 11:27:04 crc kubenswrapper[4824]: I0121 11:27:04.168210 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-7cf777545f-f8drb" podStartSLOduration=2.03794652 podStartE2EDuration="4.168195486s" podCreationTimestamp="2026-01-21 11:27:00 +0000 UTC" firstStartedPulling="2026-01-21 11:27:01.364393736 +0000 UTC m=+963.657423028" lastFinishedPulling="2026-01-21 11:27:03.494642702 +0000 UTC m=+965.787671994" observedRunningTime="2026-01-21 11:27:04.157287521 +0000 UTC m=+966.450316814" watchObservedRunningTime="2026-01-21 11:27:04.168195486 +0000 UTC m=+966.461224779" Jan 21 11:27:04 crc kubenswrapper[4824]: I0121 11:27:04.179506 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Jan 21 11:27:04 crc kubenswrapper[4824]: E0121 11:27:04.179851 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62025fca-0943-48d2-b139-c55b0c7df358" containerName="cinder-api-log" Jan 21 11:27:04 crc kubenswrapper[4824]: I0121 11:27:04.179864 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="62025fca-0943-48d2-b139-c55b0c7df358" containerName="cinder-api-log" Jan 21 11:27:04 crc kubenswrapper[4824]: E0121 11:27:04.179875 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62025fca-0943-48d2-b139-c55b0c7df358" containerName="cinder-api" Jan 21 11:27:04 crc kubenswrapper[4824]: I0121 11:27:04.179881 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="62025fca-0943-48d2-b139-c55b0c7df358" containerName="cinder-api" Jan 21 11:27:04 crc kubenswrapper[4824]: E0121 11:27:04.179892 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f68506fe-ac80-41e7-87c0-c7b5ac8b5362" containerName="init" Jan 21 11:27:04 crc kubenswrapper[4824]: I0121 11:27:04.179897 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="f68506fe-ac80-41e7-87c0-c7b5ac8b5362" containerName="init" Jan 21 11:27:04 crc kubenswrapper[4824]: E0121 11:27:04.179908 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d8c8cff-4340-4eb4-94b7-062968a14848" containerName="extract-utilities" Jan 21 11:27:04 crc kubenswrapper[4824]: I0121 11:27:04.179914 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d8c8cff-4340-4eb4-94b7-062968a14848" containerName="extract-utilities" Jan 21 11:27:04 crc kubenswrapper[4824]: E0121 11:27:04.179923 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d8c8cff-4340-4eb4-94b7-062968a14848" containerName="registry-server" Jan 21 11:27:04 crc kubenswrapper[4824]: I0121 11:27:04.179928 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d8c8cff-4340-4eb4-94b7-062968a14848" containerName="registry-server" Jan 21 11:27:04 crc kubenswrapper[4824]: E0121 11:27:04.179938 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d8c8cff-4340-4eb4-94b7-062968a14848" containerName="extract-content" Jan 21 11:27:04 crc kubenswrapper[4824]: I0121 11:27:04.179943 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d8c8cff-4340-4eb4-94b7-062968a14848" containerName="extract-content" Jan 21 11:27:04 crc kubenswrapper[4824]: I0121 11:27:04.180178 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="62025fca-0943-48d2-b139-c55b0c7df358" containerName="cinder-api" Jan 21 11:27:04 crc kubenswrapper[4824]: I0121 11:27:04.180192 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="1d8c8cff-4340-4eb4-94b7-062968a14848" containerName="registry-server" Jan 21 11:27:04 crc kubenswrapper[4824]: I0121 11:27:04.180198 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="62025fca-0943-48d2-b139-c55b0c7df358" containerName="cinder-api-log" Jan 21 11:27:04 crc kubenswrapper[4824]: I0121 11:27:04.180218 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="f68506fe-ac80-41e7-87c0-c7b5ac8b5362" containerName="init" Jan 21 11:27:04 crc kubenswrapper[4824]: I0121 11:27:04.181020 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Jan 21 11:27:04 crc kubenswrapper[4824]: I0121 11:27:04.183184 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Jan 21 11:27:04 crc kubenswrapper[4824]: I0121 11:27:04.183331 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Jan 21 11:27:04 crc kubenswrapper[4824]: I0121 11:27:04.183520 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Jan 21 11:27:04 crc kubenswrapper[4824]: I0121 11:27:04.186784 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Jan 21 11:27:04 crc kubenswrapper[4824]: I0121 11:27:04.186917 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-79b7cfb8cd-ktpp5" podStartSLOduration=2.129948567 podStartE2EDuration="4.186902442s" podCreationTimestamp="2026-01-21 11:27:00 +0000 UTC" firstStartedPulling="2026-01-21 11:27:01.443907528 +0000 UTC m=+963.736936819" lastFinishedPulling="2026-01-21 11:27:03.500861402 +0000 UTC m=+965.793890694" observedRunningTime="2026-01-21 11:27:04.179489399 +0000 UTC m=+966.472518691" watchObservedRunningTime="2026-01-21 11:27:04.186902442 +0000 UTC m=+966.479931734" Jan 21 11:27:04 crc kubenswrapper[4824]: I0121 11:27:04.344472 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff9de537-ecd2-469d-b90e-5857bbf64c2d-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"ff9de537-ecd2-469d-b90e-5857bbf64c2d\") " pod="openstack/cinder-api-0" Jan 21 11:27:04 crc kubenswrapper[4824]: I0121 11:27:04.344516 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff9de537-ecd2-469d-b90e-5857bbf64c2d-config-data\") pod \"cinder-api-0\" (UID: \"ff9de537-ecd2-469d-b90e-5857bbf64c2d\") " pod="openstack/cinder-api-0" Jan 21 11:27:04 crc kubenswrapper[4824]: I0121 11:27:04.344548 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ff9de537-ecd2-469d-b90e-5857bbf64c2d-etc-machine-id\") pod \"cinder-api-0\" (UID: \"ff9de537-ecd2-469d-b90e-5857bbf64c2d\") " pod="openstack/cinder-api-0" Jan 21 11:27:04 crc kubenswrapper[4824]: I0121 11:27:04.344613 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ff9de537-ecd2-469d-b90e-5857bbf64c2d-logs\") pod \"cinder-api-0\" (UID: \"ff9de537-ecd2-469d-b90e-5857bbf64c2d\") " pod="openstack/cinder-api-0" Jan 21 11:27:04 crc kubenswrapper[4824]: I0121 11:27:04.344667 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ff9de537-ecd2-469d-b90e-5857bbf64c2d-scripts\") pod \"cinder-api-0\" (UID: \"ff9de537-ecd2-469d-b90e-5857bbf64c2d\") " pod="openstack/cinder-api-0" Jan 21 11:27:04 crc kubenswrapper[4824]: I0121 11:27:04.344716 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ff9de537-ecd2-469d-b90e-5857bbf64c2d-config-data-custom\") pod \"cinder-api-0\" (UID: \"ff9de537-ecd2-469d-b90e-5857bbf64c2d\") " pod="openstack/cinder-api-0" Jan 21 11:27:04 crc kubenswrapper[4824]: I0121 11:27:04.344750 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ff9de537-ecd2-469d-b90e-5857bbf64c2d-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"ff9de537-ecd2-469d-b90e-5857bbf64c2d\") " pod="openstack/cinder-api-0" Jan 21 11:27:04 crc kubenswrapper[4824]: I0121 11:27:04.344802 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-447gr\" (UniqueName: \"kubernetes.io/projected/ff9de537-ecd2-469d-b90e-5857bbf64c2d-kube-api-access-447gr\") pod \"cinder-api-0\" (UID: \"ff9de537-ecd2-469d-b90e-5857bbf64c2d\") " pod="openstack/cinder-api-0" Jan 21 11:27:04 crc kubenswrapper[4824]: I0121 11:27:04.344819 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ff9de537-ecd2-469d-b90e-5857bbf64c2d-public-tls-certs\") pod \"cinder-api-0\" (UID: \"ff9de537-ecd2-469d-b90e-5857bbf64c2d\") " pod="openstack/cinder-api-0" Jan 21 11:27:04 crc kubenswrapper[4824]: I0121 11:27:04.445764 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ff9de537-ecd2-469d-b90e-5857bbf64c2d-scripts\") pod \"cinder-api-0\" (UID: \"ff9de537-ecd2-469d-b90e-5857bbf64c2d\") " pod="openstack/cinder-api-0" Jan 21 11:27:04 crc kubenswrapper[4824]: I0121 11:27:04.445839 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ff9de537-ecd2-469d-b90e-5857bbf64c2d-config-data-custom\") pod \"cinder-api-0\" (UID: \"ff9de537-ecd2-469d-b90e-5857bbf64c2d\") " pod="openstack/cinder-api-0" Jan 21 11:27:04 crc kubenswrapper[4824]: I0121 11:27:04.445863 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ff9de537-ecd2-469d-b90e-5857bbf64c2d-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"ff9de537-ecd2-469d-b90e-5857bbf64c2d\") " pod="openstack/cinder-api-0" Jan 21 11:27:04 crc kubenswrapper[4824]: I0121 11:27:04.445919 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ff9de537-ecd2-469d-b90e-5857bbf64c2d-public-tls-certs\") pod \"cinder-api-0\" (UID: \"ff9de537-ecd2-469d-b90e-5857bbf64c2d\") " pod="openstack/cinder-api-0" Jan 21 11:27:04 crc kubenswrapper[4824]: I0121 11:27:04.445934 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-447gr\" (UniqueName: \"kubernetes.io/projected/ff9de537-ecd2-469d-b90e-5857bbf64c2d-kube-api-access-447gr\") pod \"cinder-api-0\" (UID: \"ff9de537-ecd2-469d-b90e-5857bbf64c2d\") " pod="openstack/cinder-api-0" Jan 21 11:27:04 crc kubenswrapper[4824]: I0121 11:27:04.445999 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff9de537-ecd2-469d-b90e-5857bbf64c2d-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"ff9de537-ecd2-469d-b90e-5857bbf64c2d\") " pod="openstack/cinder-api-0" Jan 21 11:27:04 crc kubenswrapper[4824]: I0121 11:27:04.446019 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff9de537-ecd2-469d-b90e-5857bbf64c2d-config-data\") pod \"cinder-api-0\" (UID: \"ff9de537-ecd2-469d-b90e-5857bbf64c2d\") " pod="openstack/cinder-api-0" Jan 21 11:27:04 crc kubenswrapper[4824]: I0121 11:27:04.446043 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ff9de537-ecd2-469d-b90e-5857bbf64c2d-etc-machine-id\") pod \"cinder-api-0\" (UID: \"ff9de537-ecd2-469d-b90e-5857bbf64c2d\") " pod="openstack/cinder-api-0" Jan 21 11:27:04 crc kubenswrapper[4824]: I0121 11:27:04.446100 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ff9de537-ecd2-469d-b90e-5857bbf64c2d-logs\") pod \"cinder-api-0\" (UID: \"ff9de537-ecd2-469d-b90e-5857bbf64c2d\") " pod="openstack/cinder-api-0" Jan 21 11:27:04 crc kubenswrapper[4824]: I0121 11:27:04.446433 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ff9de537-ecd2-469d-b90e-5857bbf64c2d-logs\") pod \"cinder-api-0\" (UID: \"ff9de537-ecd2-469d-b90e-5857bbf64c2d\") " pod="openstack/cinder-api-0" Jan 21 11:27:04 crc kubenswrapper[4824]: I0121 11:27:04.447071 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ff9de537-ecd2-469d-b90e-5857bbf64c2d-etc-machine-id\") pod \"cinder-api-0\" (UID: \"ff9de537-ecd2-469d-b90e-5857bbf64c2d\") " pod="openstack/cinder-api-0" Jan 21 11:27:04 crc kubenswrapper[4824]: I0121 11:27:04.449190 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ff9de537-ecd2-469d-b90e-5857bbf64c2d-scripts\") pod \"cinder-api-0\" (UID: \"ff9de537-ecd2-469d-b90e-5857bbf64c2d\") " pod="openstack/cinder-api-0" Jan 21 11:27:04 crc kubenswrapper[4824]: I0121 11:27:04.450147 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ff9de537-ecd2-469d-b90e-5857bbf64c2d-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"ff9de537-ecd2-469d-b90e-5857bbf64c2d\") " pod="openstack/cinder-api-0" Jan 21 11:27:04 crc kubenswrapper[4824]: I0121 11:27:04.450537 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff9de537-ecd2-469d-b90e-5857bbf64c2d-config-data\") pod \"cinder-api-0\" (UID: \"ff9de537-ecd2-469d-b90e-5857bbf64c2d\") " pod="openstack/cinder-api-0" Jan 21 11:27:04 crc kubenswrapper[4824]: I0121 11:27:04.451526 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff9de537-ecd2-469d-b90e-5857bbf64c2d-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"ff9de537-ecd2-469d-b90e-5857bbf64c2d\") " pod="openstack/cinder-api-0" Jan 21 11:27:04 crc kubenswrapper[4824]: I0121 11:27:04.452163 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ff9de537-ecd2-469d-b90e-5857bbf64c2d-config-data-custom\") pod \"cinder-api-0\" (UID: \"ff9de537-ecd2-469d-b90e-5857bbf64c2d\") " pod="openstack/cinder-api-0" Jan 21 11:27:04 crc kubenswrapper[4824]: I0121 11:27:04.453602 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ff9de537-ecd2-469d-b90e-5857bbf64c2d-public-tls-certs\") pod \"cinder-api-0\" (UID: \"ff9de537-ecd2-469d-b90e-5857bbf64c2d\") " pod="openstack/cinder-api-0" Jan 21 11:27:04 crc kubenswrapper[4824]: I0121 11:27:04.466619 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-447gr\" (UniqueName: \"kubernetes.io/projected/ff9de537-ecd2-469d-b90e-5857bbf64c2d-kube-api-access-447gr\") pod \"cinder-api-0\" (UID: \"ff9de537-ecd2-469d-b90e-5857bbf64c2d\") " pod="openstack/cinder-api-0" Jan 21 11:27:04 crc kubenswrapper[4824]: I0121 11:27:04.502350 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Jan 21 11:27:04 crc kubenswrapper[4824]: I0121 11:27:04.906387 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Jan 21 11:27:05 crc kubenswrapper[4824]: I0121 11:27:05.136082 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"ff9de537-ecd2-469d-b90e-5857bbf64c2d","Type":"ContainerStarted","Data":"67ee0780e6bb858ef355fe024566b8fb623152b9372fb6f8aad3e487374938d7"} Jan 21 11:27:05 crc kubenswrapper[4824]: I0121 11:27:05.571888 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Jan 21 11:27:06 crc kubenswrapper[4824]: I0121 11:27:06.057588 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="62025fca-0943-48d2-b139-c55b0c7df358" path="/var/lib/kubelet/pods/62025fca-0943-48d2-b139-c55b0c7df358/volumes" Jan 21 11:27:06 crc kubenswrapper[4824]: I0121 11:27:06.143904 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"ff9de537-ecd2-469d-b90e-5857bbf64c2d","Type":"ContainerStarted","Data":"4300f85a676d71a079b02f932e8d9dae11e471e7fe4d4a61effa67b37578b02c"} Jan 21 11:27:06 crc kubenswrapper[4824]: I0121 11:27:06.143940 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"ff9de537-ecd2-469d-b90e-5857bbf64c2d","Type":"ContainerStarted","Data":"f2c098bd6680501c78682d7bd8ce5e2c495edc0d917df3678d84a806e3076625"} Jan 21 11:27:06 crc kubenswrapper[4824]: I0121 11:27:06.163469 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=2.163455262 podStartE2EDuration="2.163455262s" podCreationTimestamp="2026-01-21 11:27:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:27:06.162747236 +0000 UTC m=+968.455776528" watchObservedRunningTime="2026-01-21 11:27:06.163455262 +0000 UTC m=+968.456484554" Jan 21 11:27:06 crc kubenswrapper[4824]: I0121 11:27:06.252507 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-6c7f965fb4-8slr2"] Jan 21 11:27:06 crc kubenswrapper[4824]: I0121 11:27:06.253715 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6c7f965fb4-8slr2" Jan 21 11:27:06 crc kubenswrapper[4824]: I0121 11:27:06.255600 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Jan 21 11:27:06 crc kubenswrapper[4824]: I0121 11:27:06.255902 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Jan 21 11:27:06 crc kubenswrapper[4824]: I0121 11:27:06.265626 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-6c7f965fb4-8slr2"] Jan 21 11:27:06 crc kubenswrapper[4824]: I0121 11:27:06.390002 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3f7f97b5-8abd-4351-8396-cf1b76c376a7-config-data\") pod \"barbican-api-6c7f965fb4-8slr2\" (UID: \"3f7f97b5-8abd-4351-8396-cf1b76c376a7\") " pod="openstack/barbican-api-6c7f965fb4-8slr2" Jan 21 11:27:06 crc kubenswrapper[4824]: I0121 11:27:06.390078 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3f7f97b5-8abd-4351-8396-cf1b76c376a7-public-tls-certs\") pod \"barbican-api-6c7f965fb4-8slr2\" (UID: \"3f7f97b5-8abd-4351-8396-cf1b76c376a7\") " pod="openstack/barbican-api-6c7f965fb4-8slr2" Jan 21 11:27:06 crc kubenswrapper[4824]: I0121 11:27:06.390153 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3f7f97b5-8abd-4351-8396-cf1b76c376a7-logs\") pod \"barbican-api-6c7f965fb4-8slr2\" (UID: \"3f7f97b5-8abd-4351-8396-cf1b76c376a7\") " pod="openstack/barbican-api-6c7f965fb4-8slr2" Jan 21 11:27:06 crc kubenswrapper[4824]: I0121 11:27:06.390217 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f7f97b5-8abd-4351-8396-cf1b76c376a7-combined-ca-bundle\") pod \"barbican-api-6c7f965fb4-8slr2\" (UID: \"3f7f97b5-8abd-4351-8396-cf1b76c376a7\") " pod="openstack/barbican-api-6c7f965fb4-8slr2" Jan 21 11:27:06 crc kubenswrapper[4824]: I0121 11:27:06.390239 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3f7f97b5-8abd-4351-8396-cf1b76c376a7-config-data-custom\") pod \"barbican-api-6c7f965fb4-8slr2\" (UID: \"3f7f97b5-8abd-4351-8396-cf1b76c376a7\") " pod="openstack/barbican-api-6c7f965fb4-8slr2" Jan 21 11:27:06 crc kubenswrapper[4824]: I0121 11:27:06.390285 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3f7f97b5-8abd-4351-8396-cf1b76c376a7-internal-tls-certs\") pod \"barbican-api-6c7f965fb4-8slr2\" (UID: \"3f7f97b5-8abd-4351-8396-cf1b76c376a7\") " pod="openstack/barbican-api-6c7f965fb4-8slr2" Jan 21 11:27:06 crc kubenswrapper[4824]: I0121 11:27:06.390316 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8f268\" (UniqueName: \"kubernetes.io/projected/3f7f97b5-8abd-4351-8396-cf1b76c376a7-kube-api-access-8f268\") pod \"barbican-api-6c7f965fb4-8slr2\" (UID: \"3f7f97b5-8abd-4351-8396-cf1b76c376a7\") " pod="openstack/barbican-api-6c7f965fb4-8slr2" Jan 21 11:27:06 crc kubenswrapper[4824]: I0121 11:27:06.491486 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3f7f97b5-8abd-4351-8396-cf1b76c376a7-internal-tls-certs\") pod \"barbican-api-6c7f965fb4-8slr2\" (UID: \"3f7f97b5-8abd-4351-8396-cf1b76c376a7\") " pod="openstack/barbican-api-6c7f965fb4-8slr2" Jan 21 11:27:06 crc kubenswrapper[4824]: I0121 11:27:06.491539 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8f268\" (UniqueName: \"kubernetes.io/projected/3f7f97b5-8abd-4351-8396-cf1b76c376a7-kube-api-access-8f268\") pod \"barbican-api-6c7f965fb4-8slr2\" (UID: \"3f7f97b5-8abd-4351-8396-cf1b76c376a7\") " pod="openstack/barbican-api-6c7f965fb4-8slr2" Jan 21 11:27:06 crc kubenswrapper[4824]: I0121 11:27:06.491631 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3f7f97b5-8abd-4351-8396-cf1b76c376a7-config-data\") pod \"barbican-api-6c7f965fb4-8slr2\" (UID: \"3f7f97b5-8abd-4351-8396-cf1b76c376a7\") " pod="openstack/barbican-api-6c7f965fb4-8slr2" Jan 21 11:27:06 crc kubenswrapper[4824]: I0121 11:27:06.491673 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3f7f97b5-8abd-4351-8396-cf1b76c376a7-public-tls-certs\") pod \"barbican-api-6c7f965fb4-8slr2\" (UID: \"3f7f97b5-8abd-4351-8396-cf1b76c376a7\") " pod="openstack/barbican-api-6c7f965fb4-8slr2" Jan 21 11:27:06 crc kubenswrapper[4824]: I0121 11:27:06.491744 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3f7f97b5-8abd-4351-8396-cf1b76c376a7-logs\") pod \"barbican-api-6c7f965fb4-8slr2\" (UID: \"3f7f97b5-8abd-4351-8396-cf1b76c376a7\") " pod="openstack/barbican-api-6c7f965fb4-8slr2" Jan 21 11:27:06 crc kubenswrapper[4824]: I0121 11:27:06.491805 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f7f97b5-8abd-4351-8396-cf1b76c376a7-combined-ca-bundle\") pod \"barbican-api-6c7f965fb4-8slr2\" (UID: \"3f7f97b5-8abd-4351-8396-cf1b76c376a7\") " pod="openstack/barbican-api-6c7f965fb4-8slr2" Jan 21 11:27:06 crc kubenswrapper[4824]: I0121 11:27:06.491837 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3f7f97b5-8abd-4351-8396-cf1b76c376a7-config-data-custom\") pod \"barbican-api-6c7f965fb4-8slr2\" (UID: \"3f7f97b5-8abd-4351-8396-cf1b76c376a7\") " pod="openstack/barbican-api-6c7f965fb4-8slr2" Jan 21 11:27:06 crc kubenswrapper[4824]: I0121 11:27:06.493094 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3f7f97b5-8abd-4351-8396-cf1b76c376a7-logs\") pod \"barbican-api-6c7f965fb4-8slr2\" (UID: \"3f7f97b5-8abd-4351-8396-cf1b76c376a7\") " pod="openstack/barbican-api-6c7f965fb4-8slr2" Jan 21 11:27:06 crc kubenswrapper[4824]: I0121 11:27:06.496395 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3f7f97b5-8abd-4351-8396-cf1b76c376a7-public-tls-certs\") pod \"barbican-api-6c7f965fb4-8slr2\" (UID: \"3f7f97b5-8abd-4351-8396-cf1b76c376a7\") " pod="openstack/barbican-api-6c7f965fb4-8slr2" Jan 21 11:27:06 crc kubenswrapper[4824]: I0121 11:27:06.496507 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3f7f97b5-8abd-4351-8396-cf1b76c376a7-config-data-custom\") pod \"barbican-api-6c7f965fb4-8slr2\" (UID: \"3f7f97b5-8abd-4351-8396-cf1b76c376a7\") " pod="openstack/barbican-api-6c7f965fb4-8slr2" Jan 21 11:27:06 crc kubenswrapper[4824]: I0121 11:27:06.496847 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f7f97b5-8abd-4351-8396-cf1b76c376a7-combined-ca-bundle\") pod \"barbican-api-6c7f965fb4-8slr2\" (UID: \"3f7f97b5-8abd-4351-8396-cf1b76c376a7\") " pod="openstack/barbican-api-6c7f965fb4-8slr2" Jan 21 11:27:06 crc kubenswrapper[4824]: I0121 11:27:06.497833 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3f7f97b5-8abd-4351-8396-cf1b76c376a7-config-data\") pod \"barbican-api-6c7f965fb4-8slr2\" (UID: \"3f7f97b5-8abd-4351-8396-cf1b76c376a7\") " pod="openstack/barbican-api-6c7f965fb4-8slr2" Jan 21 11:27:06 crc kubenswrapper[4824]: I0121 11:27:06.498254 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3f7f97b5-8abd-4351-8396-cf1b76c376a7-internal-tls-certs\") pod \"barbican-api-6c7f965fb4-8slr2\" (UID: \"3f7f97b5-8abd-4351-8396-cf1b76c376a7\") " pod="openstack/barbican-api-6c7f965fb4-8slr2" Jan 21 11:27:06 crc kubenswrapper[4824]: I0121 11:27:06.507535 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8f268\" (UniqueName: \"kubernetes.io/projected/3f7f97b5-8abd-4351-8396-cf1b76c376a7-kube-api-access-8f268\") pod \"barbican-api-6c7f965fb4-8slr2\" (UID: \"3f7f97b5-8abd-4351-8396-cf1b76c376a7\") " pod="openstack/barbican-api-6c7f965fb4-8slr2" Jan 21 11:27:06 crc kubenswrapper[4824]: I0121 11:27:06.578303 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6c7f965fb4-8slr2" Jan 21 11:27:06 crc kubenswrapper[4824]: I0121 11:27:06.966534 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-6c7f965fb4-8slr2"] Jan 21 11:27:06 crc kubenswrapper[4824]: W0121 11:27:06.969921 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3f7f97b5_8abd_4351_8396_cf1b76c376a7.slice/crio-62d120dae566220242a7fc924ec317d9e8ba45834de3906e26ea5a79327dfa28 WatchSource:0}: Error finding container 62d120dae566220242a7fc924ec317d9e8ba45834de3906e26ea5a79327dfa28: Status 404 returned error can't find the container with id 62d120dae566220242a7fc924ec317d9e8ba45834de3906e26ea5a79327dfa28 Jan 21 11:27:07 crc kubenswrapper[4824]: I0121 11:27:07.152237 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6c7f965fb4-8slr2" event={"ID":"3f7f97b5-8abd-4351-8396-cf1b76c376a7","Type":"ContainerStarted","Data":"e7a548bf434fd3b356d458e6223ea4dc751528f6ea6cf4d9b657539003cfd16e"} Jan 21 11:27:07 crc kubenswrapper[4824]: I0121 11:27:07.152277 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6c7f965fb4-8slr2" event={"ID":"3f7f97b5-8abd-4351-8396-cf1b76c376a7","Type":"ContainerStarted","Data":"62d120dae566220242a7fc924ec317d9e8ba45834de3906e26ea5a79327dfa28"} Jan 21 11:27:07 crc kubenswrapper[4824]: I0121 11:27:07.152376 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Jan 21 11:27:07 crc kubenswrapper[4824]: I0121 11:27:07.379505 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-6b9b444cf4-q85qq" Jan 21 11:27:08 crc kubenswrapper[4824]: I0121 11:27:08.161045 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6c7f965fb4-8slr2" event={"ID":"3f7f97b5-8abd-4351-8396-cf1b76c376a7","Type":"ContainerStarted","Data":"2c342342909bc38f0c94dba0cbbc1e6ee1d8016636b9f1985115fed0b825a876"} Jan 21 11:27:08 crc kubenswrapper[4824]: I0121 11:27:08.161442 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-6c7f965fb4-8slr2" Jan 21 11:27:08 crc kubenswrapper[4824]: I0121 11:27:08.161473 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-6c7f965fb4-8slr2" Jan 21 11:27:08 crc kubenswrapper[4824]: I0121 11:27:08.184662 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-6c7f965fb4-8slr2" podStartSLOduration=2.184646841 podStartE2EDuration="2.184646841s" podCreationTimestamp="2026-01-21 11:27:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:27:08.179527715 +0000 UTC m=+970.472557007" watchObservedRunningTime="2026-01-21 11:27:08.184646841 +0000 UTC m=+970.477676133" Jan 21 11:27:08 crc kubenswrapper[4824]: I0121 11:27:08.670468 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-6b9b444cf4-q85qq" Jan 21 11:27:10 crc kubenswrapper[4824]: I0121 11:27:10.749337 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Jan 21 11:27:10 crc kubenswrapper[4824]: I0121 11:27:10.789357 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Jan 21 11:27:11 crc kubenswrapper[4824]: I0121 11:27:11.009108 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5c9776ccc5-q2xgx" Jan 21 11:27:11 crc kubenswrapper[4824]: I0121 11:27:11.051704 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-vbmkp"] Jan 21 11:27:11 crc kubenswrapper[4824]: I0121 11:27:11.051884 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-55f844cf75-vbmkp" podUID="bb4ea3d2-63c7-4829-b6b8-8112476493fa" containerName="dnsmasq-dns" containerID="cri-o://25b08d9ba24d1cbae3803365bca42ff0e044f0f812800389a684ef8eb1362dbe" gracePeriod=10 Jan 21 11:27:11 crc kubenswrapper[4824]: I0121 11:27:11.194495 4824 generic.go:334] "Generic (PLEG): container finished" podID="bb4ea3d2-63c7-4829-b6b8-8112476493fa" containerID="25b08d9ba24d1cbae3803365bca42ff0e044f0f812800389a684ef8eb1362dbe" exitCode=0 Jan 21 11:27:11 crc kubenswrapper[4824]: I0121 11:27:11.194661 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="a5721831-a6aa-4d42-b62a-041dbe5c23ad" containerName="cinder-scheduler" containerID="cri-o://945d74317b4fe36380b7b590e48fb68b917191cdb410f4c0e01be7c34a318d74" gracePeriod=30 Jan 21 11:27:11 crc kubenswrapper[4824]: I0121 11:27:11.194889 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-vbmkp" event={"ID":"bb4ea3d2-63c7-4829-b6b8-8112476493fa","Type":"ContainerDied","Data":"25b08d9ba24d1cbae3803365bca42ff0e044f0f812800389a684ef8eb1362dbe"} Jan 21 11:27:11 crc kubenswrapper[4824]: I0121 11:27:11.195119 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="a5721831-a6aa-4d42-b62a-041dbe5c23ad" containerName="probe" containerID="cri-o://b1d6ea0f0147f78d0d5492b8698d35f699b27a92f4975bcb0d9a7a8cc6032f2a" gracePeriod=30 Jan 21 11:27:11 crc kubenswrapper[4824]: I0121 11:27:11.600152 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55f844cf75-vbmkp" Jan 21 11:27:11 crc kubenswrapper[4824]: I0121 11:27:11.678238 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bb4ea3d2-63c7-4829-b6b8-8112476493fa-ovsdbserver-sb\") pod \"bb4ea3d2-63c7-4829-b6b8-8112476493fa\" (UID: \"bb4ea3d2-63c7-4829-b6b8-8112476493fa\") " Jan 21 11:27:11 crc kubenswrapper[4824]: I0121 11:27:11.678281 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-clj2h\" (UniqueName: \"kubernetes.io/projected/bb4ea3d2-63c7-4829-b6b8-8112476493fa-kube-api-access-clj2h\") pod \"bb4ea3d2-63c7-4829-b6b8-8112476493fa\" (UID: \"bb4ea3d2-63c7-4829-b6b8-8112476493fa\") " Jan 21 11:27:11 crc kubenswrapper[4824]: I0121 11:27:11.678306 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bb4ea3d2-63c7-4829-b6b8-8112476493fa-dns-svc\") pod \"bb4ea3d2-63c7-4829-b6b8-8112476493fa\" (UID: \"bb4ea3d2-63c7-4829-b6b8-8112476493fa\") " Jan 21 11:27:11 crc kubenswrapper[4824]: I0121 11:27:11.678331 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bb4ea3d2-63c7-4829-b6b8-8112476493fa-ovsdbserver-nb\") pod \"bb4ea3d2-63c7-4829-b6b8-8112476493fa\" (UID: \"bb4ea3d2-63c7-4829-b6b8-8112476493fa\") " Jan 21 11:27:11 crc kubenswrapper[4824]: I0121 11:27:11.678355 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bb4ea3d2-63c7-4829-b6b8-8112476493fa-config\") pod \"bb4ea3d2-63c7-4829-b6b8-8112476493fa\" (UID: \"bb4ea3d2-63c7-4829-b6b8-8112476493fa\") " Jan 21 11:27:11 crc kubenswrapper[4824]: I0121 11:27:11.678427 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bb4ea3d2-63c7-4829-b6b8-8112476493fa-dns-swift-storage-0\") pod \"bb4ea3d2-63c7-4829-b6b8-8112476493fa\" (UID: \"bb4ea3d2-63c7-4829-b6b8-8112476493fa\") " Jan 21 11:27:11 crc kubenswrapper[4824]: I0121 11:27:11.685179 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bb4ea3d2-63c7-4829-b6b8-8112476493fa-kube-api-access-clj2h" (OuterVolumeSpecName: "kube-api-access-clj2h") pod "bb4ea3d2-63c7-4829-b6b8-8112476493fa" (UID: "bb4ea3d2-63c7-4829-b6b8-8112476493fa"). InnerVolumeSpecName "kube-api-access-clj2h". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:27:11 crc kubenswrapper[4824]: I0121 11:27:11.715764 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bb4ea3d2-63c7-4829-b6b8-8112476493fa-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "bb4ea3d2-63c7-4829-b6b8-8112476493fa" (UID: "bb4ea3d2-63c7-4829-b6b8-8112476493fa"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:27:11 crc kubenswrapper[4824]: I0121 11:27:11.718976 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bb4ea3d2-63c7-4829-b6b8-8112476493fa-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "bb4ea3d2-63c7-4829-b6b8-8112476493fa" (UID: "bb4ea3d2-63c7-4829-b6b8-8112476493fa"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:27:11 crc kubenswrapper[4824]: I0121 11:27:11.719626 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bb4ea3d2-63c7-4829-b6b8-8112476493fa-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "bb4ea3d2-63c7-4829-b6b8-8112476493fa" (UID: "bb4ea3d2-63c7-4829-b6b8-8112476493fa"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:27:11 crc kubenswrapper[4824]: I0121 11:27:11.727252 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bb4ea3d2-63c7-4829-b6b8-8112476493fa-config" (OuterVolumeSpecName: "config") pod "bb4ea3d2-63c7-4829-b6b8-8112476493fa" (UID: "bb4ea3d2-63c7-4829-b6b8-8112476493fa"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:27:11 crc kubenswrapper[4824]: I0121 11:27:11.730814 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bb4ea3d2-63c7-4829-b6b8-8112476493fa-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "bb4ea3d2-63c7-4829-b6b8-8112476493fa" (UID: "bb4ea3d2-63c7-4829-b6b8-8112476493fa"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:27:11 crc kubenswrapper[4824]: I0121 11:27:11.787230 4824 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bb4ea3d2-63c7-4829-b6b8-8112476493fa-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:11 crc kubenswrapper[4824]: I0121 11:27:11.787263 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-clj2h\" (UniqueName: \"kubernetes.io/projected/bb4ea3d2-63c7-4829-b6b8-8112476493fa-kube-api-access-clj2h\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:11 crc kubenswrapper[4824]: I0121 11:27:11.787275 4824 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bb4ea3d2-63c7-4829-b6b8-8112476493fa-dns-svc\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:11 crc kubenswrapper[4824]: I0121 11:27:11.787283 4824 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bb4ea3d2-63c7-4829-b6b8-8112476493fa-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:11 crc kubenswrapper[4824]: I0121 11:27:11.787291 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bb4ea3d2-63c7-4829-b6b8-8112476493fa-config\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:11 crc kubenswrapper[4824]: I0121 11:27:11.787299 4824 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bb4ea3d2-63c7-4829-b6b8-8112476493fa-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:11 crc kubenswrapper[4824]: I0121 11:27:11.862320 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-79bfcdcfbd-s2v5h" Jan 21 11:27:12 crc kubenswrapper[4824]: I0121 11:27:12.202719 4824 generic.go:334] "Generic (PLEG): container finished" podID="a5721831-a6aa-4d42-b62a-041dbe5c23ad" containerID="b1d6ea0f0147f78d0d5492b8698d35f699b27a92f4975bcb0d9a7a8cc6032f2a" exitCode=0 Jan 21 11:27:12 crc kubenswrapper[4824]: I0121 11:27:12.202751 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"a5721831-a6aa-4d42-b62a-041dbe5c23ad","Type":"ContainerDied","Data":"b1d6ea0f0147f78d0d5492b8698d35f699b27a92f4975bcb0d9a7a8cc6032f2a"} Jan 21 11:27:12 crc kubenswrapper[4824]: I0121 11:27:12.204744 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-vbmkp" event={"ID":"bb4ea3d2-63c7-4829-b6b8-8112476493fa","Type":"ContainerDied","Data":"5e519680bc7c55a0ecc86ba490dcbb3969a313e5b2bc06cbb94da1a485a26d3b"} Jan 21 11:27:12 crc kubenswrapper[4824]: I0121 11:27:12.204849 4824 scope.go:117] "RemoveContainer" containerID="25b08d9ba24d1cbae3803365bca42ff0e044f0f812800389a684ef8eb1362dbe" Jan 21 11:27:12 crc kubenswrapper[4824]: I0121 11:27:12.205055 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55f844cf75-vbmkp" Jan 21 11:27:12 crc kubenswrapper[4824]: I0121 11:27:12.221593 4824 scope.go:117] "RemoveContainer" containerID="f5715a5c2dbadf161e906011876459ee94ba9a1523d9455c1d55ac180e574139" Jan 21 11:27:12 crc kubenswrapper[4824]: I0121 11:27:12.222921 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-vbmkp"] Jan 21 11:27:12 crc kubenswrapper[4824]: I0121 11:27:12.234492 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-vbmkp"] Jan 21 11:27:13 crc kubenswrapper[4824]: I0121 11:27:13.597531 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Jan 21 11:27:13 crc kubenswrapper[4824]: I0121 11:27:13.617314 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a5721831-a6aa-4d42-b62a-041dbe5c23ad-config-data\") pod \"a5721831-a6aa-4d42-b62a-041dbe5c23ad\" (UID: \"a5721831-a6aa-4d42-b62a-041dbe5c23ad\") " Jan 21 11:27:13 crc kubenswrapper[4824]: I0121 11:27:13.617359 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a5721831-a6aa-4d42-b62a-041dbe5c23ad-etc-machine-id\") pod \"a5721831-a6aa-4d42-b62a-041dbe5c23ad\" (UID: \"a5721831-a6aa-4d42-b62a-041dbe5c23ad\") " Jan 21 11:27:13 crc kubenswrapper[4824]: I0121 11:27:13.617391 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5721831-a6aa-4d42-b62a-041dbe5c23ad-combined-ca-bundle\") pod \"a5721831-a6aa-4d42-b62a-041dbe5c23ad\" (UID: \"a5721831-a6aa-4d42-b62a-041dbe5c23ad\") " Jan 21 11:27:13 crc kubenswrapper[4824]: I0121 11:27:13.617425 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a5721831-a6aa-4d42-b62a-041dbe5c23ad-config-data-custom\") pod \"a5721831-a6aa-4d42-b62a-041dbe5c23ad\" (UID: \"a5721831-a6aa-4d42-b62a-041dbe5c23ad\") " Jan 21 11:27:13 crc kubenswrapper[4824]: I0121 11:27:13.617440 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cmmwh\" (UniqueName: \"kubernetes.io/projected/a5721831-a6aa-4d42-b62a-041dbe5c23ad-kube-api-access-cmmwh\") pod \"a5721831-a6aa-4d42-b62a-041dbe5c23ad\" (UID: \"a5721831-a6aa-4d42-b62a-041dbe5c23ad\") " Jan 21 11:27:13 crc kubenswrapper[4824]: I0121 11:27:13.617443 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a5721831-a6aa-4d42-b62a-041dbe5c23ad-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "a5721831-a6aa-4d42-b62a-041dbe5c23ad" (UID: "a5721831-a6aa-4d42-b62a-041dbe5c23ad"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 21 11:27:13 crc kubenswrapper[4824]: I0121 11:27:13.617563 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a5721831-a6aa-4d42-b62a-041dbe5c23ad-scripts\") pod \"a5721831-a6aa-4d42-b62a-041dbe5c23ad\" (UID: \"a5721831-a6aa-4d42-b62a-041dbe5c23ad\") " Jan 21 11:27:13 crc kubenswrapper[4824]: I0121 11:27:13.618465 4824 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a5721831-a6aa-4d42-b62a-041dbe5c23ad-etc-machine-id\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:13 crc kubenswrapper[4824]: I0121 11:27:13.622126 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a5721831-a6aa-4d42-b62a-041dbe5c23ad-scripts" (OuterVolumeSpecName: "scripts") pod "a5721831-a6aa-4d42-b62a-041dbe5c23ad" (UID: "a5721831-a6aa-4d42-b62a-041dbe5c23ad"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:27:13 crc kubenswrapper[4824]: I0121 11:27:13.625100 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a5721831-a6aa-4d42-b62a-041dbe5c23ad-kube-api-access-cmmwh" (OuterVolumeSpecName: "kube-api-access-cmmwh") pod "a5721831-a6aa-4d42-b62a-041dbe5c23ad" (UID: "a5721831-a6aa-4d42-b62a-041dbe5c23ad"). InnerVolumeSpecName "kube-api-access-cmmwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:27:13 crc kubenswrapper[4824]: I0121 11:27:13.625099 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a5721831-a6aa-4d42-b62a-041dbe5c23ad-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "a5721831-a6aa-4d42-b62a-041dbe5c23ad" (UID: "a5721831-a6aa-4d42-b62a-041dbe5c23ad"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:27:13 crc kubenswrapper[4824]: I0121 11:27:13.662663 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a5721831-a6aa-4d42-b62a-041dbe5c23ad-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a5721831-a6aa-4d42-b62a-041dbe5c23ad" (UID: "a5721831-a6aa-4d42-b62a-041dbe5c23ad"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:27:13 crc kubenswrapper[4824]: I0121 11:27:13.688584 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a5721831-a6aa-4d42-b62a-041dbe5c23ad-config-data" (OuterVolumeSpecName: "config-data") pod "a5721831-a6aa-4d42-b62a-041dbe5c23ad" (UID: "a5721831-a6aa-4d42-b62a-041dbe5c23ad"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:27:13 crc kubenswrapper[4824]: I0121 11:27:13.719863 4824 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a5721831-a6aa-4d42-b62a-041dbe5c23ad-config-data-custom\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:13 crc kubenswrapper[4824]: I0121 11:27:13.719887 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cmmwh\" (UniqueName: \"kubernetes.io/projected/a5721831-a6aa-4d42-b62a-041dbe5c23ad-kube-api-access-cmmwh\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:13 crc kubenswrapper[4824]: I0121 11:27:13.719897 4824 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a5721831-a6aa-4d42-b62a-041dbe5c23ad-scripts\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:13 crc kubenswrapper[4824]: I0121 11:27:13.719907 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a5721831-a6aa-4d42-b62a-041dbe5c23ad-config-data\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:13 crc kubenswrapper[4824]: I0121 11:27:13.719914 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5721831-a6aa-4d42-b62a-041dbe5c23ad-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:13 crc kubenswrapper[4824]: I0121 11:27:13.854305 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-659565f76f-jdg2q" Jan 21 11:27:13 crc kubenswrapper[4824]: I0121 11:27:13.894013 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-79bfcdcfbd-s2v5h"] Jan 21 11:27:13 crc kubenswrapper[4824]: I0121 11:27:13.894170 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-79bfcdcfbd-s2v5h" podUID="d9302662-76b1-4acc-b9e3-ffde925d34b2" containerName="neutron-api" containerID="cri-o://ff804ebe93733fa9e697132922f8e30c6ee75d058b7cb8afecec7da791271c35" gracePeriod=30 Jan 21 11:27:13 crc kubenswrapper[4824]: I0121 11:27:13.894470 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-79bfcdcfbd-s2v5h" podUID="d9302662-76b1-4acc-b9e3-ffde925d34b2" containerName="neutron-httpd" containerID="cri-o://eb6d7862490b4afbdbbe422c8984287dc94d6e4a39261b3846440fb920c542fb" gracePeriod=30 Jan 21 11:27:14 crc kubenswrapper[4824]: I0121 11:27:14.070247 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bb4ea3d2-63c7-4829-b6b8-8112476493fa" path="/var/lib/kubelet/pods/bb4ea3d2-63c7-4829-b6b8-8112476493fa/volumes" Jan 21 11:27:14 crc kubenswrapper[4824]: I0121 11:27:14.222788 4824 generic.go:334] "Generic (PLEG): container finished" podID="d9302662-76b1-4acc-b9e3-ffde925d34b2" containerID="eb6d7862490b4afbdbbe422c8984287dc94d6e4a39261b3846440fb920c542fb" exitCode=0 Jan 21 11:27:14 crc kubenswrapper[4824]: I0121 11:27:14.222851 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-79bfcdcfbd-s2v5h" event={"ID":"d9302662-76b1-4acc-b9e3-ffde925d34b2","Type":"ContainerDied","Data":"eb6d7862490b4afbdbbe422c8984287dc94d6e4a39261b3846440fb920c542fb"} Jan 21 11:27:14 crc kubenswrapper[4824]: I0121 11:27:14.224740 4824 generic.go:334] "Generic (PLEG): container finished" podID="a5721831-a6aa-4d42-b62a-041dbe5c23ad" containerID="945d74317b4fe36380b7b590e48fb68b917191cdb410f4c0e01be7c34a318d74" exitCode=0 Jan 21 11:27:14 crc kubenswrapper[4824]: I0121 11:27:14.224785 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Jan 21 11:27:14 crc kubenswrapper[4824]: I0121 11:27:14.224783 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"a5721831-a6aa-4d42-b62a-041dbe5c23ad","Type":"ContainerDied","Data":"945d74317b4fe36380b7b590e48fb68b917191cdb410f4c0e01be7c34a318d74"} Jan 21 11:27:14 crc kubenswrapper[4824]: I0121 11:27:14.224922 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"a5721831-a6aa-4d42-b62a-041dbe5c23ad","Type":"ContainerDied","Data":"ccfadddd012e5287b137e7bb05a0f3a68e04a10bc60fc278e3e35eab10438b65"} Jan 21 11:27:14 crc kubenswrapper[4824]: I0121 11:27:14.224945 4824 scope.go:117] "RemoveContainer" containerID="b1d6ea0f0147f78d0d5492b8698d35f699b27a92f4975bcb0d9a7a8cc6032f2a" Jan 21 11:27:14 crc kubenswrapper[4824]: I0121 11:27:14.240914 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Jan 21 11:27:14 crc kubenswrapper[4824]: I0121 11:27:14.243397 4824 scope.go:117] "RemoveContainer" containerID="945d74317b4fe36380b7b590e48fb68b917191cdb410f4c0e01be7c34a318d74" Jan 21 11:27:14 crc kubenswrapper[4824]: I0121 11:27:14.246875 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Jan 21 11:27:14 crc kubenswrapper[4824]: I0121 11:27:14.264914 4824 scope.go:117] "RemoveContainer" containerID="b1d6ea0f0147f78d0d5492b8698d35f699b27a92f4975bcb0d9a7a8cc6032f2a" Jan 21 11:27:14 crc kubenswrapper[4824]: I0121 11:27:14.268025 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Jan 21 11:27:14 crc kubenswrapper[4824]: E0121 11:27:14.268378 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5721831-a6aa-4d42-b62a-041dbe5c23ad" containerName="probe" Jan 21 11:27:14 crc kubenswrapper[4824]: I0121 11:27:14.268391 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5721831-a6aa-4d42-b62a-041dbe5c23ad" containerName="probe" Jan 21 11:27:14 crc kubenswrapper[4824]: E0121 11:27:14.268415 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb4ea3d2-63c7-4829-b6b8-8112476493fa" containerName="init" Jan 21 11:27:14 crc kubenswrapper[4824]: I0121 11:27:14.268420 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb4ea3d2-63c7-4829-b6b8-8112476493fa" containerName="init" Jan 21 11:27:14 crc kubenswrapper[4824]: E0121 11:27:14.268429 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb4ea3d2-63c7-4829-b6b8-8112476493fa" containerName="dnsmasq-dns" Jan 21 11:27:14 crc kubenswrapper[4824]: I0121 11:27:14.268435 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb4ea3d2-63c7-4829-b6b8-8112476493fa" containerName="dnsmasq-dns" Jan 21 11:27:14 crc kubenswrapper[4824]: E0121 11:27:14.268448 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5721831-a6aa-4d42-b62a-041dbe5c23ad" containerName="cinder-scheduler" Jan 21 11:27:14 crc kubenswrapper[4824]: I0121 11:27:14.268453 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5721831-a6aa-4d42-b62a-041dbe5c23ad" containerName="cinder-scheduler" Jan 21 11:27:14 crc kubenswrapper[4824]: I0121 11:27:14.268632 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="a5721831-a6aa-4d42-b62a-041dbe5c23ad" containerName="cinder-scheduler" Jan 21 11:27:14 crc kubenswrapper[4824]: I0121 11:27:14.268645 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="bb4ea3d2-63c7-4829-b6b8-8112476493fa" containerName="dnsmasq-dns" Jan 21 11:27:14 crc kubenswrapper[4824]: I0121 11:27:14.268667 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="a5721831-a6aa-4d42-b62a-041dbe5c23ad" containerName="probe" Jan 21 11:27:14 crc kubenswrapper[4824]: E0121 11:27:14.269163 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b1d6ea0f0147f78d0d5492b8698d35f699b27a92f4975bcb0d9a7a8cc6032f2a\": container with ID starting with b1d6ea0f0147f78d0d5492b8698d35f699b27a92f4975bcb0d9a7a8cc6032f2a not found: ID does not exist" containerID="b1d6ea0f0147f78d0d5492b8698d35f699b27a92f4975bcb0d9a7a8cc6032f2a" Jan 21 11:27:14 crc kubenswrapper[4824]: I0121 11:27:14.269195 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b1d6ea0f0147f78d0d5492b8698d35f699b27a92f4975bcb0d9a7a8cc6032f2a"} err="failed to get container status \"b1d6ea0f0147f78d0d5492b8698d35f699b27a92f4975bcb0d9a7a8cc6032f2a\": rpc error: code = NotFound desc = could not find container \"b1d6ea0f0147f78d0d5492b8698d35f699b27a92f4975bcb0d9a7a8cc6032f2a\": container with ID starting with b1d6ea0f0147f78d0d5492b8698d35f699b27a92f4975bcb0d9a7a8cc6032f2a not found: ID does not exist" Jan 21 11:27:14 crc kubenswrapper[4824]: I0121 11:27:14.269220 4824 scope.go:117] "RemoveContainer" containerID="945d74317b4fe36380b7b590e48fb68b917191cdb410f4c0e01be7c34a318d74" Jan 21 11:27:14 crc kubenswrapper[4824]: I0121 11:27:14.269509 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Jan 21 11:27:14 crc kubenswrapper[4824]: E0121 11:27:14.269569 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"945d74317b4fe36380b7b590e48fb68b917191cdb410f4c0e01be7c34a318d74\": container with ID starting with 945d74317b4fe36380b7b590e48fb68b917191cdb410f4c0e01be7c34a318d74 not found: ID does not exist" containerID="945d74317b4fe36380b7b590e48fb68b917191cdb410f4c0e01be7c34a318d74" Jan 21 11:27:14 crc kubenswrapper[4824]: I0121 11:27:14.269592 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"945d74317b4fe36380b7b590e48fb68b917191cdb410f4c0e01be7c34a318d74"} err="failed to get container status \"945d74317b4fe36380b7b590e48fb68b917191cdb410f4c0e01be7c34a318d74\": rpc error: code = NotFound desc = could not find container \"945d74317b4fe36380b7b590e48fb68b917191cdb410f4c0e01be7c34a318d74\": container with ID starting with 945d74317b4fe36380b7b590e48fb68b917191cdb410f4c0e01be7c34a318d74 not found: ID does not exist" Jan 21 11:27:14 crc kubenswrapper[4824]: I0121 11:27:14.270975 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Jan 21 11:27:14 crc kubenswrapper[4824]: I0121 11:27:14.278795 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Jan 21 11:27:14 crc kubenswrapper[4824]: I0121 11:27:14.328413 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e0f0267-58d2-4690-9187-8ddf03ef082e-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"2e0f0267-58d2-4690-9187-8ddf03ef082e\") " pod="openstack/cinder-scheduler-0" Jan 21 11:27:14 crc kubenswrapper[4824]: I0121 11:27:14.328535 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2e0f0267-58d2-4690-9187-8ddf03ef082e-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"2e0f0267-58d2-4690-9187-8ddf03ef082e\") " pod="openstack/cinder-scheduler-0" Jan 21 11:27:14 crc kubenswrapper[4824]: I0121 11:27:14.328710 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2e0f0267-58d2-4690-9187-8ddf03ef082e-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"2e0f0267-58d2-4690-9187-8ddf03ef082e\") " pod="openstack/cinder-scheduler-0" Jan 21 11:27:14 crc kubenswrapper[4824]: I0121 11:27:14.329316 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9fdvv\" (UniqueName: \"kubernetes.io/projected/2e0f0267-58d2-4690-9187-8ddf03ef082e-kube-api-access-9fdvv\") pod \"cinder-scheduler-0\" (UID: \"2e0f0267-58d2-4690-9187-8ddf03ef082e\") " pod="openstack/cinder-scheduler-0" Jan 21 11:27:14 crc kubenswrapper[4824]: I0121 11:27:14.329361 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2e0f0267-58d2-4690-9187-8ddf03ef082e-scripts\") pod \"cinder-scheduler-0\" (UID: \"2e0f0267-58d2-4690-9187-8ddf03ef082e\") " pod="openstack/cinder-scheduler-0" Jan 21 11:27:14 crc kubenswrapper[4824]: I0121 11:27:14.329378 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e0f0267-58d2-4690-9187-8ddf03ef082e-config-data\") pod \"cinder-scheduler-0\" (UID: \"2e0f0267-58d2-4690-9187-8ddf03ef082e\") " pod="openstack/cinder-scheduler-0" Jan 21 11:27:14 crc kubenswrapper[4824]: I0121 11:27:14.430748 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9fdvv\" (UniqueName: \"kubernetes.io/projected/2e0f0267-58d2-4690-9187-8ddf03ef082e-kube-api-access-9fdvv\") pod \"cinder-scheduler-0\" (UID: \"2e0f0267-58d2-4690-9187-8ddf03ef082e\") " pod="openstack/cinder-scheduler-0" Jan 21 11:27:14 crc kubenswrapper[4824]: I0121 11:27:14.430798 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2e0f0267-58d2-4690-9187-8ddf03ef082e-scripts\") pod \"cinder-scheduler-0\" (UID: \"2e0f0267-58d2-4690-9187-8ddf03ef082e\") " pod="openstack/cinder-scheduler-0" Jan 21 11:27:14 crc kubenswrapper[4824]: I0121 11:27:14.430820 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e0f0267-58d2-4690-9187-8ddf03ef082e-config-data\") pod \"cinder-scheduler-0\" (UID: \"2e0f0267-58d2-4690-9187-8ddf03ef082e\") " pod="openstack/cinder-scheduler-0" Jan 21 11:27:14 crc kubenswrapper[4824]: I0121 11:27:14.430863 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e0f0267-58d2-4690-9187-8ddf03ef082e-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"2e0f0267-58d2-4690-9187-8ddf03ef082e\") " pod="openstack/cinder-scheduler-0" Jan 21 11:27:14 crc kubenswrapper[4824]: I0121 11:27:14.430895 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2e0f0267-58d2-4690-9187-8ddf03ef082e-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"2e0f0267-58d2-4690-9187-8ddf03ef082e\") " pod="openstack/cinder-scheduler-0" Jan 21 11:27:14 crc kubenswrapper[4824]: I0121 11:27:14.430926 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2e0f0267-58d2-4690-9187-8ddf03ef082e-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"2e0f0267-58d2-4690-9187-8ddf03ef082e\") " pod="openstack/cinder-scheduler-0" Jan 21 11:27:14 crc kubenswrapper[4824]: I0121 11:27:14.431025 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2e0f0267-58d2-4690-9187-8ddf03ef082e-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"2e0f0267-58d2-4690-9187-8ddf03ef082e\") " pod="openstack/cinder-scheduler-0" Jan 21 11:27:14 crc kubenswrapper[4824]: I0121 11:27:14.435144 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2e0f0267-58d2-4690-9187-8ddf03ef082e-scripts\") pod \"cinder-scheduler-0\" (UID: \"2e0f0267-58d2-4690-9187-8ddf03ef082e\") " pod="openstack/cinder-scheduler-0" Jan 21 11:27:14 crc kubenswrapper[4824]: I0121 11:27:14.435310 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2e0f0267-58d2-4690-9187-8ddf03ef082e-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"2e0f0267-58d2-4690-9187-8ddf03ef082e\") " pod="openstack/cinder-scheduler-0" Jan 21 11:27:14 crc kubenswrapper[4824]: I0121 11:27:14.435929 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e0f0267-58d2-4690-9187-8ddf03ef082e-config-data\") pod \"cinder-scheduler-0\" (UID: \"2e0f0267-58d2-4690-9187-8ddf03ef082e\") " pod="openstack/cinder-scheduler-0" Jan 21 11:27:14 crc kubenswrapper[4824]: I0121 11:27:14.436372 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e0f0267-58d2-4690-9187-8ddf03ef082e-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"2e0f0267-58d2-4690-9187-8ddf03ef082e\") " pod="openstack/cinder-scheduler-0" Jan 21 11:27:14 crc kubenswrapper[4824]: I0121 11:27:14.447849 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9fdvv\" (UniqueName: \"kubernetes.io/projected/2e0f0267-58d2-4690-9187-8ddf03ef082e-kube-api-access-9fdvv\") pod \"cinder-scheduler-0\" (UID: \"2e0f0267-58d2-4690-9187-8ddf03ef082e\") " pod="openstack/cinder-scheduler-0" Jan 21 11:27:14 crc kubenswrapper[4824]: I0121 11:27:14.585492 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Jan 21 11:27:14 crc kubenswrapper[4824]: I0121 11:27:14.976505 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Jan 21 11:27:14 crc kubenswrapper[4824]: W0121 11:27:14.987046 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2e0f0267_58d2_4690_9187_8ddf03ef082e.slice/crio-e4b13faacb70f79ca1a74ffe6705e4d4d8ba2ae3dd39c51fa021fe0341cbc621 WatchSource:0}: Error finding container e4b13faacb70f79ca1a74ffe6705e4d4d8ba2ae3dd39c51fa021fe0341cbc621: Status 404 returned error can't find the container with id e4b13faacb70f79ca1a74ffe6705e4d4d8ba2ae3dd39c51fa021fe0341cbc621 Jan 21 11:27:15 crc kubenswrapper[4824]: I0121 11:27:15.233310 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"2e0f0267-58d2-4690-9187-8ddf03ef082e","Type":"ContainerStarted","Data":"e4b13faacb70f79ca1a74ffe6705e4d4d8ba2ae3dd39c51fa021fe0341cbc621"} Jan 21 11:27:16 crc kubenswrapper[4824]: I0121 11:27:16.064884 4824 patch_prober.go:28] interesting pod/machine-config-daemon-zqjg9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 21 11:27:16 crc kubenswrapper[4824]: I0121 11:27:16.065275 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 21 11:27:16 crc kubenswrapper[4824]: I0121 11:27:16.072309 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a5721831-a6aa-4d42-b62a-041dbe5c23ad" path="/var/lib/kubelet/pods/a5721831-a6aa-4d42-b62a-041dbe5c23ad/volumes" Jan 21 11:27:16 crc kubenswrapper[4824]: I0121 11:27:16.073519 4824 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" Jan 21 11:27:16 crc kubenswrapper[4824]: I0121 11:27:16.074147 4824 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"365550d435931ba3a56d32c848ae2c97f8d7721cf74faadf03f5fc5d15e5b5a5"} pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 21 11:27:16 crc kubenswrapper[4824]: I0121 11:27:16.074198 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" containerName="machine-config-daemon" containerID="cri-o://365550d435931ba3a56d32c848ae2c97f8d7721cf74faadf03f5fc5d15e5b5a5" gracePeriod=600 Jan 21 11:27:16 crc kubenswrapper[4824]: I0121 11:27:16.079631 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Jan 21 11:27:16 crc kubenswrapper[4824]: I0121 11:27:16.246013 4824 generic.go:334] "Generic (PLEG): container finished" podID="33f3d922-4ffe-409b-a49a-d88c85898260" containerID="365550d435931ba3a56d32c848ae2c97f8d7721cf74faadf03f5fc5d15e5b5a5" exitCode=0 Jan 21 11:27:16 crc kubenswrapper[4824]: I0121 11:27:16.246250 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" event={"ID":"33f3d922-4ffe-409b-a49a-d88c85898260","Type":"ContainerDied","Data":"365550d435931ba3a56d32c848ae2c97f8d7721cf74faadf03f5fc5d15e5b5a5"} Jan 21 11:27:16 crc kubenswrapper[4824]: I0121 11:27:16.246281 4824 scope.go:117] "RemoveContainer" containerID="0c03e9f8dae3ff83ead8b2e38560a1562de223bd29105640743d2bd8e0cdca44" Jan 21 11:27:16 crc kubenswrapper[4824]: I0121 11:27:16.252409 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"2e0f0267-58d2-4690-9187-8ddf03ef082e","Type":"ContainerStarted","Data":"17331213015ca691828e0bb69e808e23e5f81297d08aad89e5b16c716d858917"} Jan 21 11:27:16 crc kubenswrapper[4824]: I0121 11:27:16.252439 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"2e0f0267-58d2-4690-9187-8ddf03ef082e","Type":"ContainerStarted","Data":"f11ed05be21bc138f79d8f98c30bec5e9c23f6fd84a904ab2f95213a4528812d"} Jan 21 11:27:16 crc kubenswrapper[4824]: I0121 11:27:16.256398 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-8654bbf998-p8crm" Jan 21 11:27:16 crc kubenswrapper[4824]: I0121 11:27:16.271833 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=2.271821918 podStartE2EDuration="2.271821918s" podCreationTimestamp="2026-01-21 11:27:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:27:16.265323351 +0000 UTC m=+978.558352643" watchObservedRunningTime="2026-01-21 11:27:16.271821918 +0000 UTC m=+978.564851209" Jan 21 11:27:16 crc kubenswrapper[4824]: I0121 11:27:16.380920 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-8654bbf998-p8crm" Jan 21 11:27:16 crc kubenswrapper[4824]: I0121 11:27:16.511819 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-55f844cf75-vbmkp" podUID="bb4ea3d2-63c7-4829-b6b8-8112476493fa" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.150:5353: i/o timeout" Jan 21 11:27:17 crc kubenswrapper[4824]: I0121 11:27:17.262159 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" event={"ID":"33f3d922-4ffe-409b-a49a-d88c85898260","Type":"ContainerStarted","Data":"74ad8e1889db466232cc3528a7e51d12f4be65b09faed3c15716a73518537690"} Jan 21 11:27:17 crc kubenswrapper[4824]: I0121 11:27:17.264680 4824 generic.go:334] "Generic (PLEG): container finished" podID="d9302662-76b1-4acc-b9e3-ffde925d34b2" containerID="ff804ebe93733fa9e697132922f8e30c6ee75d058b7cb8afecec7da791271c35" exitCode=0 Jan 21 11:27:17 crc kubenswrapper[4824]: I0121 11:27:17.264778 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-79bfcdcfbd-s2v5h" event={"ID":"d9302662-76b1-4acc-b9e3-ffde925d34b2","Type":"ContainerDied","Data":"ff804ebe93733fa9e697132922f8e30c6ee75d058b7cb8afecec7da791271c35"} Jan 21 11:27:17 crc kubenswrapper[4824]: I0121 11:27:17.264831 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-79bfcdcfbd-s2v5h" event={"ID":"d9302662-76b1-4acc-b9e3-ffde925d34b2","Type":"ContainerDied","Data":"2df3211773129d1b5d01ef3e6ca36fb1965741e7764384f48d28c04cb0d0626d"} Jan 21 11:27:17 crc kubenswrapper[4824]: I0121 11:27:17.264843 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2df3211773129d1b5d01ef3e6ca36fb1965741e7764384f48d28c04cb0d0626d" Jan 21 11:27:17 crc kubenswrapper[4824]: I0121 11:27:17.276527 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-79bfcdcfbd-s2v5h" Jan 21 11:27:17 crc kubenswrapper[4824]: I0121 11:27:17.287768 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/d9302662-76b1-4acc-b9e3-ffde925d34b2-ovndb-tls-certs\") pod \"d9302662-76b1-4acc-b9e3-ffde925d34b2\" (UID: \"d9302662-76b1-4acc-b9e3-ffde925d34b2\") " Jan 21 11:27:17 crc kubenswrapper[4824]: I0121 11:27:17.287863 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/d9302662-76b1-4acc-b9e3-ffde925d34b2-httpd-config\") pod \"d9302662-76b1-4acc-b9e3-ffde925d34b2\" (UID: \"d9302662-76b1-4acc-b9e3-ffde925d34b2\") " Jan 21 11:27:17 crc kubenswrapper[4824]: I0121 11:27:17.287984 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9x8cf\" (UniqueName: \"kubernetes.io/projected/d9302662-76b1-4acc-b9e3-ffde925d34b2-kube-api-access-9x8cf\") pod \"d9302662-76b1-4acc-b9e3-ffde925d34b2\" (UID: \"d9302662-76b1-4acc-b9e3-ffde925d34b2\") " Jan 21 11:27:17 crc kubenswrapper[4824]: I0121 11:27:17.288123 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/d9302662-76b1-4acc-b9e3-ffde925d34b2-config\") pod \"d9302662-76b1-4acc-b9e3-ffde925d34b2\" (UID: \"d9302662-76b1-4acc-b9e3-ffde925d34b2\") " Jan 21 11:27:17 crc kubenswrapper[4824]: I0121 11:27:17.288145 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9302662-76b1-4acc-b9e3-ffde925d34b2-combined-ca-bundle\") pod \"d9302662-76b1-4acc-b9e3-ffde925d34b2\" (UID: \"d9302662-76b1-4acc-b9e3-ffde925d34b2\") " Jan 21 11:27:17 crc kubenswrapper[4824]: I0121 11:27:17.294007 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d9302662-76b1-4acc-b9e3-ffde925d34b2-kube-api-access-9x8cf" (OuterVolumeSpecName: "kube-api-access-9x8cf") pod "d9302662-76b1-4acc-b9e3-ffde925d34b2" (UID: "d9302662-76b1-4acc-b9e3-ffde925d34b2"). InnerVolumeSpecName "kube-api-access-9x8cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:27:17 crc kubenswrapper[4824]: I0121 11:27:17.305043 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d9302662-76b1-4acc-b9e3-ffde925d34b2-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "d9302662-76b1-4acc-b9e3-ffde925d34b2" (UID: "d9302662-76b1-4acc-b9e3-ffde925d34b2"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:27:17 crc kubenswrapper[4824]: I0121 11:27:17.337057 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d9302662-76b1-4acc-b9e3-ffde925d34b2-config" (OuterVolumeSpecName: "config") pod "d9302662-76b1-4acc-b9e3-ffde925d34b2" (UID: "d9302662-76b1-4acc-b9e3-ffde925d34b2"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:27:17 crc kubenswrapper[4824]: I0121 11:27:17.361914 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d9302662-76b1-4acc-b9e3-ffde925d34b2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d9302662-76b1-4acc-b9e3-ffde925d34b2" (UID: "d9302662-76b1-4acc-b9e3-ffde925d34b2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:27:17 crc kubenswrapper[4824]: I0121 11:27:17.376438 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d9302662-76b1-4acc-b9e3-ffde925d34b2-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "d9302662-76b1-4acc-b9e3-ffde925d34b2" (UID: "d9302662-76b1-4acc-b9e3-ffde925d34b2"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:27:17 crc kubenswrapper[4824]: I0121 11:27:17.390235 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/d9302662-76b1-4acc-b9e3-ffde925d34b2-config\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:17 crc kubenswrapper[4824]: I0121 11:27:17.390262 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9302662-76b1-4acc-b9e3-ffde925d34b2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:17 crc kubenswrapper[4824]: I0121 11:27:17.390274 4824 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/d9302662-76b1-4acc-b9e3-ffde925d34b2-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:17 crc kubenswrapper[4824]: I0121 11:27:17.390281 4824 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/d9302662-76b1-4acc-b9e3-ffde925d34b2-httpd-config\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:17 crc kubenswrapper[4824]: I0121 11:27:17.390289 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9x8cf\" (UniqueName: \"kubernetes.io/projected/d9302662-76b1-4acc-b9e3-ffde925d34b2-kube-api-access-9x8cf\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:17 crc kubenswrapper[4824]: I0121 11:27:17.883367 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-6c7f965fb4-8slr2" Jan 21 11:27:17 crc kubenswrapper[4824]: I0121 11:27:17.982186 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-6c7f965fb4-8slr2" Jan 21 11:27:18 crc kubenswrapper[4824]: I0121 11:27:18.042027 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-6b9b444cf4-q85qq"] Jan 21 11:27:18 crc kubenswrapper[4824]: I0121 11:27:18.042234 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-6b9b444cf4-q85qq" podUID="4ed8e1b1-57c3-4b01-bbd7-9dc5125f4f94" containerName="barbican-api-log" containerID="cri-o://41e3cbe62a3cfe1f1349c3e05b9d5169c6ca9dd78e1f915854df40b91a504d55" gracePeriod=30 Jan 21 11:27:18 crc kubenswrapper[4824]: I0121 11:27:18.042365 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-6b9b444cf4-q85qq" podUID="4ed8e1b1-57c3-4b01-bbd7-9dc5125f4f94" containerName="barbican-api" containerID="cri-o://820df2cc441bc3b9b3b9761588c4025b1f941d3a17f1376789c1aa427ba78ecd" gracePeriod=30 Jan 21 11:27:18 crc kubenswrapper[4824]: I0121 11:27:18.280611 4824 generic.go:334] "Generic (PLEG): container finished" podID="4ed8e1b1-57c3-4b01-bbd7-9dc5125f4f94" containerID="41e3cbe62a3cfe1f1349c3e05b9d5169c6ca9dd78e1f915854df40b91a504d55" exitCode=143 Jan 21 11:27:18 crc kubenswrapper[4824]: I0121 11:27:18.281057 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6b9b444cf4-q85qq" event={"ID":"4ed8e1b1-57c3-4b01-bbd7-9dc5125f4f94","Type":"ContainerDied","Data":"41e3cbe62a3cfe1f1349c3e05b9d5169c6ca9dd78e1f915854df40b91a504d55"} Jan 21 11:27:18 crc kubenswrapper[4824]: I0121 11:27:18.281149 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-79bfcdcfbd-s2v5h" Jan 21 11:27:18 crc kubenswrapper[4824]: I0121 11:27:18.299183 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-79bfcdcfbd-s2v5h"] Jan 21 11:27:18 crc kubenswrapper[4824]: I0121 11:27:18.305463 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-79bfcdcfbd-s2v5h"] Jan 21 11:27:18 crc kubenswrapper[4824]: I0121 11:27:18.752191 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-6c5fc769c9-spmzr" Jan 21 11:27:19 crc kubenswrapper[4824]: I0121 11:27:19.586181 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Jan 21 11:27:20 crc kubenswrapper[4824]: I0121 11:27:20.057161 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d9302662-76b1-4acc-b9e3-ffde925d34b2" path="/var/lib/kubelet/pods/d9302662-76b1-4acc-b9e3-ffde925d34b2/volumes" Jan 21 11:27:20 crc kubenswrapper[4824]: I0121 11:27:20.794839 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Jan 21 11:27:20 crc kubenswrapper[4824]: E0121 11:27:20.795781 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d9302662-76b1-4acc-b9e3-ffde925d34b2" containerName="neutron-api" Jan 21 11:27:20 crc kubenswrapper[4824]: I0121 11:27:20.795873 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="d9302662-76b1-4acc-b9e3-ffde925d34b2" containerName="neutron-api" Jan 21 11:27:20 crc kubenswrapper[4824]: E0121 11:27:20.795943 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d9302662-76b1-4acc-b9e3-ffde925d34b2" containerName="neutron-httpd" Jan 21 11:27:20 crc kubenswrapper[4824]: I0121 11:27:20.796017 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="d9302662-76b1-4acc-b9e3-ffde925d34b2" containerName="neutron-httpd" Jan 21 11:27:20 crc kubenswrapper[4824]: I0121 11:27:20.796235 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="d9302662-76b1-4acc-b9e3-ffde925d34b2" containerName="neutron-httpd" Jan 21 11:27:20 crc kubenswrapper[4824]: I0121 11:27:20.796292 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="d9302662-76b1-4acc-b9e3-ffde925d34b2" containerName="neutron-api" Jan 21 11:27:20 crc kubenswrapper[4824]: I0121 11:27:20.796839 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Jan 21 11:27:20 crc kubenswrapper[4824]: I0121 11:27:20.799337 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Jan 21 11:27:20 crc kubenswrapper[4824]: I0121 11:27:20.799337 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Jan 21 11:27:20 crc kubenswrapper[4824]: I0121 11:27:20.799388 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-mp8h2" Jan 21 11:27:20 crc kubenswrapper[4824]: I0121 11:27:20.805516 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Jan 21 11:27:20 crc kubenswrapper[4824]: I0121 11:27:20.841452 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/1af94d08-6d1e-4a97-8f82-27a1755d17a1-openstack-config-secret\") pod \"openstackclient\" (UID: \"1af94d08-6d1e-4a97-8f82-27a1755d17a1\") " pod="openstack/openstackclient" Jan 21 11:27:20 crc kubenswrapper[4824]: I0121 11:27:20.841495 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/1af94d08-6d1e-4a97-8f82-27a1755d17a1-openstack-config\") pod \"openstackclient\" (UID: \"1af94d08-6d1e-4a97-8f82-27a1755d17a1\") " pod="openstack/openstackclient" Jan 21 11:27:20 crc kubenswrapper[4824]: I0121 11:27:20.841780 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1af94d08-6d1e-4a97-8f82-27a1755d17a1-combined-ca-bundle\") pod \"openstackclient\" (UID: \"1af94d08-6d1e-4a97-8f82-27a1755d17a1\") " pod="openstack/openstackclient" Jan 21 11:27:20 crc kubenswrapper[4824]: I0121 11:27:20.841848 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tzwj8\" (UniqueName: \"kubernetes.io/projected/1af94d08-6d1e-4a97-8f82-27a1755d17a1-kube-api-access-tzwj8\") pod \"openstackclient\" (UID: \"1af94d08-6d1e-4a97-8f82-27a1755d17a1\") " pod="openstack/openstackclient" Jan 21 11:27:20 crc kubenswrapper[4824]: I0121 11:27:20.943896 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1af94d08-6d1e-4a97-8f82-27a1755d17a1-combined-ca-bundle\") pod \"openstackclient\" (UID: \"1af94d08-6d1e-4a97-8f82-27a1755d17a1\") " pod="openstack/openstackclient" Jan 21 11:27:20 crc kubenswrapper[4824]: I0121 11:27:20.943949 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tzwj8\" (UniqueName: \"kubernetes.io/projected/1af94d08-6d1e-4a97-8f82-27a1755d17a1-kube-api-access-tzwj8\") pod \"openstackclient\" (UID: \"1af94d08-6d1e-4a97-8f82-27a1755d17a1\") " pod="openstack/openstackclient" Jan 21 11:27:20 crc kubenswrapper[4824]: I0121 11:27:20.944088 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/1af94d08-6d1e-4a97-8f82-27a1755d17a1-openstack-config-secret\") pod \"openstackclient\" (UID: \"1af94d08-6d1e-4a97-8f82-27a1755d17a1\") " pod="openstack/openstackclient" Jan 21 11:27:20 crc kubenswrapper[4824]: I0121 11:27:20.944119 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/1af94d08-6d1e-4a97-8f82-27a1755d17a1-openstack-config\") pod \"openstackclient\" (UID: \"1af94d08-6d1e-4a97-8f82-27a1755d17a1\") " pod="openstack/openstackclient" Jan 21 11:27:20 crc kubenswrapper[4824]: I0121 11:27:20.944858 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/1af94d08-6d1e-4a97-8f82-27a1755d17a1-openstack-config\") pod \"openstackclient\" (UID: \"1af94d08-6d1e-4a97-8f82-27a1755d17a1\") " pod="openstack/openstackclient" Jan 21 11:27:20 crc kubenswrapper[4824]: I0121 11:27:20.948719 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1af94d08-6d1e-4a97-8f82-27a1755d17a1-combined-ca-bundle\") pod \"openstackclient\" (UID: \"1af94d08-6d1e-4a97-8f82-27a1755d17a1\") " pod="openstack/openstackclient" Jan 21 11:27:20 crc kubenswrapper[4824]: I0121 11:27:20.954749 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/1af94d08-6d1e-4a97-8f82-27a1755d17a1-openstack-config-secret\") pod \"openstackclient\" (UID: \"1af94d08-6d1e-4a97-8f82-27a1755d17a1\") " pod="openstack/openstackclient" Jan 21 11:27:20 crc kubenswrapper[4824]: I0121 11:27:20.963247 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tzwj8\" (UniqueName: \"kubernetes.io/projected/1af94d08-6d1e-4a97-8f82-27a1755d17a1-kube-api-access-tzwj8\") pod \"openstackclient\" (UID: \"1af94d08-6d1e-4a97-8f82-27a1755d17a1\") " pod="openstack/openstackclient" Jan 21 11:27:21 crc kubenswrapper[4824]: I0121 11:27:21.114033 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Jan 21 11:27:21 crc kubenswrapper[4824]: I0121 11:27:21.198802 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-6b9b444cf4-q85qq" podUID="4ed8e1b1-57c3-4b01-bbd7-9dc5125f4f94" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.163:9311/healthcheck\": read tcp 10.217.0.2:44672->10.217.0.163:9311: read: connection reset by peer" Jan 21 11:27:21 crc kubenswrapper[4824]: I0121 11:27:21.199096 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-6b9b444cf4-q85qq" podUID="4ed8e1b1-57c3-4b01-bbd7-9dc5125f4f94" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.163:9311/healthcheck\": read tcp 10.217.0.2:44670->10.217.0.163:9311: read: connection reset by peer" Jan 21 11:27:21 crc kubenswrapper[4824]: I0121 11:27:21.302460 4824 generic.go:334] "Generic (PLEG): container finished" podID="4ed8e1b1-57c3-4b01-bbd7-9dc5125f4f94" containerID="820df2cc441bc3b9b3b9761588c4025b1f941d3a17f1376789c1aa427ba78ecd" exitCode=0 Jan 21 11:27:21 crc kubenswrapper[4824]: I0121 11:27:21.302653 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6b9b444cf4-q85qq" event={"ID":"4ed8e1b1-57c3-4b01-bbd7-9dc5125f4f94","Type":"ContainerDied","Data":"820df2cc441bc3b9b3b9761588c4025b1f941d3a17f1376789c1aa427ba78ecd"} Jan 21 11:27:21 crc kubenswrapper[4824]: I0121 11:27:21.467073 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6b9b444cf4-q85qq" Jan 21 11:27:21 crc kubenswrapper[4824]: I0121 11:27:21.510487 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Jan 21 11:27:21 crc kubenswrapper[4824]: W0121 11:27:21.511259 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1af94d08_6d1e_4a97_8f82_27a1755d17a1.slice/crio-1bea49d26452dfba81b66bbb6ec52dc54c631874ef4a66db9dfc3284bdf2b640 WatchSource:0}: Error finding container 1bea49d26452dfba81b66bbb6ec52dc54c631874ef4a66db9dfc3284bdf2b640: Status 404 returned error can't find the container with id 1bea49d26452dfba81b66bbb6ec52dc54c631874ef4a66db9dfc3284bdf2b640 Jan 21 11:27:21 crc kubenswrapper[4824]: I0121 11:27:21.552728 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vnlw6\" (UniqueName: \"kubernetes.io/projected/4ed8e1b1-57c3-4b01-bbd7-9dc5125f4f94-kube-api-access-vnlw6\") pod \"4ed8e1b1-57c3-4b01-bbd7-9dc5125f4f94\" (UID: \"4ed8e1b1-57c3-4b01-bbd7-9dc5125f4f94\") " Jan 21 11:27:21 crc kubenswrapper[4824]: I0121 11:27:21.552794 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4ed8e1b1-57c3-4b01-bbd7-9dc5125f4f94-config-data\") pod \"4ed8e1b1-57c3-4b01-bbd7-9dc5125f4f94\" (UID: \"4ed8e1b1-57c3-4b01-bbd7-9dc5125f4f94\") " Jan 21 11:27:21 crc kubenswrapper[4824]: I0121 11:27:21.552820 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4ed8e1b1-57c3-4b01-bbd7-9dc5125f4f94-config-data-custom\") pod \"4ed8e1b1-57c3-4b01-bbd7-9dc5125f4f94\" (UID: \"4ed8e1b1-57c3-4b01-bbd7-9dc5125f4f94\") " Jan 21 11:27:21 crc kubenswrapper[4824]: I0121 11:27:21.552884 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4ed8e1b1-57c3-4b01-bbd7-9dc5125f4f94-logs\") pod \"4ed8e1b1-57c3-4b01-bbd7-9dc5125f4f94\" (UID: \"4ed8e1b1-57c3-4b01-bbd7-9dc5125f4f94\") " Jan 21 11:27:21 crc kubenswrapper[4824]: I0121 11:27:21.552911 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ed8e1b1-57c3-4b01-bbd7-9dc5125f4f94-combined-ca-bundle\") pod \"4ed8e1b1-57c3-4b01-bbd7-9dc5125f4f94\" (UID: \"4ed8e1b1-57c3-4b01-bbd7-9dc5125f4f94\") " Jan 21 11:27:21 crc kubenswrapper[4824]: I0121 11:27:21.553286 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4ed8e1b1-57c3-4b01-bbd7-9dc5125f4f94-logs" (OuterVolumeSpecName: "logs") pod "4ed8e1b1-57c3-4b01-bbd7-9dc5125f4f94" (UID: "4ed8e1b1-57c3-4b01-bbd7-9dc5125f4f94"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:27:21 crc kubenswrapper[4824]: I0121 11:27:21.553518 4824 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4ed8e1b1-57c3-4b01-bbd7-9dc5125f4f94-logs\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:21 crc kubenswrapper[4824]: I0121 11:27:21.557510 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4ed8e1b1-57c3-4b01-bbd7-9dc5125f4f94-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "4ed8e1b1-57c3-4b01-bbd7-9dc5125f4f94" (UID: "4ed8e1b1-57c3-4b01-bbd7-9dc5125f4f94"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:27:21 crc kubenswrapper[4824]: I0121 11:27:21.557901 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4ed8e1b1-57c3-4b01-bbd7-9dc5125f4f94-kube-api-access-vnlw6" (OuterVolumeSpecName: "kube-api-access-vnlw6") pod "4ed8e1b1-57c3-4b01-bbd7-9dc5125f4f94" (UID: "4ed8e1b1-57c3-4b01-bbd7-9dc5125f4f94"). InnerVolumeSpecName "kube-api-access-vnlw6". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:27:21 crc kubenswrapper[4824]: I0121 11:27:21.574598 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4ed8e1b1-57c3-4b01-bbd7-9dc5125f4f94-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4ed8e1b1-57c3-4b01-bbd7-9dc5125f4f94" (UID: "4ed8e1b1-57c3-4b01-bbd7-9dc5125f4f94"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:27:21 crc kubenswrapper[4824]: I0121 11:27:21.591682 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4ed8e1b1-57c3-4b01-bbd7-9dc5125f4f94-config-data" (OuterVolumeSpecName: "config-data") pod "4ed8e1b1-57c3-4b01-bbd7-9dc5125f4f94" (UID: "4ed8e1b1-57c3-4b01-bbd7-9dc5125f4f94"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:27:21 crc kubenswrapper[4824]: I0121 11:27:21.655172 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ed8e1b1-57c3-4b01-bbd7-9dc5125f4f94-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:21 crc kubenswrapper[4824]: I0121 11:27:21.655908 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vnlw6\" (UniqueName: \"kubernetes.io/projected/4ed8e1b1-57c3-4b01-bbd7-9dc5125f4f94-kube-api-access-vnlw6\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:21 crc kubenswrapper[4824]: I0121 11:27:21.656028 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4ed8e1b1-57c3-4b01-bbd7-9dc5125f4f94-config-data\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:21 crc kubenswrapper[4824]: I0121 11:27:21.656086 4824 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4ed8e1b1-57c3-4b01-bbd7-9dc5125f4f94-config-data-custom\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:22 crc kubenswrapper[4824]: I0121 11:27:22.318080 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6b9b444cf4-q85qq" event={"ID":"4ed8e1b1-57c3-4b01-bbd7-9dc5125f4f94","Type":"ContainerDied","Data":"62531d41bed5d935b440d701f3d3a82064300927572a77364b43daf9ada74a10"} Jan 21 11:27:22 crc kubenswrapper[4824]: I0121 11:27:22.318307 4824 scope.go:117] "RemoveContainer" containerID="820df2cc441bc3b9b3b9761588c4025b1f941d3a17f1376789c1aa427ba78ecd" Jan 21 11:27:22 crc kubenswrapper[4824]: I0121 11:27:22.318456 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6b9b444cf4-q85qq" Jan 21 11:27:22 crc kubenswrapper[4824]: I0121 11:27:22.321527 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"1af94d08-6d1e-4a97-8f82-27a1755d17a1","Type":"ContainerStarted","Data":"1bea49d26452dfba81b66bbb6ec52dc54c631874ef4a66db9dfc3284bdf2b640"} Jan 21 11:27:22 crc kubenswrapper[4824]: I0121 11:27:22.338780 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-6b9b444cf4-q85qq"] Jan 21 11:27:22 crc kubenswrapper[4824]: I0121 11:27:22.344899 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-6b9b444cf4-q85qq"] Jan 21 11:27:22 crc kubenswrapper[4824]: I0121 11:27:22.349461 4824 scope.go:117] "RemoveContainer" containerID="41e3cbe62a3cfe1f1349c3e05b9d5169c6ca9dd78e1f915854df40b91a504d55" Jan 21 11:27:22 crc kubenswrapper[4824]: I0121 11:27:22.846247 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Jan 21 11:27:22 crc kubenswrapper[4824]: I0121 11:27:22.846442 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="46f0fa4b-479a-42c5-b9b2-3f30a079be48" containerName="glance-log" containerID="cri-o://953a68ff3f5b22fd88f3a1a2babc253f6f9d0648a9b5bca14c07b85059d64e1a" gracePeriod=30 Jan 21 11:27:22 crc kubenswrapper[4824]: I0121 11:27:22.846506 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="46f0fa4b-479a-42c5-b9b2-3f30a079be48" containerName="glance-httpd" containerID="cri-o://77d449bfcfde658bccf9c84cc9cff76978535cecd3aaeeb5e306a982392e5cf4" gracePeriod=30 Jan 21 11:27:23 crc kubenswrapper[4824]: I0121 11:27:23.330417 4824 generic.go:334] "Generic (PLEG): container finished" podID="46f0fa4b-479a-42c5-b9b2-3f30a079be48" containerID="953a68ff3f5b22fd88f3a1a2babc253f6f9d0648a9b5bca14c07b85059d64e1a" exitCode=143 Jan 21 11:27:23 crc kubenswrapper[4824]: I0121 11:27:23.330473 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"46f0fa4b-479a-42c5-b9b2-3f30a079be48","Type":"ContainerDied","Data":"953a68ff3f5b22fd88f3a1a2babc253f6f9d0648a9b5bca14c07b85059d64e1a"} Jan 21 11:27:23 crc kubenswrapper[4824]: I0121 11:27:23.370256 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Jan 21 11:27:23 crc kubenswrapper[4824]: I0121 11:27:23.370593 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="bae23da5-948a-4883-b88d-d6ed5d6ccba7" containerName="ceilometer-central-agent" containerID="cri-o://7b2e891014d0053a1d16ec02f1fc280a3a0cbb96983be7316f8030e04e57c2be" gracePeriod=30 Jan 21 11:27:23 crc kubenswrapper[4824]: I0121 11:27:23.370645 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="bae23da5-948a-4883-b88d-d6ed5d6ccba7" containerName="sg-core" containerID="cri-o://1568d27a18b9471d6e4be5cfb1adad4950254a106a787756913dd9ddb73cf5f9" gracePeriod=30 Jan 21 11:27:23 crc kubenswrapper[4824]: I0121 11:27:23.370670 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="bae23da5-948a-4883-b88d-d6ed5d6ccba7" containerName="ceilometer-notification-agent" containerID="cri-o://3b7c3d7bba0374d53d0ac03b9cb2de415ba41cc2525693fb788df9d2d6551fe4" gracePeriod=30 Jan 21 11:27:23 crc kubenswrapper[4824]: I0121 11:27:23.370702 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="bae23da5-948a-4883-b88d-d6ed5d6ccba7" containerName="proxy-httpd" containerID="cri-o://8eb8e5c1518e1715f46644f48ab6d186888980058e318f472bef1f90eb8e3c37" gracePeriod=30 Jan 21 11:27:23 crc kubenswrapper[4824]: I0121 11:27:23.374676 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="bae23da5-948a-4883-b88d-d6ed5d6ccba7" containerName="proxy-httpd" probeResult="failure" output="Get \"http://10.217.0.156:3000/\": EOF" Jan 21 11:27:24 crc kubenswrapper[4824]: I0121 11:27:24.062717 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4ed8e1b1-57c3-4b01-bbd7-9dc5125f4f94" path="/var/lib/kubelet/pods/4ed8e1b1-57c3-4b01-bbd7-9dc5125f4f94/volumes" Jan 21 11:27:24 crc kubenswrapper[4824]: I0121 11:27:24.342654 4824 generic.go:334] "Generic (PLEG): container finished" podID="bae23da5-948a-4883-b88d-d6ed5d6ccba7" containerID="8eb8e5c1518e1715f46644f48ab6d186888980058e318f472bef1f90eb8e3c37" exitCode=0 Jan 21 11:27:24 crc kubenswrapper[4824]: I0121 11:27:24.342684 4824 generic.go:334] "Generic (PLEG): container finished" podID="bae23da5-948a-4883-b88d-d6ed5d6ccba7" containerID="1568d27a18b9471d6e4be5cfb1adad4950254a106a787756913dd9ddb73cf5f9" exitCode=2 Jan 21 11:27:24 crc kubenswrapper[4824]: I0121 11:27:24.342691 4824 generic.go:334] "Generic (PLEG): container finished" podID="bae23da5-948a-4883-b88d-d6ed5d6ccba7" containerID="7b2e891014d0053a1d16ec02f1fc280a3a0cbb96983be7316f8030e04e57c2be" exitCode=0 Jan 21 11:27:24 crc kubenswrapper[4824]: I0121 11:27:24.342711 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bae23da5-948a-4883-b88d-d6ed5d6ccba7","Type":"ContainerDied","Data":"8eb8e5c1518e1715f46644f48ab6d186888980058e318f472bef1f90eb8e3c37"} Jan 21 11:27:24 crc kubenswrapper[4824]: I0121 11:27:24.342733 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bae23da5-948a-4883-b88d-d6ed5d6ccba7","Type":"ContainerDied","Data":"1568d27a18b9471d6e4be5cfb1adad4950254a106a787756913dd9ddb73cf5f9"} Jan 21 11:27:24 crc kubenswrapper[4824]: I0121 11:27:24.342754 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bae23da5-948a-4883-b88d-d6ed5d6ccba7","Type":"ContainerDied","Data":"7b2e891014d0053a1d16ec02f1fc280a3a0cbb96983be7316f8030e04e57c2be"} Jan 21 11:27:24 crc kubenswrapper[4824]: I0121 11:27:24.753398 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Jan 21 11:27:24 crc kubenswrapper[4824]: I0121 11:27:24.967612 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-4h8mx"] Jan 21 11:27:24 crc kubenswrapper[4824]: E0121 11:27:24.969269 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ed8e1b1-57c3-4b01-bbd7-9dc5125f4f94" containerName="barbican-api-log" Jan 21 11:27:24 crc kubenswrapper[4824]: I0121 11:27:24.969294 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ed8e1b1-57c3-4b01-bbd7-9dc5125f4f94" containerName="barbican-api-log" Jan 21 11:27:24 crc kubenswrapper[4824]: E0121 11:27:24.969330 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ed8e1b1-57c3-4b01-bbd7-9dc5125f4f94" containerName="barbican-api" Jan 21 11:27:24 crc kubenswrapper[4824]: I0121 11:27:24.969336 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ed8e1b1-57c3-4b01-bbd7-9dc5125f4f94" containerName="barbican-api" Jan 21 11:27:24 crc kubenswrapper[4824]: I0121 11:27:24.969504 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="4ed8e1b1-57c3-4b01-bbd7-9dc5125f4f94" containerName="barbican-api-log" Jan 21 11:27:24 crc kubenswrapper[4824]: I0121 11:27:24.969518 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="4ed8e1b1-57c3-4b01-bbd7-9dc5125f4f94" containerName="barbican-api" Jan 21 11:27:24 crc kubenswrapper[4824]: I0121 11:27:24.970051 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-4h8mx" Jan 21 11:27:24 crc kubenswrapper[4824]: I0121 11:27:24.986671 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-4h8mx"] Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.013344 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6mwhf\" (UniqueName: \"kubernetes.io/projected/91261d07-56c8-4da3-8dd5-6f23587d77be-kube-api-access-6mwhf\") pod \"nova-api-db-create-4h8mx\" (UID: \"91261d07-56c8-4da3-8dd5-6f23587d77be\") " pod="openstack/nova-api-db-create-4h8mx" Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.013560 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/91261d07-56c8-4da3-8dd5-6f23587d77be-operator-scripts\") pod \"nova-api-db-create-4h8mx\" (UID: \"91261d07-56c8-4da3-8dd5-6f23587d77be\") " pod="openstack/nova-api-db-create-4h8mx" Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.071721 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-f2xfg"] Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.075163 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-f2xfg" Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.079940 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-f2xfg"] Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.089860 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-91fe-account-create-update-4xgv2"] Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.090657 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-91fe-account-create-update-4xgv2" Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.092201 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.100214 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-91fe-account-create-update-4xgv2"] Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.115296 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6mwhf\" (UniqueName: \"kubernetes.io/projected/91261d07-56c8-4da3-8dd5-6f23587d77be-kube-api-access-6mwhf\") pod \"nova-api-db-create-4h8mx\" (UID: \"91261d07-56c8-4da3-8dd5-6f23587d77be\") " pod="openstack/nova-api-db-create-4h8mx" Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.115369 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/60da7490-4aa1-4880-bfcb-e51a4ed99f91-operator-scripts\") pod \"nova-api-91fe-account-create-update-4xgv2\" (UID: \"60da7490-4aa1-4880-bfcb-e51a4ed99f91\") " pod="openstack/nova-api-91fe-account-create-update-4xgv2" Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.115433 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/91261d07-56c8-4da3-8dd5-6f23587d77be-operator-scripts\") pod \"nova-api-db-create-4h8mx\" (UID: \"91261d07-56c8-4da3-8dd5-6f23587d77be\") " pod="openstack/nova-api-db-create-4h8mx" Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.115656 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tgm2l\" (UniqueName: \"kubernetes.io/projected/60da7490-4aa1-4880-bfcb-e51a4ed99f91-kube-api-access-tgm2l\") pod \"nova-api-91fe-account-create-update-4xgv2\" (UID: \"60da7490-4aa1-4880-bfcb-e51a4ed99f91\") " pod="openstack/nova-api-91fe-account-create-update-4xgv2" Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.115794 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/12e7657f-78fe-4b5c-b614-4449e94366b9-operator-scripts\") pod \"nova-cell0-db-create-f2xfg\" (UID: \"12e7657f-78fe-4b5c-b614-4449e94366b9\") " pod="openstack/nova-cell0-db-create-f2xfg" Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.115829 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cpkmt\" (UniqueName: \"kubernetes.io/projected/12e7657f-78fe-4b5c-b614-4449e94366b9-kube-api-access-cpkmt\") pod \"nova-cell0-db-create-f2xfg\" (UID: \"12e7657f-78fe-4b5c-b614-4449e94366b9\") " pod="openstack/nova-cell0-db-create-f2xfg" Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.116122 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/91261d07-56c8-4da3-8dd5-6f23587d77be-operator-scripts\") pod \"nova-api-db-create-4h8mx\" (UID: \"91261d07-56c8-4da3-8dd5-6f23587d77be\") " pod="openstack/nova-api-db-create-4h8mx" Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.130529 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6mwhf\" (UniqueName: \"kubernetes.io/projected/91261d07-56c8-4da3-8dd5-6f23587d77be-kube-api-access-6mwhf\") pod \"nova-api-db-create-4h8mx\" (UID: \"91261d07-56c8-4da3-8dd5-6f23587d77be\") " pod="openstack/nova-api-db-create-4h8mx" Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.217335 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/60da7490-4aa1-4880-bfcb-e51a4ed99f91-operator-scripts\") pod \"nova-api-91fe-account-create-update-4xgv2\" (UID: \"60da7490-4aa1-4880-bfcb-e51a4ed99f91\") " pod="openstack/nova-api-91fe-account-create-update-4xgv2" Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.217448 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tgm2l\" (UniqueName: \"kubernetes.io/projected/60da7490-4aa1-4880-bfcb-e51a4ed99f91-kube-api-access-tgm2l\") pod \"nova-api-91fe-account-create-update-4xgv2\" (UID: \"60da7490-4aa1-4880-bfcb-e51a4ed99f91\") " pod="openstack/nova-api-91fe-account-create-update-4xgv2" Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.217513 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/12e7657f-78fe-4b5c-b614-4449e94366b9-operator-scripts\") pod \"nova-cell0-db-create-f2xfg\" (UID: \"12e7657f-78fe-4b5c-b614-4449e94366b9\") " pod="openstack/nova-cell0-db-create-f2xfg" Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.217537 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cpkmt\" (UniqueName: \"kubernetes.io/projected/12e7657f-78fe-4b5c-b614-4449e94366b9-kube-api-access-cpkmt\") pod \"nova-cell0-db-create-f2xfg\" (UID: \"12e7657f-78fe-4b5c-b614-4449e94366b9\") " pod="openstack/nova-cell0-db-create-f2xfg" Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.218301 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/12e7657f-78fe-4b5c-b614-4449e94366b9-operator-scripts\") pod \"nova-cell0-db-create-f2xfg\" (UID: \"12e7657f-78fe-4b5c-b614-4449e94366b9\") " pod="openstack/nova-cell0-db-create-f2xfg" Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.219772 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/60da7490-4aa1-4880-bfcb-e51a4ed99f91-operator-scripts\") pod \"nova-api-91fe-account-create-update-4xgv2\" (UID: \"60da7490-4aa1-4880-bfcb-e51a4ed99f91\") " pod="openstack/nova-api-91fe-account-create-update-4xgv2" Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.235382 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tgm2l\" (UniqueName: \"kubernetes.io/projected/60da7490-4aa1-4880-bfcb-e51a4ed99f91-kube-api-access-tgm2l\") pod \"nova-api-91fe-account-create-update-4xgv2\" (UID: \"60da7490-4aa1-4880-bfcb-e51a4ed99f91\") " pod="openstack/nova-api-91fe-account-create-update-4xgv2" Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.238668 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cpkmt\" (UniqueName: \"kubernetes.io/projected/12e7657f-78fe-4b5c-b614-4449e94366b9-kube-api-access-cpkmt\") pod \"nova-cell0-db-create-f2xfg\" (UID: \"12e7657f-78fe-4b5c-b614-4449e94366b9\") " pod="openstack/nova-cell0-db-create-f2xfg" Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.275143 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-dhvfz"] Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.283882 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-4h8mx" Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.305235 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-dhvfz"] Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.305284 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-90fb-account-create-update-m4n4l"] Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.305372 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-dhvfz" Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.306691 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-90fb-account-create-update-m4n4l" Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.308426 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.320063 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b24eefe0-9cb1-4672-85d8-6dfe2babebff-operator-scripts\") pod \"nova-cell0-90fb-account-create-update-m4n4l\" (UID: \"b24eefe0-9cb1-4672-85d8-6dfe2babebff\") " pod="openstack/nova-cell0-90fb-account-create-update-m4n4l" Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.320146 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d5162c70-9359-42d5-a02d-16140ac4e366-operator-scripts\") pod \"nova-cell1-db-create-dhvfz\" (UID: \"d5162c70-9359-42d5-a02d-16140ac4e366\") " pod="openstack/nova-cell1-db-create-dhvfz" Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.322935 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-90fb-account-create-update-m4n4l"] Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.335009 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-btjhr\" (UniqueName: \"kubernetes.io/projected/d5162c70-9359-42d5-a02d-16140ac4e366-kube-api-access-btjhr\") pod \"nova-cell1-db-create-dhvfz\" (UID: \"d5162c70-9359-42d5-a02d-16140ac4e366\") " pod="openstack/nova-cell1-db-create-dhvfz" Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.335115 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lzjdm\" (UniqueName: \"kubernetes.io/projected/b24eefe0-9cb1-4672-85d8-6dfe2babebff-kube-api-access-lzjdm\") pod \"nova-cell0-90fb-account-create-update-m4n4l\" (UID: \"b24eefe0-9cb1-4672-85d8-6dfe2babebff\") " pod="openstack/nova-cell0-90fb-account-create-update-m4n4l" Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.387241 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-f2xfg" Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.408508 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-91fe-account-create-update-4xgv2" Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.436673 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b24eefe0-9cb1-4672-85d8-6dfe2babebff-operator-scripts\") pod \"nova-cell0-90fb-account-create-update-m4n4l\" (UID: \"b24eefe0-9cb1-4672-85d8-6dfe2babebff\") " pod="openstack/nova-cell0-90fb-account-create-update-m4n4l" Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.436749 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d5162c70-9359-42d5-a02d-16140ac4e366-operator-scripts\") pod \"nova-cell1-db-create-dhvfz\" (UID: \"d5162c70-9359-42d5-a02d-16140ac4e366\") " pod="openstack/nova-cell1-db-create-dhvfz" Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.436876 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-btjhr\" (UniqueName: \"kubernetes.io/projected/d5162c70-9359-42d5-a02d-16140ac4e366-kube-api-access-btjhr\") pod \"nova-cell1-db-create-dhvfz\" (UID: \"d5162c70-9359-42d5-a02d-16140ac4e366\") " pod="openstack/nova-cell1-db-create-dhvfz" Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.436971 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lzjdm\" (UniqueName: \"kubernetes.io/projected/b24eefe0-9cb1-4672-85d8-6dfe2babebff-kube-api-access-lzjdm\") pod \"nova-cell0-90fb-account-create-update-m4n4l\" (UID: \"b24eefe0-9cb1-4672-85d8-6dfe2babebff\") " pod="openstack/nova-cell0-90fb-account-create-update-m4n4l" Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.437571 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b24eefe0-9cb1-4672-85d8-6dfe2babebff-operator-scripts\") pod \"nova-cell0-90fb-account-create-update-m4n4l\" (UID: \"b24eefe0-9cb1-4672-85d8-6dfe2babebff\") " pod="openstack/nova-cell0-90fb-account-create-update-m4n4l" Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.437623 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d5162c70-9359-42d5-a02d-16140ac4e366-operator-scripts\") pod \"nova-cell1-db-create-dhvfz\" (UID: \"d5162c70-9359-42d5-a02d-16140ac4e366\") " pod="openstack/nova-cell1-db-create-dhvfz" Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.472657 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lzjdm\" (UniqueName: \"kubernetes.io/projected/b24eefe0-9cb1-4672-85d8-6dfe2babebff-kube-api-access-lzjdm\") pod \"nova-cell0-90fb-account-create-update-m4n4l\" (UID: \"b24eefe0-9cb1-4672-85d8-6dfe2babebff\") " pod="openstack/nova-cell0-90fb-account-create-update-m4n4l" Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.480617 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-btjhr\" (UniqueName: \"kubernetes.io/projected/d5162c70-9359-42d5-a02d-16140ac4e366-kube-api-access-btjhr\") pod \"nova-cell1-db-create-dhvfz\" (UID: \"d5162c70-9359-42d5-a02d-16140ac4e366\") " pod="openstack/nova-cell1-db-create-dhvfz" Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.485537 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-16cb-account-create-update-qjbx7"] Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.488608 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-16cb-account-create-update-qjbx7" Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.490254 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.494805 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-16cb-account-create-update-qjbx7"] Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.538807 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-frs7l\" (UniqueName: \"kubernetes.io/projected/34afc8af-6ce5-4ee0-917b-326bd60d6237-kube-api-access-frs7l\") pod \"nova-cell1-16cb-account-create-update-qjbx7\" (UID: \"34afc8af-6ce5-4ee0-917b-326bd60d6237\") " pod="openstack/nova-cell1-16cb-account-create-update-qjbx7" Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.539056 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/34afc8af-6ce5-4ee0-917b-326bd60d6237-operator-scripts\") pod \"nova-cell1-16cb-account-create-update-qjbx7\" (UID: \"34afc8af-6ce5-4ee0-917b-326bd60d6237\") " pod="openstack/nova-cell1-16cb-account-create-update-qjbx7" Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.622412 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-dhvfz" Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.628920 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-90fb-account-create-update-m4n4l" Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.632758 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-79f4c948b7-4hz4p"] Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.633908 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-79f4c948b7-4hz4p" Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.637302 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.637340 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.637528 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.640777 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/34afc8af-6ce5-4ee0-917b-326bd60d6237-operator-scripts\") pod \"nova-cell1-16cb-account-create-update-qjbx7\" (UID: \"34afc8af-6ce5-4ee0-917b-326bd60d6237\") " pod="openstack/nova-cell1-16cb-account-create-update-qjbx7" Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.640881 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-frs7l\" (UniqueName: \"kubernetes.io/projected/34afc8af-6ce5-4ee0-917b-326bd60d6237-kube-api-access-frs7l\") pod \"nova-cell1-16cb-account-create-update-qjbx7\" (UID: \"34afc8af-6ce5-4ee0-917b-326bd60d6237\") " pod="openstack/nova-cell1-16cb-account-create-update-qjbx7" Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.641388 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/34afc8af-6ce5-4ee0-917b-326bd60d6237-operator-scripts\") pod \"nova-cell1-16cb-account-create-update-qjbx7\" (UID: \"34afc8af-6ce5-4ee0-917b-326bd60d6237\") " pod="openstack/nova-cell1-16cb-account-create-update-qjbx7" Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.643230 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-79f4c948b7-4hz4p"] Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.661734 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-frs7l\" (UniqueName: \"kubernetes.io/projected/34afc8af-6ce5-4ee0-917b-326bd60d6237-kube-api-access-frs7l\") pod \"nova-cell1-16cb-account-create-update-qjbx7\" (UID: \"34afc8af-6ce5-4ee0-917b-326bd60d6237\") " pod="openstack/nova-cell1-16cb-account-create-update-qjbx7" Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.741772 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tjr4p\" (UniqueName: \"kubernetes.io/projected/01d4ce4d-3728-4f4c-9fbf-ce58793b6b7d-kube-api-access-tjr4p\") pod \"swift-proxy-79f4c948b7-4hz4p\" (UID: \"01d4ce4d-3728-4f4c-9fbf-ce58793b6b7d\") " pod="openstack/swift-proxy-79f4c948b7-4hz4p" Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.741988 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/01d4ce4d-3728-4f4c-9fbf-ce58793b6b7d-public-tls-certs\") pod \"swift-proxy-79f4c948b7-4hz4p\" (UID: \"01d4ce4d-3728-4f4c-9fbf-ce58793b6b7d\") " pod="openstack/swift-proxy-79f4c948b7-4hz4p" Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.742085 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/01d4ce4d-3728-4f4c-9fbf-ce58793b6b7d-internal-tls-certs\") pod \"swift-proxy-79f4c948b7-4hz4p\" (UID: \"01d4ce4d-3728-4f4c-9fbf-ce58793b6b7d\") " pod="openstack/swift-proxy-79f4c948b7-4hz4p" Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.742126 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/01d4ce4d-3728-4f4c-9fbf-ce58793b6b7d-config-data\") pod \"swift-proxy-79f4c948b7-4hz4p\" (UID: \"01d4ce4d-3728-4f4c-9fbf-ce58793b6b7d\") " pod="openstack/swift-proxy-79f4c948b7-4hz4p" Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.742292 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01d4ce4d-3728-4f4c-9fbf-ce58793b6b7d-combined-ca-bundle\") pod \"swift-proxy-79f4c948b7-4hz4p\" (UID: \"01d4ce4d-3728-4f4c-9fbf-ce58793b6b7d\") " pod="openstack/swift-proxy-79f4c948b7-4hz4p" Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.742351 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/01d4ce4d-3728-4f4c-9fbf-ce58793b6b7d-run-httpd\") pod \"swift-proxy-79f4c948b7-4hz4p\" (UID: \"01d4ce4d-3728-4f4c-9fbf-ce58793b6b7d\") " pod="openstack/swift-proxy-79f4c948b7-4hz4p" Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.742382 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/01d4ce4d-3728-4f4c-9fbf-ce58793b6b7d-etc-swift\") pod \"swift-proxy-79f4c948b7-4hz4p\" (UID: \"01d4ce4d-3728-4f4c-9fbf-ce58793b6b7d\") " pod="openstack/swift-proxy-79f4c948b7-4hz4p" Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.742400 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/01d4ce4d-3728-4f4c-9fbf-ce58793b6b7d-log-httpd\") pod \"swift-proxy-79f4c948b7-4hz4p\" (UID: \"01d4ce4d-3728-4f4c-9fbf-ce58793b6b7d\") " pod="openstack/swift-proxy-79f4c948b7-4hz4p" Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.839590 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-16cb-account-create-update-qjbx7" Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.843761 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/01d4ce4d-3728-4f4c-9fbf-ce58793b6b7d-internal-tls-certs\") pod \"swift-proxy-79f4c948b7-4hz4p\" (UID: \"01d4ce4d-3728-4f4c-9fbf-ce58793b6b7d\") " pod="openstack/swift-proxy-79f4c948b7-4hz4p" Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.843804 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/01d4ce4d-3728-4f4c-9fbf-ce58793b6b7d-config-data\") pod \"swift-proxy-79f4c948b7-4hz4p\" (UID: \"01d4ce4d-3728-4f4c-9fbf-ce58793b6b7d\") " pod="openstack/swift-proxy-79f4c948b7-4hz4p" Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.843845 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01d4ce4d-3728-4f4c-9fbf-ce58793b6b7d-combined-ca-bundle\") pod \"swift-proxy-79f4c948b7-4hz4p\" (UID: \"01d4ce4d-3728-4f4c-9fbf-ce58793b6b7d\") " pod="openstack/swift-proxy-79f4c948b7-4hz4p" Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.843862 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/01d4ce4d-3728-4f4c-9fbf-ce58793b6b7d-run-httpd\") pod \"swift-proxy-79f4c948b7-4hz4p\" (UID: \"01d4ce4d-3728-4f4c-9fbf-ce58793b6b7d\") " pod="openstack/swift-proxy-79f4c948b7-4hz4p" Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.843878 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/01d4ce4d-3728-4f4c-9fbf-ce58793b6b7d-etc-swift\") pod \"swift-proxy-79f4c948b7-4hz4p\" (UID: \"01d4ce4d-3728-4f4c-9fbf-ce58793b6b7d\") " pod="openstack/swift-proxy-79f4c948b7-4hz4p" Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.843893 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/01d4ce4d-3728-4f4c-9fbf-ce58793b6b7d-log-httpd\") pod \"swift-proxy-79f4c948b7-4hz4p\" (UID: \"01d4ce4d-3728-4f4c-9fbf-ce58793b6b7d\") " pod="openstack/swift-proxy-79f4c948b7-4hz4p" Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.843917 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tjr4p\" (UniqueName: \"kubernetes.io/projected/01d4ce4d-3728-4f4c-9fbf-ce58793b6b7d-kube-api-access-tjr4p\") pod \"swift-proxy-79f4c948b7-4hz4p\" (UID: \"01d4ce4d-3728-4f4c-9fbf-ce58793b6b7d\") " pod="openstack/swift-proxy-79f4c948b7-4hz4p" Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.843938 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/01d4ce4d-3728-4f4c-9fbf-ce58793b6b7d-public-tls-certs\") pod \"swift-proxy-79f4c948b7-4hz4p\" (UID: \"01d4ce4d-3728-4f4c-9fbf-ce58793b6b7d\") " pod="openstack/swift-proxy-79f4c948b7-4hz4p" Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.844673 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/01d4ce4d-3728-4f4c-9fbf-ce58793b6b7d-run-httpd\") pod \"swift-proxy-79f4c948b7-4hz4p\" (UID: \"01d4ce4d-3728-4f4c-9fbf-ce58793b6b7d\") " pod="openstack/swift-proxy-79f4c948b7-4hz4p" Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.846174 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/01d4ce4d-3728-4f4c-9fbf-ce58793b6b7d-log-httpd\") pod \"swift-proxy-79f4c948b7-4hz4p\" (UID: \"01d4ce4d-3728-4f4c-9fbf-ce58793b6b7d\") " pod="openstack/swift-proxy-79f4c948b7-4hz4p" Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.847383 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01d4ce4d-3728-4f4c-9fbf-ce58793b6b7d-combined-ca-bundle\") pod \"swift-proxy-79f4c948b7-4hz4p\" (UID: \"01d4ce4d-3728-4f4c-9fbf-ce58793b6b7d\") " pod="openstack/swift-proxy-79f4c948b7-4hz4p" Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.847409 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/01d4ce4d-3728-4f4c-9fbf-ce58793b6b7d-public-tls-certs\") pod \"swift-proxy-79f4c948b7-4hz4p\" (UID: \"01d4ce4d-3728-4f4c-9fbf-ce58793b6b7d\") " pod="openstack/swift-proxy-79f4c948b7-4hz4p" Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.848548 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/01d4ce4d-3728-4f4c-9fbf-ce58793b6b7d-internal-tls-certs\") pod \"swift-proxy-79f4c948b7-4hz4p\" (UID: \"01d4ce4d-3728-4f4c-9fbf-ce58793b6b7d\") " pod="openstack/swift-proxy-79f4c948b7-4hz4p" Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.850352 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/01d4ce4d-3728-4f4c-9fbf-ce58793b6b7d-config-data\") pod \"swift-proxy-79f4c948b7-4hz4p\" (UID: \"01d4ce4d-3728-4f4c-9fbf-ce58793b6b7d\") " pod="openstack/swift-proxy-79f4c948b7-4hz4p" Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.850437 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/01d4ce4d-3728-4f4c-9fbf-ce58793b6b7d-etc-swift\") pod \"swift-proxy-79f4c948b7-4hz4p\" (UID: \"01d4ce4d-3728-4f4c-9fbf-ce58793b6b7d\") " pod="openstack/swift-proxy-79f4c948b7-4hz4p" Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.860045 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tjr4p\" (UniqueName: \"kubernetes.io/projected/01d4ce4d-3728-4f4c-9fbf-ce58793b6b7d-kube-api-access-tjr4p\") pod \"swift-proxy-79f4c948b7-4hz4p\" (UID: \"01d4ce4d-3728-4f4c-9fbf-ce58793b6b7d\") " pod="openstack/swift-proxy-79f4c948b7-4hz4p" Jan 21 11:27:25 crc kubenswrapper[4824]: I0121 11:27:25.963085 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-79f4c948b7-4hz4p" Jan 21 11:27:26 crc kubenswrapper[4824]: I0121 11:27:26.359598 4824 generic.go:334] "Generic (PLEG): container finished" podID="bae23da5-948a-4883-b88d-d6ed5d6ccba7" containerID="3b7c3d7bba0374d53d0ac03b9cb2de415ba41cc2525693fb788df9d2d6551fe4" exitCode=0 Jan 21 11:27:26 crc kubenswrapper[4824]: I0121 11:27:26.359671 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bae23da5-948a-4883-b88d-d6ed5d6ccba7","Type":"ContainerDied","Data":"3b7c3d7bba0374d53d0ac03b9cb2de415ba41cc2525693fb788df9d2d6551fe4"} Jan 21 11:27:26 crc kubenswrapper[4824]: I0121 11:27:26.362264 4824 generic.go:334] "Generic (PLEG): container finished" podID="46f0fa4b-479a-42c5-b9b2-3f30a079be48" containerID="77d449bfcfde658bccf9c84cc9cff76978535cecd3aaeeb5e306a982392e5cf4" exitCode=0 Jan 21 11:27:26 crc kubenswrapper[4824]: I0121 11:27:26.362292 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"46f0fa4b-479a-42c5-b9b2-3f30a079be48","Type":"ContainerDied","Data":"77d449bfcfde658bccf9c84cc9cff76978535cecd3aaeeb5e306a982392e5cf4"} Jan 21 11:27:28 crc kubenswrapper[4824]: I0121 11:27:28.704410 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 21 11:27:28 crc kubenswrapper[4824]: I0121 11:27:28.806982 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bae23da5-948a-4883-b88d-d6ed5d6ccba7-scripts\") pod \"bae23da5-948a-4883-b88d-d6ed5d6ccba7\" (UID: \"bae23da5-948a-4883-b88d-d6ed5d6ccba7\") " Jan 21 11:27:28 crc kubenswrapper[4824]: I0121 11:27:28.807103 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bae23da5-948a-4883-b88d-d6ed5d6ccba7-config-data\") pod \"bae23da5-948a-4883-b88d-d6ed5d6ccba7\" (UID: \"bae23da5-948a-4883-b88d-d6ed5d6ccba7\") " Jan 21 11:27:28 crc kubenswrapper[4824]: I0121 11:27:28.807128 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bae23da5-948a-4883-b88d-d6ed5d6ccba7-log-httpd\") pod \"bae23da5-948a-4883-b88d-d6ed5d6ccba7\" (UID: \"bae23da5-948a-4883-b88d-d6ed5d6ccba7\") " Jan 21 11:27:28 crc kubenswrapper[4824]: I0121 11:27:28.807159 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bae23da5-948a-4883-b88d-d6ed5d6ccba7-sg-core-conf-yaml\") pod \"bae23da5-948a-4883-b88d-d6ed5d6ccba7\" (UID: \"bae23da5-948a-4883-b88d-d6ed5d6ccba7\") " Jan 21 11:27:28 crc kubenswrapper[4824]: I0121 11:27:28.807219 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nmjb9\" (UniqueName: \"kubernetes.io/projected/bae23da5-948a-4883-b88d-d6ed5d6ccba7-kube-api-access-nmjb9\") pod \"bae23da5-948a-4883-b88d-d6ed5d6ccba7\" (UID: \"bae23da5-948a-4883-b88d-d6ed5d6ccba7\") " Jan 21 11:27:28 crc kubenswrapper[4824]: I0121 11:27:28.807240 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bae23da5-948a-4883-b88d-d6ed5d6ccba7-combined-ca-bundle\") pod \"bae23da5-948a-4883-b88d-d6ed5d6ccba7\" (UID: \"bae23da5-948a-4883-b88d-d6ed5d6ccba7\") " Jan 21 11:27:28 crc kubenswrapper[4824]: I0121 11:27:28.807291 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bae23da5-948a-4883-b88d-d6ed5d6ccba7-run-httpd\") pod \"bae23da5-948a-4883-b88d-d6ed5d6ccba7\" (UID: \"bae23da5-948a-4883-b88d-d6ed5d6ccba7\") " Jan 21 11:27:28 crc kubenswrapper[4824]: I0121 11:27:28.808073 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bae23da5-948a-4883-b88d-d6ed5d6ccba7-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "bae23da5-948a-4883-b88d-d6ed5d6ccba7" (UID: "bae23da5-948a-4883-b88d-d6ed5d6ccba7"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:27:28 crc kubenswrapper[4824]: I0121 11:27:28.811424 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bae23da5-948a-4883-b88d-d6ed5d6ccba7-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "bae23da5-948a-4883-b88d-d6ed5d6ccba7" (UID: "bae23da5-948a-4883-b88d-d6ed5d6ccba7"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:27:28 crc kubenswrapper[4824]: I0121 11:27:28.812762 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bae23da5-948a-4883-b88d-d6ed5d6ccba7-scripts" (OuterVolumeSpecName: "scripts") pod "bae23da5-948a-4883-b88d-d6ed5d6ccba7" (UID: "bae23da5-948a-4883-b88d-d6ed5d6ccba7"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:27:28 crc kubenswrapper[4824]: I0121 11:27:28.815628 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bae23da5-948a-4883-b88d-d6ed5d6ccba7-kube-api-access-nmjb9" (OuterVolumeSpecName: "kube-api-access-nmjb9") pod "bae23da5-948a-4883-b88d-d6ed5d6ccba7" (UID: "bae23da5-948a-4883-b88d-d6ed5d6ccba7"). InnerVolumeSpecName "kube-api-access-nmjb9". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:27:28 crc kubenswrapper[4824]: I0121 11:27:28.817884 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Jan 21 11:27:28 crc kubenswrapper[4824]: I0121 11:27:28.839268 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bae23da5-948a-4883-b88d-d6ed5d6ccba7-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "bae23da5-948a-4883-b88d-d6ed5d6ccba7" (UID: "bae23da5-948a-4883-b88d-d6ed5d6ccba7"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:27:28 crc kubenswrapper[4824]: I0121 11:27:28.882941 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bae23da5-948a-4883-b88d-d6ed5d6ccba7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bae23da5-948a-4883-b88d-d6ed5d6ccba7" (UID: "bae23da5-948a-4883-b88d-d6ed5d6ccba7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:27:28 crc kubenswrapper[4824]: I0121 11:27:28.907991 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bae23da5-948a-4883-b88d-d6ed5d6ccba7-config-data" (OuterVolumeSpecName: "config-data") pod "bae23da5-948a-4883-b88d-d6ed5d6ccba7" (UID: "bae23da5-948a-4883-b88d-d6ed5d6ccba7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:27:28 crc kubenswrapper[4824]: I0121 11:27:28.908449 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/46f0fa4b-479a-42c5-b9b2-3f30a079be48-config-data\") pod \"46f0fa4b-479a-42c5-b9b2-3f30a079be48\" (UID: \"46f0fa4b-479a-42c5-b9b2-3f30a079be48\") " Jan 21 11:27:28 crc kubenswrapper[4824]: I0121 11:27:28.908520 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"46f0fa4b-479a-42c5-b9b2-3f30a079be48\" (UID: \"46f0fa4b-479a-42c5-b9b2-3f30a079be48\") " Jan 21 11:27:28 crc kubenswrapper[4824]: I0121 11:27:28.908580 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/46f0fa4b-479a-42c5-b9b2-3f30a079be48-internal-tls-certs\") pod \"46f0fa4b-479a-42c5-b9b2-3f30a079be48\" (UID: \"46f0fa4b-479a-42c5-b9b2-3f30a079be48\") " Jan 21 11:27:28 crc kubenswrapper[4824]: I0121 11:27:28.908685 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/46f0fa4b-479a-42c5-b9b2-3f30a079be48-logs\") pod \"46f0fa4b-479a-42c5-b9b2-3f30a079be48\" (UID: \"46f0fa4b-479a-42c5-b9b2-3f30a079be48\") " Jan 21 11:27:28 crc kubenswrapper[4824]: I0121 11:27:28.908706 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46f0fa4b-479a-42c5-b9b2-3f30a079be48-combined-ca-bundle\") pod \"46f0fa4b-479a-42c5-b9b2-3f30a079be48\" (UID: \"46f0fa4b-479a-42c5-b9b2-3f30a079be48\") " Jan 21 11:27:28 crc kubenswrapper[4824]: I0121 11:27:28.908742 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/46f0fa4b-479a-42c5-b9b2-3f30a079be48-httpd-run\") pod \"46f0fa4b-479a-42c5-b9b2-3f30a079be48\" (UID: \"46f0fa4b-479a-42c5-b9b2-3f30a079be48\") " Jan 21 11:27:28 crc kubenswrapper[4824]: I0121 11:27:28.908837 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/46f0fa4b-479a-42c5-b9b2-3f30a079be48-scripts\") pod \"46f0fa4b-479a-42c5-b9b2-3f30a079be48\" (UID: \"46f0fa4b-479a-42c5-b9b2-3f30a079be48\") " Jan 21 11:27:28 crc kubenswrapper[4824]: I0121 11:27:28.908863 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v8grh\" (UniqueName: \"kubernetes.io/projected/46f0fa4b-479a-42c5-b9b2-3f30a079be48-kube-api-access-v8grh\") pod \"46f0fa4b-479a-42c5-b9b2-3f30a079be48\" (UID: \"46f0fa4b-479a-42c5-b9b2-3f30a079be48\") " Jan 21 11:27:28 crc kubenswrapper[4824]: I0121 11:27:28.909377 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bae23da5-948a-4883-b88d-d6ed5d6ccba7-config-data\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:28 crc kubenswrapper[4824]: I0121 11:27:28.909397 4824 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bae23da5-948a-4883-b88d-d6ed5d6ccba7-log-httpd\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:28 crc kubenswrapper[4824]: I0121 11:27:28.909405 4824 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bae23da5-948a-4883-b88d-d6ed5d6ccba7-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:28 crc kubenswrapper[4824]: I0121 11:27:28.909416 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nmjb9\" (UniqueName: \"kubernetes.io/projected/bae23da5-948a-4883-b88d-d6ed5d6ccba7-kube-api-access-nmjb9\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:28 crc kubenswrapper[4824]: I0121 11:27:28.909425 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bae23da5-948a-4883-b88d-d6ed5d6ccba7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:28 crc kubenswrapper[4824]: I0121 11:27:28.909434 4824 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bae23da5-948a-4883-b88d-d6ed5d6ccba7-run-httpd\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:28 crc kubenswrapper[4824]: I0121 11:27:28.909442 4824 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bae23da5-948a-4883-b88d-d6ed5d6ccba7-scripts\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:28 crc kubenswrapper[4824]: I0121 11:27:28.909438 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/46f0fa4b-479a-42c5-b9b2-3f30a079be48-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "46f0fa4b-479a-42c5-b9b2-3f30a079be48" (UID: "46f0fa4b-479a-42c5-b9b2-3f30a079be48"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:27:28 crc kubenswrapper[4824]: I0121 11:27:28.909579 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/46f0fa4b-479a-42c5-b9b2-3f30a079be48-logs" (OuterVolumeSpecName: "logs") pod "46f0fa4b-479a-42c5-b9b2-3f30a079be48" (UID: "46f0fa4b-479a-42c5-b9b2-3f30a079be48"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:27:28 crc kubenswrapper[4824]: I0121 11:27:28.911340 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage07-crc" (OuterVolumeSpecName: "glance") pod "46f0fa4b-479a-42c5-b9b2-3f30a079be48" (UID: "46f0fa4b-479a-42c5-b9b2-3f30a079be48"). InnerVolumeSpecName "local-storage07-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Jan 21 11:27:28 crc kubenswrapper[4824]: I0121 11:27:28.911949 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/46f0fa4b-479a-42c5-b9b2-3f30a079be48-kube-api-access-v8grh" (OuterVolumeSpecName: "kube-api-access-v8grh") pod "46f0fa4b-479a-42c5-b9b2-3f30a079be48" (UID: "46f0fa4b-479a-42c5-b9b2-3f30a079be48"). InnerVolumeSpecName "kube-api-access-v8grh". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:27:28 crc kubenswrapper[4824]: I0121 11:27:28.912417 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/46f0fa4b-479a-42c5-b9b2-3f30a079be48-scripts" (OuterVolumeSpecName: "scripts") pod "46f0fa4b-479a-42c5-b9b2-3f30a079be48" (UID: "46f0fa4b-479a-42c5-b9b2-3f30a079be48"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:27:28 crc kubenswrapper[4824]: I0121 11:27:28.942044 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/46f0fa4b-479a-42c5-b9b2-3f30a079be48-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "46f0fa4b-479a-42c5-b9b2-3f30a079be48" (UID: "46f0fa4b-479a-42c5-b9b2-3f30a079be48"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:27:28 crc kubenswrapper[4824]: I0121 11:27:28.948787 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/46f0fa4b-479a-42c5-b9b2-3f30a079be48-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "46f0fa4b-479a-42c5-b9b2-3f30a079be48" (UID: "46f0fa4b-479a-42c5-b9b2-3f30a079be48"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:27:28 crc kubenswrapper[4824]: I0121 11:27:28.955184 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/46f0fa4b-479a-42c5-b9b2-3f30a079be48-config-data" (OuterVolumeSpecName: "config-data") pod "46f0fa4b-479a-42c5-b9b2-3f30a079be48" (UID: "46f0fa4b-479a-42c5-b9b2-3f30a079be48"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:27:29 crc kubenswrapper[4824]: I0121 11:27:29.024258 4824 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/46f0fa4b-479a-42c5-b9b2-3f30a079be48-httpd-run\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:29 crc kubenswrapper[4824]: I0121 11:27:29.024290 4824 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/46f0fa4b-479a-42c5-b9b2-3f30a079be48-scripts\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:29 crc kubenswrapper[4824]: I0121 11:27:29.024304 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v8grh\" (UniqueName: \"kubernetes.io/projected/46f0fa4b-479a-42c5-b9b2-3f30a079be48-kube-api-access-v8grh\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:29 crc kubenswrapper[4824]: I0121 11:27:29.024313 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/46f0fa4b-479a-42c5-b9b2-3f30a079be48-config-data\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:29 crc kubenswrapper[4824]: I0121 11:27:29.024339 4824 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" " Jan 21 11:27:29 crc kubenswrapper[4824]: I0121 11:27:29.024385 4824 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/46f0fa4b-479a-42c5-b9b2-3f30a079be48-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:29 crc kubenswrapper[4824]: I0121 11:27:29.024395 4824 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/46f0fa4b-479a-42c5-b9b2-3f30a079be48-logs\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:29 crc kubenswrapper[4824]: I0121 11:27:29.024404 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46f0fa4b-479a-42c5-b9b2-3f30a079be48-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:29 crc kubenswrapper[4824]: I0121 11:27:29.039798 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-16cb-account-create-update-qjbx7"] Jan 21 11:27:29 crc kubenswrapper[4824]: I0121 11:27:29.061348 4824 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage07-crc" (UniqueName: "kubernetes.io/local-volume/local-storage07-crc") on node "crc" Jan 21 11:27:29 crc kubenswrapper[4824]: I0121 11:27:29.065170 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-4h8mx"] Jan 21 11:27:29 crc kubenswrapper[4824]: I0121 11:27:29.126709 4824 reconciler_common.go:293] "Volume detached for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:29 crc kubenswrapper[4824]: I0121 11:27:29.198184 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-91fe-account-create-update-4xgv2"] Jan 21 11:27:29 crc kubenswrapper[4824]: I0121 11:27:29.205418 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-dhvfz"] Jan 21 11:27:29 crc kubenswrapper[4824]: W0121 11:27:29.213176 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb24eefe0_9cb1_4672_85d8_6dfe2babebff.slice/crio-5ca82bb30aa3c9ac045630af69b3a4a8697ebeccee0b68dc49d5dad6be111c3a WatchSource:0}: Error finding container 5ca82bb30aa3c9ac045630af69b3a4a8697ebeccee0b68dc49d5dad6be111c3a: Status 404 returned error can't find the container with id 5ca82bb30aa3c9ac045630af69b3a4a8697ebeccee0b68dc49d5dad6be111c3a Jan 21 11:27:29 crc kubenswrapper[4824]: I0121 11:27:29.217154 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-90fb-account-create-update-m4n4l"] Jan 21 11:27:29 crc kubenswrapper[4824]: I0121 11:27:29.361425 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-f2xfg"] Jan 21 11:27:29 crc kubenswrapper[4824]: I0121 11:27:29.387042 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-90fb-account-create-update-m4n4l" event={"ID":"b24eefe0-9cb1-4672-85d8-6dfe2babebff","Type":"ContainerStarted","Data":"7d5e936f6ab78e748abd4aabb374f95965e7eaaf526fd1a755b675eed79144ec"} Jan 21 11:27:29 crc kubenswrapper[4824]: I0121 11:27:29.387086 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-90fb-account-create-update-m4n4l" event={"ID":"b24eefe0-9cb1-4672-85d8-6dfe2babebff","Type":"ContainerStarted","Data":"5ca82bb30aa3c9ac045630af69b3a4a8697ebeccee0b68dc49d5dad6be111c3a"} Jan 21 11:27:29 crc kubenswrapper[4824]: I0121 11:27:29.390255 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bae23da5-948a-4883-b88d-d6ed5d6ccba7","Type":"ContainerDied","Data":"b8cc173acd912ad690d3e7aa92c24c20b537f220a873ef96f5148fc3b668366a"} Jan 21 11:27:29 crc kubenswrapper[4824]: I0121 11:27:29.390293 4824 scope.go:117] "RemoveContainer" containerID="8eb8e5c1518e1715f46644f48ab6d186888980058e318f472bef1f90eb8e3c37" Jan 21 11:27:29 crc kubenswrapper[4824]: I0121 11:27:29.390344 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 21 11:27:29 crc kubenswrapper[4824]: I0121 11:27:29.392181 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"1af94d08-6d1e-4a97-8f82-27a1755d17a1","Type":"ContainerStarted","Data":"f8d2b72296ca177ea36b5db87f12950c56e50668e1e64af2fed5ab52cefc2cc3"} Jan 21 11:27:29 crc kubenswrapper[4824]: I0121 11:27:29.395722 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-dhvfz" event={"ID":"d5162c70-9359-42d5-a02d-16140ac4e366","Type":"ContainerStarted","Data":"6fa0076da1636efc5cc0178f55ca211d393f930b6c63c3d85cac4eedd9f2d288"} Jan 21 11:27:29 crc kubenswrapper[4824]: I0121 11:27:29.395757 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-dhvfz" event={"ID":"d5162c70-9359-42d5-a02d-16140ac4e366","Type":"ContainerStarted","Data":"2c6b2c2ef954db67b738c424fe09e651c130b39cd2f9702c3ab0d17616f3d362"} Jan 21 11:27:29 crc kubenswrapper[4824]: I0121 11:27:29.401894 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-90fb-account-create-update-m4n4l" podStartSLOduration=4.401882188 podStartE2EDuration="4.401882188s" podCreationTimestamp="2026-01-21 11:27:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:27:29.399528649 +0000 UTC m=+991.692557940" watchObservedRunningTime="2026-01-21 11:27:29.401882188 +0000 UTC m=+991.694911481" Jan 21 11:27:29 crc kubenswrapper[4824]: I0121 11:27:29.401921 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-91fe-account-create-update-4xgv2" event={"ID":"60da7490-4aa1-4880-bfcb-e51a4ed99f91","Type":"ContainerStarted","Data":"a9bd46298904e9e7c4cccbdd1cecf51e20eaa4ef17b0e64f0ae86c7cca0ec79d"} Jan 21 11:27:29 crc kubenswrapper[4824]: I0121 11:27:29.401981 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-91fe-account-create-update-4xgv2" event={"ID":"60da7490-4aa1-4880-bfcb-e51a4ed99f91","Type":"ContainerStarted","Data":"72b068529352cd2fd0d98380f92cbe1ad287be50708214d7b3205789afcc691d"} Jan 21 11:27:29 crc kubenswrapper[4824]: I0121 11:27:29.405793 4824 generic.go:334] "Generic (PLEG): container finished" podID="91261d07-56c8-4da3-8dd5-6f23587d77be" containerID="06b41897897bde2d466371c85738035a8ab9ee6530d1dfecdba6e71d57264bfe" exitCode=0 Jan 21 11:27:29 crc kubenswrapper[4824]: I0121 11:27:29.405874 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-4h8mx" event={"ID":"91261d07-56c8-4da3-8dd5-6f23587d77be","Type":"ContainerDied","Data":"06b41897897bde2d466371c85738035a8ab9ee6530d1dfecdba6e71d57264bfe"} Jan 21 11:27:29 crc kubenswrapper[4824]: I0121 11:27:29.405911 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-4h8mx" event={"ID":"91261d07-56c8-4da3-8dd5-6f23587d77be","Type":"ContainerStarted","Data":"0d0a9f9151b1e3a86a4c82524d30beed3cbaa6676c1501c4fd1a3117b7a53765"} Jan 21 11:27:29 crc kubenswrapper[4824]: I0121 11:27:29.410352 4824 generic.go:334] "Generic (PLEG): container finished" podID="34afc8af-6ce5-4ee0-917b-326bd60d6237" containerID="7ab5f4aea14ccac0c6081bda9c7469c2c70bff8d5179b2221cca7a70044cada5" exitCode=0 Jan 21 11:27:29 crc kubenswrapper[4824]: I0121 11:27:29.410421 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-16cb-account-create-update-qjbx7" event={"ID":"34afc8af-6ce5-4ee0-917b-326bd60d6237","Type":"ContainerDied","Data":"7ab5f4aea14ccac0c6081bda9c7469c2c70bff8d5179b2221cca7a70044cada5"} Jan 21 11:27:29 crc kubenswrapper[4824]: I0121 11:27:29.410440 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-16cb-account-create-update-qjbx7" event={"ID":"34afc8af-6ce5-4ee0-917b-326bd60d6237","Type":"ContainerStarted","Data":"ff32875367cf89cd527e03854db7b2ba8ad3ebac9843661fa1849f3051811312"} Jan 21 11:27:29 crc kubenswrapper[4824]: I0121 11:27:29.416543 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"46f0fa4b-479a-42c5-b9b2-3f30a079be48","Type":"ContainerDied","Data":"5dc758f1e11243d2a910301ab89754d05c98dfa5446bb49c248a5833a44c6ce9"} Jan 21 11:27:29 crc kubenswrapper[4824]: I0121 11:27:29.416628 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Jan 21 11:27:29 crc kubenswrapper[4824]: I0121 11:27:29.421418 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=2.387191921 podStartE2EDuration="9.421400364s" podCreationTimestamp="2026-01-21 11:27:20 +0000 UTC" firstStartedPulling="2026-01-21 11:27:21.513346991 +0000 UTC m=+983.806376282" lastFinishedPulling="2026-01-21 11:27:28.547555434 +0000 UTC m=+990.840584725" observedRunningTime="2026-01-21 11:27:29.41321287 +0000 UTC m=+991.706242163" watchObservedRunningTime="2026-01-21 11:27:29.421400364 +0000 UTC m=+991.714429656" Jan 21 11:27:29 crc kubenswrapper[4824]: W0121 11:27:29.429731 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod12e7657f_78fe_4b5c_b614_4449e94366b9.slice/crio-62ab48f611e01d791ea972b5bd6b22f81360725b9ea98ca5dc5298079b346589 WatchSource:0}: Error finding container 62ab48f611e01d791ea972b5bd6b22f81360725b9ea98ca5dc5298079b346589: Status 404 returned error can't find the container with id 62ab48f611e01d791ea972b5bd6b22f81360725b9ea98ca5dc5298079b346589 Jan 21 11:27:29 crc kubenswrapper[4824]: I0121 11:27:29.439438 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-79f4c948b7-4hz4p"] Jan 21 11:27:29 crc kubenswrapper[4824]: I0121 11:27:29.443412 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-db-create-dhvfz" podStartSLOduration=4.443397575 podStartE2EDuration="4.443397575s" podCreationTimestamp="2026-01-21 11:27:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:27:29.428470159 +0000 UTC m=+991.721499451" watchObservedRunningTime="2026-01-21 11:27:29.443397575 +0000 UTC m=+991.736426867" Jan 21 11:27:29 crc kubenswrapper[4824]: I0121 11:27:29.460884 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-91fe-account-create-update-4xgv2" podStartSLOduration=4.460871335 podStartE2EDuration="4.460871335s" podCreationTimestamp="2026-01-21 11:27:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:27:29.442264799 +0000 UTC m=+991.735294091" watchObservedRunningTime="2026-01-21 11:27:29.460871335 +0000 UTC m=+991.753900627" Jan 21 11:27:29 crc kubenswrapper[4824]: I0121 11:27:29.756996 4824 scope.go:117] "RemoveContainer" containerID="1568d27a18b9471d6e4be5cfb1adad4950254a106a787756913dd9ddb73cf5f9" Jan 21 11:27:29 crc kubenswrapper[4824]: I0121 11:27:29.786573 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Jan 21 11:27:29 crc kubenswrapper[4824]: I0121 11:27:29.792154 4824 scope.go:117] "RemoveContainer" containerID="3b7c3d7bba0374d53d0ac03b9cb2de415ba41cc2525693fb788df9d2d6551fe4" Jan 21 11:27:29 crc kubenswrapper[4824]: I0121 11:27:29.797760 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Jan 21 11:27:29 crc kubenswrapper[4824]: I0121 11:27:29.808913 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Jan 21 11:27:29 crc kubenswrapper[4824]: I0121 11:27:29.818070 4824 scope.go:117] "RemoveContainer" containerID="7b2e891014d0053a1d16ec02f1fc280a3a0cbb96983be7316f8030e04e57c2be" Jan 21 11:27:29 crc kubenswrapper[4824]: I0121 11:27:29.824998 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Jan 21 11:27:29 crc kubenswrapper[4824]: I0121 11:27:29.841622 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Jan 21 11:27:29 crc kubenswrapper[4824]: E0121 11:27:29.841996 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bae23da5-948a-4883-b88d-d6ed5d6ccba7" containerName="ceilometer-notification-agent" Jan 21 11:27:29 crc kubenswrapper[4824]: I0121 11:27:29.842015 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="bae23da5-948a-4883-b88d-d6ed5d6ccba7" containerName="ceilometer-notification-agent" Jan 21 11:27:29 crc kubenswrapper[4824]: E0121 11:27:29.842026 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bae23da5-948a-4883-b88d-d6ed5d6ccba7" containerName="sg-core" Jan 21 11:27:29 crc kubenswrapper[4824]: I0121 11:27:29.842032 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="bae23da5-948a-4883-b88d-d6ed5d6ccba7" containerName="sg-core" Jan 21 11:27:29 crc kubenswrapper[4824]: E0121 11:27:29.842043 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46f0fa4b-479a-42c5-b9b2-3f30a079be48" containerName="glance-log" Jan 21 11:27:29 crc kubenswrapper[4824]: I0121 11:27:29.842049 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="46f0fa4b-479a-42c5-b9b2-3f30a079be48" containerName="glance-log" Jan 21 11:27:29 crc kubenswrapper[4824]: E0121 11:27:29.842066 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bae23da5-948a-4883-b88d-d6ed5d6ccba7" containerName="ceilometer-central-agent" Jan 21 11:27:29 crc kubenswrapper[4824]: I0121 11:27:29.842071 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="bae23da5-948a-4883-b88d-d6ed5d6ccba7" containerName="ceilometer-central-agent" Jan 21 11:27:29 crc kubenswrapper[4824]: E0121 11:27:29.842084 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bae23da5-948a-4883-b88d-d6ed5d6ccba7" containerName="proxy-httpd" Jan 21 11:27:29 crc kubenswrapper[4824]: I0121 11:27:29.842089 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="bae23da5-948a-4883-b88d-d6ed5d6ccba7" containerName="proxy-httpd" Jan 21 11:27:29 crc kubenswrapper[4824]: E0121 11:27:29.842096 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46f0fa4b-479a-42c5-b9b2-3f30a079be48" containerName="glance-httpd" Jan 21 11:27:29 crc kubenswrapper[4824]: I0121 11:27:29.842102 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="46f0fa4b-479a-42c5-b9b2-3f30a079be48" containerName="glance-httpd" Jan 21 11:27:29 crc kubenswrapper[4824]: I0121 11:27:29.842258 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="bae23da5-948a-4883-b88d-d6ed5d6ccba7" containerName="ceilometer-notification-agent" Jan 21 11:27:29 crc kubenswrapper[4824]: I0121 11:27:29.842271 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="46f0fa4b-479a-42c5-b9b2-3f30a079be48" containerName="glance-httpd" Jan 21 11:27:29 crc kubenswrapper[4824]: I0121 11:27:29.842282 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="bae23da5-948a-4883-b88d-d6ed5d6ccba7" containerName="proxy-httpd" Jan 21 11:27:29 crc kubenswrapper[4824]: I0121 11:27:29.842310 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="46f0fa4b-479a-42c5-b9b2-3f30a079be48" containerName="glance-log" Jan 21 11:27:29 crc kubenswrapper[4824]: I0121 11:27:29.842321 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="bae23da5-948a-4883-b88d-d6ed5d6ccba7" containerName="ceilometer-central-agent" Jan 21 11:27:29 crc kubenswrapper[4824]: I0121 11:27:29.842333 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="bae23da5-948a-4883-b88d-d6ed5d6ccba7" containerName="sg-core" Jan 21 11:27:29 crc kubenswrapper[4824]: I0121 11:27:29.843709 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 21 11:27:29 crc kubenswrapper[4824]: I0121 11:27:29.846475 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Jan 21 11:27:29 crc kubenswrapper[4824]: I0121 11:27:29.846640 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Jan 21 11:27:29 crc kubenswrapper[4824]: I0121 11:27:29.848503 4824 scope.go:117] "RemoveContainer" containerID="77d449bfcfde658bccf9c84cc9cff76978535cecd3aaeeb5e306a982392e5cf4" Jan 21 11:27:29 crc kubenswrapper[4824]: I0121 11:27:29.850018 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Jan 21 11:27:29 crc kubenswrapper[4824]: I0121 11:27:29.851322 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Jan 21 11:27:29 crc kubenswrapper[4824]: I0121 11:27:29.853446 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Jan 21 11:27:29 crc kubenswrapper[4824]: I0121 11:27:29.853725 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Jan 21 11:27:29 crc kubenswrapper[4824]: I0121 11:27:29.869023 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Jan 21 11:27:29 crc kubenswrapper[4824]: I0121 11:27:29.879080 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Jan 21 11:27:29 crc kubenswrapper[4824]: I0121 11:27:29.895917 4824 scope.go:117] "RemoveContainer" containerID="953a68ff3f5b22fd88f3a1a2babc253f6f9d0648a9b5bca14c07b85059d64e1a" Jan 21 11:27:29 crc kubenswrapper[4824]: I0121 11:27:29.942430 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7419ae4a-e8ea-4384-a9b8-3a3e02e27f25-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"7419ae4a-e8ea-4384-a9b8-3a3e02e27f25\") " pod="openstack/glance-default-internal-api-0" Jan 21 11:27:29 crc kubenswrapper[4824]: I0121 11:27:29.942695 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a89ff227-6009-4377-8f9f-34ac4078646f-log-httpd\") pod \"ceilometer-0\" (UID: \"a89ff227-6009-4377-8f9f-34ac4078646f\") " pod="openstack/ceilometer-0" Jan 21 11:27:29 crc kubenswrapper[4824]: I0121 11:27:29.942726 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a89ff227-6009-4377-8f9f-34ac4078646f-scripts\") pod \"ceilometer-0\" (UID: \"a89ff227-6009-4377-8f9f-34ac4078646f\") " pod="openstack/ceilometer-0" Jan 21 11:27:29 crc kubenswrapper[4824]: I0121 11:27:29.942757 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"7419ae4a-e8ea-4384-a9b8-3a3e02e27f25\") " pod="openstack/glance-default-internal-api-0" Jan 21 11:27:29 crc kubenswrapper[4824]: I0121 11:27:29.942787 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7419ae4a-e8ea-4384-a9b8-3a3e02e27f25-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"7419ae4a-e8ea-4384-a9b8-3a3e02e27f25\") " pod="openstack/glance-default-internal-api-0" Jan 21 11:27:29 crc kubenswrapper[4824]: I0121 11:27:29.942817 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7419ae4a-e8ea-4384-a9b8-3a3e02e27f25-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"7419ae4a-e8ea-4384-a9b8-3a3e02e27f25\") " pod="openstack/glance-default-internal-api-0" Jan 21 11:27:29 crc kubenswrapper[4824]: I0121 11:27:29.942845 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a89ff227-6009-4377-8f9f-34ac4078646f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a89ff227-6009-4377-8f9f-34ac4078646f\") " pod="openstack/ceilometer-0" Jan 21 11:27:29 crc kubenswrapper[4824]: I0121 11:27:29.942859 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-898h7\" (UniqueName: \"kubernetes.io/projected/a89ff227-6009-4377-8f9f-34ac4078646f-kube-api-access-898h7\") pod \"ceilometer-0\" (UID: \"a89ff227-6009-4377-8f9f-34ac4078646f\") " pod="openstack/ceilometer-0" Jan 21 11:27:29 crc kubenswrapper[4824]: I0121 11:27:29.942873 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a89ff227-6009-4377-8f9f-34ac4078646f-config-data\") pod \"ceilometer-0\" (UID: \"a89ff227-6009-4377-8f9f-34ac4078646f\") " pod="openstack/ceilometer-0" Jan 21 11:27:29 crc kubenswrapper[4824]: I0121 11:27:29.942948 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a89ff227-6009-4377-8f9f-34ac4078646f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a89ff227-6009-4377-8f9f-34ac4078646f\") " pod="openstack/ceilometer-0" Jan 21 11:27:29 crc kubenswrapper[4824]: I0121 11:27:29.943058 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7419ae4a-e8ea-4384-a9b8-3a3e02e27f25-scripts\") pod \"glance-default-internal-api-0\" (UID: \"7419ae4a-e8ea-4384-a9b8-3a3e02e27f25\") " pod="openstack/glance-default-internal-api-0" Jan 21 11:27:29 crc kubenswrapper[4824]: I0121 11:27:29.943115 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pcd2n\" (UniqueName: \"kubernetes.io/projected/7419ae4a-e8ea-4384-a9b8-3a3e02e27f25-kube-api-access-pcd2n\") pod \"glance-default-internal-api-0\" (UID: \"7419ae4a-e8ea-4384-a9b8-3a3e02e27f25\") " pod="openstack/glance-default-internal-api-0" Jan 21 11:27:29 crc kubenswrapper[4824]: I0121 11:27:29.943168 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a89ff227-6009-4377-8f9f-34ac4078646f-run-httpd\") pod \"ceilometer-0\" (UID: \"a89ff227-6009-4377-8f9f-34ac4078646f\") " pod="openstack/ceilometer-0" Jan 21 11:27:29 crc kubenswrapper[4824]: I0121 11:27:29.943185 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7419ae4a-e8ea-4384-a9b8-3a3e02e27f25-config-data\") pod \"glance-default-internal-api-0\" (UID: \"7419ae4a-e8ea-4384-a9b8-3a3e02e27f25\") " pod="openstack/glance-default-internal-api-0" Jan 21 11:27:29 crc kubenswrapper[4824]: I0121 11:27:29.943200 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7419ae4a-e8ea-4384-a9b8-3a3e02e27f25-logs\") pod \"glance-default-internal-api-0\" (UID: \"7419ae4a-e8ea-4384-a9b8-3a3e02e27f25\") " pod="openstack/glance-default-internal-api-0" Jan 21 11:27:30 crc kubenswrapper[4824]: I0121 11:27:30.045106 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7419ae4a-e8ea-4384-a9b8-3a3e02e27f25-scripts\") pod \"glance-default-internal-api-0\" (UID: \"7419ae4a-e8ea-4384-a9b8-3a3e02e27f25\") " pod="openstack/glance-default-internal-api-0" Jan 21 11:27:30 crc kubenswrapper[4824]: I0121 11:27:30.045144 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pcd2n\" (UniqueName: \"kubernetes.io/projected/7419ae4a-e8ea-4384-a9b8-3a3e02e27f25-kube-api-access-pcd2n\") pod \"glance-default-internal-api-0\" (UID: \"7419ae4a-e8ea-4384-a9b8-3a3e02e27f25\") " pod="openstack/glance-default-internal-api-0" Jan 21 11:27:30 crc kubenswrapper[4824]: I0121 11:27:30.045187 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a89ff227-6009-4377-8f9f-34ac4078646f-run-httpd\") pod \"ceilometer-0\" (UID: \"a89ff227-6009-4377-8f9f-34ac4078646f\") " pod="openstack/ceilometer-0" Jan 21 11:27:30 crc kubenswrapper[4824]: I0121 11:27:30.045206 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7419ae4a-e8ea-4384-a9b8-3a3e02e27f25-config-data\") pod \"glance-default-internal-api-0\" (UID: \"7419ae4a-e8ea-4384-a9b8-3a3e02e27f25\") " pod="openstack/glance-default-internal-api-0" Jan 21 11:27:30 crc kubenswrapper[4824]: I0121 11:27:30.045225 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7419ae4a-e8ea-4384-a9b8-3a3e02e27f25-logs\") pod \"glance-default-internal-api-0\" (UID: \"7419ae4a-e8ea-4384-a9b8-3a3e02e27f25\") " pod="openstack/glance-default-internal-api-0" Jan 21 11:27:30 crc kubenswrapper[4824]: I0121 11:27:30.045284 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7419ae4a-e8ea-4384-a9b8-3a3e02e27f25-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"7419ae4a-e8ea-4384-a9b8-3a3e02e27f25\") " pod="openstack/glance-default-internal-api-0" Jan 21 11:27:30 crc kubenswrapper[4824]: I0121 11:27:30.045309 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a89ff227-6009-4377-8f9f-34ac4078646f-log-httpd\") pod \"ceilometer-0\" (UID: \"a89ff227-6009-4377-8f9f-34ac4078646f\") " pod="openstack/ceilometer-0" Jan 21 11:27:30 crc kubenswrapper[4824]: I0121 11:27:30.045332 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a89ff227-6009-4377-8f9f-34ac4078646f-scripts\") pod \"ceilometer-0\" (UID: \"a89ff227-6009-4377-8f9f-34ac4078646f\") " pod="openstack/ceilometer-0" Jan 21 11:27:30 crc kubenswrapper[4824]: I0121 11:27:30.045354 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"7419ae4a-e8ea-4384-a9b8-3a3e02e27f25\") " pod="openstack/glance-default-internal-api-0" Jan 21 11:27:30 crc kubenswrapper[4824]: I0121 11:27:30.045381 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7419ae4a-e8ea-4384-a9b8-3a3e02e27f25-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"7419ae4a-e8ea-4384-a9b8-3a3e02e27f25\") " pod="openstack/glance-default-internal-api-0" Jan 21 11:27:30 crc kubenswrapper[4824]: I0121 11:27:30.045405 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7419ae4a-e8ea-4384-a9b8-3a3e02e27f25-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"7419ae4a-e8ea-4384-a9b8-3a3e02e27f25\") " pod="openstack/glance-default-internal-api-0" Jan 21 11:27:30 crc kubenswrapper[4824]: I0121 11:27:30.045433 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a89ff227-6009-4377-8f9f-34ac4078646f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a89ff227-6009-4377-8f9f-34ac4078646f\") " pod="openstack/ceilometer-0" Jan 21 11:27:30 crc kubenswrapper[4824]: I0121 11:27:30.045447 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-898h7\" (UniqueName: \"kubernetes.io/projected/a89ff227-6009-4377-8f9f-34ac4078646f-kube-api-access-898h7\") pod \"ceilometer-0\" (UID: \"a89ff227-6009-4377-8f9f-34ac4078646f\") " pod="openstack/ceilometer-0" Jan 21 11:27:30 crc kubenswrapper[4824]: I0121 11:27:30.045463 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a89ff227-6009-4377-8f9f-34ac4078646f-config-data\") pod \"ceilometer-0\" (UID: \"a89ff227-6009-4377-8f9f-34ac4078646f\") " pod="openstack/ceilometer-0" Jan 21 11:27:30 crc kubenswrapper[4824]: I0121 11:27:30.045534 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a89ff227-6009-4377-8f9f-34ac4078646f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a89ff227-6009-4377-8f9f-34ac4078646f\") " pod="openstack/ceilometer-0" Jan 21 11:27:30 crc kubenswrapper[4824]: I0121 11:27:30.045674 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a89ff227-6009-4377-8f9f-34ac4078646f-run-httpd\") pod \"ceilometer-0\" (UID: \"a89ff227-6009-4377-8f9f-34ac4078646f\") " pod="openstack/ceilometer-0" Jan 21 11:27:30 crc kubenswrapper[4824]: I0121 11:27:30.045684 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7419ae4a-e8ea-4384-a9b8-3a3e02e27f25-logs\") pod \"glance-default-internal-api-0\" (UID: \"7419ae4a-e8ea-4384-a9b8-3a3e02e27f25\") " pod="openstack/glance-default-internal-api-0" Jan 21 11:27:30 crc kubenswrapper[4824]: I0121 11:27:30.046006 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a89ff227-6009-4377-8f9f-34ac4078646f-log-httpd\") pod \"ceilometer-0\" (UID: \"a89ff227-6009-4377-8f9f-34ac4078646f\") " pod="openstack/ceilometer-0" Jan 21 11:27:30 crc kubenswrapper[4824]: I0121 11:27:30.046156 4824 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"7419ae4a-e8ea-4384-a9b8-3a3e02e27f25\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/glance-default-internal-api-0" Jan 21 11:27:30 crc kubenswrapper[4824]: I0121 11:27:30.046154 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7419ae4a-e8ea-4384-a9b8-3a3e02e27f25-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"7419ae4a-e8ea-4384-a9b8-3a3e02e27f25\") " pod="openstack/glance-default-internal-api-0" Jan 21 11:27:30 crc kubenswrapper[4824]: I0121 11:27:30.050157 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7419ae4a-e8ea-4384-a9b8-3a3e02e27f25-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"7419ae4a-e8ea-4384-a9b8-3a3e02e27f25\") " pod="openstack/glance-default-internal-api-0" Jan 21 11:27:30 crc kubenswrapper[4824]: I0121 11:27:30.050515 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a89ff227-6009-4377-8f9f-34ac4078646f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a89ff227-6009-4377-8f9f-34ac4078646f\") " pod="openstack/ceilometer-0" Jan 21 11:27:30 crc kubenswrapper[4824]: I0121 11:27:30.050524 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a89ff227-6009-4377-8f9f-34ac4078646f-config-data\") pod \"ceilometer-0\" (UID: \"a89ff227-6009-4377-8f9f-34ac4078646f\") " pod="openstack/ceilometer-0" Jan 21 11:27:30 crc kubenswrapper[4824]: I0121 11:27:30.051161 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a89ff227-6009-4377-8f9f-34ac4078646f-scripts\") pod \"ceilometer-0\" (UID: \"a89ff227-6009-4377-8f9f-34ac4078646f\") " pod="openstack/ceilometer-0" Jan 21 11:27:30 crc kubenswrapper[4824]: I0121 11:27:30.054952 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7419ae4a-e8ea-4384-a9b8-3a3e02e27f25-config-data\") pod \"glance-default-internal-api-0\" (UID: \"7419ae4a-e8ea-4384-a9b8-3a3e02e27f25\") " pod="openstack/glance-default-internal-api-0" Jan 21 11:27:30 crc kubenswrapper[4824]: I0121 11:27:30.057912 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a89ff227-6009-4377-8f9f-34ac4078646f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a89ff227-6009-4377-8f9f-34ac4078646f\") " pod="openstack/ceilometer-0" Jan 21 11:27:30 crc kubenswrapper[4824]: I0121 11:27:30.063566 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7419ae4a-e8ea-4384-a9b8-3a3e02e27f25-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"7419ae4a-e8ea-4384-a9b8-3a3e02e27f25\") " pod="openstack/glance-default-internal-api-0" Jan 21 11:27:30 crc kubenswrapper[4824]: I0121 11:27:30.065663 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-898h7\" (UniqueName: \"kubernetes.io/projected/a89ff227-6009-4377-8f9f-34ac4078646f-kube-api-access-898h7\") pod \"ceilometer-0\" (UID: \"a89ff227-6009-4377-8f9f-34ac4078646f\") " pod="openstack/ceilometer-0" Jan 21 11:27:30 crc kubenswrapper[4824]: I0121 11:27:30.068198 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7419ae4a-e8ea-4384-a9b8-3a3e02e27f25-scripts\") pod \"glance-default-internal-api-0\" (UID: \"7419ae4a-e8ea-4384-a9b8-3a3e02e27f25\") " pod="openstack/glance-default-internal-api-0" Jan 21 11:27:30 crc kubenswrapper[4824]: I0121 11:27:30.069018 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="46f0fa4b-479a-42c5-b9b2-3f30a079be48" path="/var/lib/kubelet/pods/46f0fa4b-479a-42c5-b9b2-3f30a079be48/volumes" Jan 21 11:27:30 crc kubenswrapper[4824]: I0121 11:27:30.069852 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bae23da5-948a-4883-b88d-d6ed5d6ccba7" path="/var/lib/kubelet/pods/bae23da5-948a-4883-b88d-d6ed5d6ccba7/volumes" Jan 21 11:27:30 crc kubenswrapper[4824]: I0121 11:27:30.074137 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pcd2n\" (UniqueName: \"kubernetes.io/projected/7419ae4a-e8ea-4384-a9b8-3a3e02e27f25-kube-api-access-pcd2n\") pod \"glance-default-internal-api-0\" (UID: \"7419ae4a-e8ea-4384-a9b8-3a3e02e27f25\") " pod="openstack/glance-default-internal-api-0" Jan 21 11:27:30 crc kubenswrapper[4824]: I0121 11:27:30.098199 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"7419ae4a-e8ea-4384-a9b8-3a3e02e27f25\") " pod="openstack/glance-default-internal-api-0" Jan 21 11:27:30 crc kubenswrapper[4824]: I0121 11:27:30.171341 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 21 11:27:30 crc kubenswrapper[4824]: I0121 11:27:30.176599 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Jan 21 11:27:30 crc kubenswrapper[4824]: I0121 11:27:30.426695 4824 generic.go:334] "Generic (PLEG): container finished" podID="d5162c70-9359-42d5-a02d-16140ac4e366" containerID="6fa0076da1636efc5cc0178f55ca211d393f930b6c63c3d85cac4eedd9f2d288" exitCode=0 Jan 21 11:27:30 crc kubenswrapper[4824]: I0121 11:27:30.426904 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-dhvfz" event={"ID":"d5162c70-9359-42d5-a02d-16140ac4e366","Type":"ContainerDied","Data":"6fa0076da1636efc5cc0178f55ca211d393f930b6c63c3d85cac4eedd9f2d288"} Jan 21 11:27:30 crc kubenswrapper[4824]: I0121 11:27:30.428574 4824 generic.go:334] "Generic (PLEG): container finished" podID="60da7490-4aa1-4880-bfcb-e51a4ed99f91" containerID="a9bd46298904e9e7c4cccbdd1cecf51e20eaa4ef17b0e64f0ae86c7cca0ec79d" exitCode=0 Jan 21 11:27:30 crc kubenswrapper[4824]: I0121 11:27:30.428611 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-91fe-account-create-update-4xgv2" event={"ID":"60da7490-4aa1-4880-bfcb-e51a4ed99f91","Type":"ContainerDied","Data":"a9bd46298904e9e7c4cccbdd1cecf51e20eaa4ef17b0e64f0ae86c7cca0ec79d"} Jan 21 11:27:30 crc kubenswrapper[4824]: I0121 11:27:30.429867 4824 generic.go:334] "Generic (PLEG): container finished" podID="12e7657f-78fe-4b5c-b614-4449e94366b9" containerID="edd12500feb323e4dab204d7803bc5160bb50ef2b5116e5ba7686e65e4e155fc" exitCode=0 Jan 21 11:27:30 crc kubenswrapper[4824]: I0121 11:27:30.429901 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-f2xfg" event={"ID":"12e7657f-78fe-4b5c-b614-4449e94366b9","Type":"ContainerDied","Data":"edd12500feb323e4dab204d7803bc5160bb50ef2b5116e5ba7686e65e4e155fc"} Jan 21 11:27:30 crc kubenswrapper[4824]: I0121 11:27:30.429914 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-f2xfg" event={"ID":"12e7657f-78fe-4b5c-b614-4449e94366b9","Type":"ContainerStarted","Data":"62ab48f611e01d791ea972b5bd6b22f81360725b9ea98ca5dc5298079b346589"} Jan 21 11:27:30 crc kubenswrapper[4824]: I0121 11:27:30.442612 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-79f4c948b7-4hz4p" event={"ID":"01d4ce4d-3728-4f4c-9fbf-ce58793b6b7d","Type":"ContainerStarted","Data":"e86601c8e5d510911b89a1fe69a3df136499236541d2dc47c46a5f7da7a5581e"} Jan 21 11:27:30 crc kubenswrapper[4824]: I0121 11:27:30.442653 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-79f4c948b7-4hz4p" event={"ID":"01d4ce4d-3728-4f4c-9fbf-ce58793b6b7d","Type":"ContainerStarted","Data":"4656bc506da33d3c8ed6fc9a4ae9d4a7673549488f4d44a12d49848538d1f1ce"} Jan 21 11:27:30 crc kubenswrapper[4824]: I0121 11:27:30.442667 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-79f4c948b7-4hz4p" event={"ID":"01d4ce4d-3728-4f4c-9fbf-ce58793b6b7d","Type":"ContainerStarted","Data":"eddc6d8deacb8d5b8460a0db4d8a825a8f5b8f3b3aa8d23bbcf51fa457d8d915"} Jan 21 11:27:30 crc kubenswrapper[4824]: I0121 11:27:30.442704 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-79f4c948b7-4hz4p" Jan 21 11:27:30 crc kubenswrapper[4824]: I0121 11:27:30.442725 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-79f4c948b7-4hz4p" Jan 21 11:27:30 crc kubenswrapper[4824]: I0121 11:27:30.445247 4824 generic.go:334] "Generic (PLEG): container finished" podID="b24eefe0-9cb1-4672-85d8-6dfe2babebff" containerID="7d5e936f6ab78e748abd4aabb374f95965e7eaaf526fd1a755b675eed79144ec" exitCode=0 Jan 21 11:27:30 crc kubenswrapper[4824]: I0121 11:27:30.445300 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-90fb-account-create-update-m4n4l" event={"ID":"b24eefe0-9cb1-4672-85d8-6dfe2babebff","Type":"ContainerDied","Data":"7d5e936f6ab78e748abd4aabb374f95965e7eaaf526fd1a755b675eed79144ec"} Jan 21 11:27:30 crc kubenswrapper[4824]: I0121 11:27:30.501142 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-79f4c948b7-4hz4p" podStartSLOduration=5.501126765 podStartE2EDuration="5.501126765s" podCreationTimestamp="2026-01-21 11:27:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:27:30.493402345 +0000 UTC m=+992.786431636" watchObservedRunningTime="2026-01-21 11:27:30.501126765 +0000 UTC m=+992.794156057" Jan 21 11:27:30 crc kubenswrapper[4824]: I0121 11:27:30.598995 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Jan 21 11:27:30 crc kubenswrapper[4824]: I0121 11:27:30.682545 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Jan 21 11:27:30 crc kubenswrapper[4824]: I0121 11:27:30.799104 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-4h8mx" Jan 21 11:27:30 crc kubenswrapper[4824]: I0121 11:27:30.803240 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-16cb-account-create-update-qjbx7" Jan 21 11:27:30 crc kubenswrapper[4824]: I0121 11:27:30.858370 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/91261d07-56c8-4da3-8dd5-6f23587d77be-operator-scripts\") pod \"91261d07-56c8-4da3-8dd5-6f23587d77be\" (UID: \"91261d07-56c8-4da3-8dd5-6f23587d77be\") " Jan 21 11:27:30 crc kubenswrapper[4824]: I0121 11:27:30.858446 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6mwhf\" (UniqueName: \"kubernetes.io/projected/91261d07-56c8-4da3-8dd5-6f23587d77be-kube-api-access-6mwhf\") pod \"91261d07-56c8-4da3-8dd5-6f23587d77be\" (UID: \"91261d07-56c8-4da3-8dd5-6f23587d77be\") " Jan 21 11:27:30 crc kubenswrapper[4824]: I0121 11:27:30.858584 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/34afc8af-6ce5-4ee0-917b-326bd60d6237-operator-scripts\") pod \"34afc8af-6ce5-4ee0-917b-326bd60d6237\" (UID: \"34afc8af-6ce5-4ee0-917b-326bd60d6237\") " Jan 21 11:27:30 crc kubenswrapper[4824]: I0121 11:27:30.858718 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-frs7l\" (UniqueName: \"kubernetes.io/projected/34afc8af-6ce5-4ee0-917b-326bd60d6237-kube-api-access-frs7l\") pod \"34afc8af-6ce5-4ee0-917b-326bd60d6237\" (UID: \"34afc8af-6ce5-4ee0-917b-326bd60d6237\") " Jan 21 11:27:30 crc kubenswrapper[4824]: I0121 11:27:30.858853 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/91261d07-56c8-4da3-8dd5-6f23587d77be-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "91261d07-56c8-4da3-8dd5-6f23587d77be" (UID: "91261d07-56c8-4da3-8dd5-6f23587d77be"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:27:30 crc kubenswrapper[4824]: I0121 11:27:30.859263 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/34afc8af-6ce5-4ee0-917b-326bd60d6237-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "34afc8af-6ce5-4ee0-917b-326bd60d6237" (UID: "34afc8af-6ce5-4ee0-917b-326bd60d6237"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:27:30 crc kubenswrapper[4824]: I0121 11:27:30.859463 4824 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/91261d07-56c8-4da3-8dd5-6f23587d77be-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:30 crc kubenswrapper[4824]: I0121 11:27:30.859487 4824 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/34afc8af-6ce5-4ee0-917b-326bd60d6237-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:30 crc kubenswrapper[4824]: I0121 11:27:30.867177 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/91261d07-56c8-4da3-8dd5-6f23587d77be-kube-api-access-6mwhf" (OuterVolumeSpecName: "kube-api-access-6mwhf") pod "91261d07-56c8-4da3-8dd5-6f23587d77be" (UID: "91261d07-56c8-4da3-8dd5-6f23587d77be"). InnerVolumeSpecName "kube-api-access-6mwhf". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:27:30 crc kubenswrapper[4824]: I0121 11:27:30.867388 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/34afc8af-6ce5-4ee0-917b-326bd60d6237-kube-api-access-frs7l" (OuterVolumeSpecName: "kube-api-access-frs7l") pod "34afc8af-6ce5-4ee0-917b-326bd60d6237" (UID: "34afc8af-6ce5-4ee0-917b-326bd60d6237"). InnerVolumeSpecName "kube-api-access-frs7l". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:27:30 crc kubenswrapper[4824]: I0121 11:27:30.961666 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6mwhf\" (UniqueName: \"kubernetes.io/projected/91261d07-56c8-4da3-8dd5-6f23587d77be-kube-api-access-6mwhf\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:30 crc kubenswrapper[4824]: I0121 11:27:30.961697 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-frs7l\" (UniqueName: \"kubernetes.io/projected/34afc8af-6ce5-4ee0-917b-326bd60d6237-kube-api-access-frs7l\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:31 crc kubenswrapper[4824]: I0121 11:27:31.460567 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"7419ae4a-e8ea-4384-a9b8-3a3e02e27f25","Type":"ContainerStarted","Data":"fa8330abdceca8284591245f5b1c8a2c098f8d7542e6d63ae0422dce6173b423"} Jan 21 11:27:31 crc kubenswrapper[4824]: I0121 11:27:31.460865 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"7419ae4a-e8ea-4384-a9b8-3a3e02e27f25","Type":"ContainerStarted","Data":"3411927c2dfcc913f64e09b90f13938fb8a275cb1f99d46bac7986f5d1d4d172"} Jan 21 11:27:31 crc kubenswrapper[4824]: I0121 11:27:31.462846 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-4h8mx" event={"ID":"91261d07-56c8-4da3-8dd5-6f23587d77be","Type":"ContainerDied","Data":"0d0a9f9151b1e3a86a4c82524d30beed3cbaa6676c1501c4fd1a3117b7a53765"} Jan 21 11:27:31 crc kubenswrapper[4824]: I0121 11:27:31.462898 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0d0a9f9151b1e3a86a4c82524d30beed3cbaa6676c1501c4fd1a3117b7a53765" Jan 21 11:27:31 crc kubenswrapper[4824]: I0121 11:27:31.463004 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-4h8mx" Jan 21 11:27:31 crc kubenswrapper[4824]: I0121 11:27:31.465245 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-16cb-account-create-update-qjbx7" event={"ID":"34afc8af-6ce5-4ee0-917b-326bd60d6237","Type":"ContainerDied","Data":"ff32875367cf89cd527e03854db7b2ba8ad3ebac9843661fa1849f3051811312"} Jan 21 11:27:31 crc kubenswrapper[4824]: I0121 11:27:31.465292 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ff32875367cf89cd527e03854db7b2ba8ad3ebac9843661fa1849f3051811312" Jan 21 11:27:31 crc kubenswrapper[4824]: I0121 11:27:31.465255 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-16cb-account-create-update-qjbx7" Jan 21 11:27:31 crc kubenswrapper[4824]: I0121 11:27:31.466387 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a89ff227-6009-4377-8f9f-34ac4078646f","Type":"ContainerStarted","Data":"02e341630b226a52243314bc26a342a8ea843da778071cfc554f827999082ef2"} Jan 21 11:27:31 crc kubenswrapper[4824]: I0121 11:27:31.466417 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a89ff227-6009-4377-8f9f-34ac4078646f","Type":"ContainerStarted","Data":"24216a6df3f372a52006263ca19c690f67c9d1bbdb8896d6d5fbff2e54961470"} Jan 21 11:27:31 crc kubenswrapper[4824]: I0121 11:27:31.774123 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Jan 21 11:27:31 crc kubenswrapper[4824]: I0121 11:27:31.774661 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="12776f5e-c372-41bd-91fe-ba7d0915050a" containerName="glance-log" containerID="cri-o://9f11951e86971107f0a8d485c5dead64b960ce2e124caba92b7e241dae731aca" gracePeriod=30 Jan 21 11:27:31 crc kubenswrapper[4824]: I0121 11:27:31.774794 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="12776f5e-c372-41bd-91fe-ba7d0915050a" containerName="glance-httpd" containerID="cri-o://c66cf375354a975633a38f62d2bcee1e315728b454ed407911dcd2b8fbd32686" gracePeriod=30 Jan 21 11:27:31 crc kubenswrapper[4824]: I0121 11:27:31.961288 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-dhvfz" Jan 21 11:27:31 crc kubenswrapper[4824]: I0121 11:27:31.988851 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-91fe-account-create-update-4xgv2" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.009787 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-90fb-account-create-update-m4n4l" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.016867 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-f2xfg" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.088339 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b24eefe0-9cb1-4672-85d8-6dfe2babebff-operator-scripts\") pod \"b24eefe0-9cb1-4672-85d8-6dfe2babebff\" (UID: \"b24eefe0-9cb1-4672-85d8-6dfe2babebff\") " Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.088484 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/60da7490-4aa1-4880-bfcb-e51a4ed99f91-operator-scripts\") pod \"60da7490-4aa1-4880-bfcb-e51a4ed99f91\" (UID: \"60da7490-4aa1-4880-bfcb-e51a4ed99f91\") " Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.088522 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d5162c70-9359-42d5-a02d-16140ac4e366-operator-scripts\") pod \"d5162c70-9359-42d5-a02d-16140ac4e366\" (UID: \"d5162c70-9359-42d5-a02d-16140ac4e366\") " Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.088544 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzjdm\" (UniqueName: \"kubernetes.io/projected/b24eefe0-9cb1-4672-85d8-6dfe2babebff-kube-api-access-lzjdm\") pod \"b24eefe0-9cb1-4672-85d8-6dfe2babebff\" (UID: \"b24eefe0-9cb1-4672-85d8-6dfe2babebff\") " Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.088584 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-btjhr\" (UniqueName: \"kubernetes.io/projected/d5162c70-9359-42d5-a02d-16140ac4e366-kube-api-access-btjhr\") pod \"d5162c70-9359-42d5-a02d-16140ac4e366\" (UID: \"d5162c70-9359-42d5-a02d-16140ac4e366\") " Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.088648 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/12e7657f-78fe-4b5c-b614-4449e94366b9-operator-scripts\") pod \"12e7657f-78fe-4b5c-b614-4449e94366b9\" (UID: \"12e7657f-78fe-4b5c-b614-4449e94366b9\") " Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.088691 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tgm2l\" (UniqueName: \"kubernetes.io/projected/60da7490-4aa1-4880-bfcb-e51a4ed99f91-kube-api-access-tgm2l\") pod \"60da7490-4aa1-4880-bfcb-e51a4ed99f91\" (UID: \"60da7490-4aa1-4880-bfcb-e51a4ed99f91\") " Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.088708 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cpkmt\" (UniqueName: \"kubernetes.io/projected/12e7657f-78fe-4b5c-b614-4449e94366b9-kube-api-access-cpkmt\") pod \"12e7657f-78fe-4b5c-b614-4449e94366b9\" (UID: \"12e7657f-78fe-4b5c-b614-4449e94366b9\") " Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.089005 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d5162c70-9359-42d5-a02d-16140ac4e366-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "d5162c70-9359-42d5-a02d-16140ac4e366" (UID: "d5162c70-9359-42d5-a02d-16140ac4e366"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.089354 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/12e7657f-78fe-4b5c-b614-4449e94366b9-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "12e7657f-78fe-4b5c-b614-4449e94366b9" (UID: "12e7657f-78fe-4b5c-b614-4449e94366b9"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.089439 4824 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d5162c70-9359-42d5-a02d-16140ac4e366-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.089523 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/60da7490-4aa1-4880-bfcb-e51a4ed99f91-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "60da7490-4aa1-4880-bfcb-e51a4ed99f91" (UID: "60da7490-4aa1-4880-bfcb-e51a4ed99f91"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.089688 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b24eefe0-9cb1-4672-85d8-6dfe2babebff-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b24eefe0-9cb1-4672-85d8-6dfe2babebff" (UID: "b24eefe0-9cb1-4672-85d8-6dfe2babebff"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.092033 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d5162c70-9359-42d5-a02d-16140ac4e366-kube-api-access-btjhr" (OuterVolumeSpecName: "kube-api-access-btjhr") pod "d5162c70-9359-42d5-a02d-16140ac4e366" (UID: "d5162c70-9359-42d5-a02d-16140ac4e366"). InnerVolumeSpecName "kube-api-access-btjhr". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.092111 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/12e7657f-78fe-4b5c-b614-4449e94366b9-kube-api-access-cpkmt" (OuterVolumeSpecName: "kube-api-access-cpkmt") pod "12e7657f-78fe-4b5c-b614-4449e94366b9" (UID: "12e7657f-78fe-4b5c-b614-4449e94366b9"). InnerVolumeSpecName "kube-api-access-cpkmt". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.092194 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/60da7490-4aa1-4880-bfcb-e51a4ed99f91-kube-api-access-tgm2l" (OuterVolumeSpecName: "kube-api-access-tgm2l") pod "60da7490-4aa1-4880-bfcb-e51a4ed99f91" (UID: "60da7490-4aa1-4880-bfcb-e51a4ed99f91"). InnerVolumeSpecName "kube-api-access-tgm2l". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.092339 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b24eefe0-9cb1-4672-85d8-6dfe2babebff-kube-api-access-lzjdm" (OuterVolumeSpecName: "kube-api-access-lzjdm") pod "b24eefe0-9cb1-4672-85d8-6dfe2babebff" (UID: "b24eefe0-9cb1-4672-85d8-6dfe2babebff"). InnerVolumeSpecName "kube-api-access-lzjdm". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.191191 4824 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/60da7490-4aa1-4880-bfcb-e51a4ed99f91-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.191482 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzjdm\" (UniqueName: \"kubernetes.io/projected/b24eefe0-9cb1-4672-85d8-6dfe2babebff-kube-api-access-lzjdm\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.191494 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-btjhr\" (UniqueName: \"kubernetes.io/projected/d5162c70-9359-42d5-a02d-16140ac4e366-kube-api-access-btjhr\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.191503 4824 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/12e7657f-78fe-4b5c-b614-4449e94366b9-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.191512 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tgm2l\" (UniqueName: \"kubernetes.io/projected/60da7490-4aa1-4880-bfcb-e51a4ed99f91-kube-api-access-tgm2l\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.191520 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cpkmt\" (UniqueName: \"kubernetes.io/projected/12e7657f-78fe-4b5c-b614-4449e94366b9-kube-api-access-cpkmt\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.191527 4824 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b24eefe0-9cb1-4672-85d8-6dfe2babebff-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.349543 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-engine-656b845fb4-qst2s"] Jan 21 11:27:32 crc kubenswrapper[4824]: E0121 11:27:32.352351 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d5162c70-9359-42d5-a02d-16140ac4e366" containerName="mariadb-database-create" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.352456 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="d5162c70-9359-42d5-a02d-16140ac4e366" containerName="mariadb-database-create" Jan 21 11:27:32 crc kubenswrapper[4824]: E0121 11:27:32.352536 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="60da7490-4aa1-4880-bfcb-e51a4ed99f91" containerName="mariadb-account-create-update" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.352584 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="60da7490-4aa1-4880-bfcb-e51a4ed99f91" containerName="mariadb-account-create-update" Jan 21 11:27:32 crc kubenswrapper[4824]: E0121 11:27:32.352634 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b24eefe0-9cb1-4672-85d8-6dfe2babebff" containerName="mariadb-account-create-update" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.352683 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="b24eefe0-9cb1-4672-85d8-6dfe2babebff" containerName="mariadb-account-create-update" Jan 21 11:27:32 crc kubenswrapper[4824]: E0121 11:27:32.352734 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34afc8af-6ce5-4ee0-917b-326bd60d6237" containerName="mariadb-account-create-update" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.352789 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="34afc8af-6ce5-4ee0-917b-326bd60d6237" containerName="mariadb-account-create-update" Jan 21 11:27:32 crc kubenswrapper[4824]: E0121 11:27:32.352844 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="12e7657f-78fe-4b5c-b614-4449e94366b9" containerName="mariadb-database-create" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.352893 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="12e7657f-78fe-4b5c-b614-4449e94366b9" containerName="mariadb-database-create" Jan 21 11:27:32 crc kubenswrapper[4824]: E0121 11:27:32.352948 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91261d07-56c8-4da3-8dd5-6f23587d77be" containerName="mariadb-database-create" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.353020 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="91261d07-56c8-4da3-8dd5-6f23587d77be" containerName="mariadb-database-create" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.353253 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="91261d07-56c8-4da3-8dd5-6f23587d77be" containerName="mariadb-database-create" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.353323 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="34afc8af-6ce5-4ee0-917b-326bd60d6237" containerName="mariadb-account-create-update" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.353381 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="12e7657f-78fe-4b5c-b614-4449e94366b9" containerName="mariadb-database-create" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.354110 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="60da7490-4aa1-4880-bfcb-e51a4ed99f91" containerName="mariadb-account-create-update" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.354186 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="b24eefe0-9cb1-4672-85d8-6dfe2babebff" containerName="mariadb-account-create-update" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.354251 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="d5162c70-9359-42d5-a02d-16140ac4e366" containerName="mariadb-database-create" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.354803 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-656b845fb4-qst2s" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.357160 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-heat-dockercfg-cj2g7" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.359033 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-engine-config-data" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.376680 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-config-data" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.376893 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-656b845fb4-qst2s"] Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.419900 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-cfnapi-667b46bf4d-5h2cq"] Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.421618 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-667b46bf4d-5h2cq" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.429816 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-cfnapi-config-data" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.452816 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-667b46bf4d-5h2cq"] Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.484292 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a89ff227-6009-4377-8f9f-34ac4078646f","Type":"ContainerStarted","Data":"01663fbfa71a8a22bc17e94621a84e2c11185ae78db507b95bd3152b51310836"} Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.487427 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"7419ae4a-e8ea-4384-a9b8-3a3e02e27f25","Type":"ContainerStarted","Data":"a12c1689cb4abdc2a588fab5461b22fd986834a8bbf20ceec638f657674996c6"} Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.493621 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-90fb-account-create-update-m4n4l" event={"ID":"b24eefe0-9cb1-4672-85d8-6dfe2babebff","Type":"ContainerDied","Data":"5ca82bb30aa3c9ac045630af69b3a4a8697ebeccee0b68dc49d5dad6be111c3a"} Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.493722 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5ca82bb30aa3c9ac045630af69b3a4a8697ebeccee0b68dc49d5dad6be111c3a" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.493667 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-90fb-account-create-update-m4n4l" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.493980 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7756b9d78c-ftf8g"] Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.495516 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7756b9d78c-ftf8g" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.495831 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-dhvfz" event={"ID":"d5162c70-9359-42d5-a02d-16140ac4e366","Type":"ContainerDied","Data":"2c6b2c2ef954db67b738c424fe09e651c130b39cd2f9702c3ab0d17616f3d362"} Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.495861 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2c6b2c2ef954db67b738c424fe09e651c130b39cd2f9702c3ab0d17616f3d362" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.495918 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-dhvfz" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.498597 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-91fe-account-create-update-4xgv2" event={"ID":"60da7490-4aa1-4880-bfcb-e51a4ed99f91","Type":"ContainerDied","Data":"72b068529352cd2fd0d98380f92cbe1ad287be50708214d7b3205789afcc691d"} Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.498628 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="72b068529352cd2fd0d98380f92cbe1ad287be50708214d7b3205789afcc691d" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.498678 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-91fe-account-create-update-4xgv2" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.502947 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/25ee21eb-c0b8-4d49-8906-4e4144abf6fc-combined-ca-bundle\") pod \"heat-cfnapi-667b46bf4d-5h2cq\" (UID: \"25ee21eb-c0b8-4d49-8906-4e4144abf6fc\") " pod="openstack/heat-cfnapi-667b46bf4d-5h2cq" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.503026 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/25ee21eb-c0b8-4d49-8906-4e4144abf6fc-config-data\") pod \"heat-cfnapi-667b46bf4d-5h2cq\" (UID: \"25ee21eb-c0b8-4d49-8906-4e4144abf6fc\") " pod="openstack/heat-cfnapi-667b46bf4d-5h2cq" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.503051 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8d66537-20e1-4f87-9077-484e264bae81-combined-ca-bundle\") pod \"heat-engine-656b845fb4-qst2s\" (UID: \"d8d66537-20e1-4f87-9077-484e264bae81\") " pod="openstack/heat-engine-656b845fb4-qst2s" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.503087 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kxlch\" (UniqueName: \"kubernetes.io/projected/25ee21eb-c0b8-4d49-8906-4e4144abf6fc-kube-api-access-kxlch\") pod \"heat-cfnapi-667b46bf4d-5h2cq\" (UID: \"25ee21eb-c0b8-4d49-8906-4e4144abf6fc\") " pod="openstack/heat-cfnapi-667b46bf4d-5h2cq" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.503110 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d8d66537-20e1-4f87-9077-484e264bae81-config-data\") pod \"heat-engine-656b845fb4-qst2s\" (UID: \"d8d66537-20e1-4f87-9077-484e264bae81\") " pod="openstack/heat-engine-656b845fb4-qst2s" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.503158 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/25ee21eb-c0b8-4d49-8906-4e4144abf6fc-config-data-custom\") pod \"heat-cfnapi-667b46bf4d-5h2cq\" (UID: \"25ee21eb-c0b8-4d49-8906-4e4144abf6fc\") " pod="openstack/heat-cfnapi-667b46bf4d-5h2cq" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.503178 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rhf6r\" (UniqueName: \"kubernetes.io/projected/d8d66537-20e1-4f87-9077-484e264bae81-kube-api-access-rhf6r\") pod \"heat-engine-656b845fb4-qst2s\" (UID: \"d8d66537-20e1-4f87-9077-484e264bae81\") " pod="openstack/heat-engine-656b845fb4-qst2s" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.503210 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d8d66537-20e1-4f87-9077-484e264bae81-config-data-custom\") pod \"heat-engine-656b845fb4-qst2s\" (UID: \"d8d66537-20e1-4f87-9077-484e264bae81\") " pod="openstack/heat-engine-656b845fb4-qst2s" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.507156 4824 generic.go:334] "Generic (PLEG): container finished" podID="12776f5e-c372-41bd-91fe-ba7d0915050a" containerID="9f11951e86971107f0a8d485c5dead64b960ce2e124caba92b7e241dae731aca" exitCode=143 Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.507229 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"12776f5e-c372-41bd-91fe-ba7d0915050a","Type":"ContainerDied","Data":"9f11951e86971107f0a8d485c5dead64b960ce2e124caba92b7e241dae731aca"} Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.511465 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-f2xfg" event={"ID":"12e7657f-78fe-4b5c-b614-4449e94366b9","Type":"ContainerDied","Data":"62ab48f611e01d791ea972b5bd6b22f81360725b9ea98ca5dc5298079b346589"} Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.511548 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-f2xfg" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.511487 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="62ab48f611e01d791ea972b5bd6b22f81360725b9ea98ca5dc5298079b346589" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.521883 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7756b9d78c-ftf8g"] Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.545555 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-api-5dc8c6bf78-szsb6"] Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.550665 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-5dc8c6bf78-szsb6" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.555223 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-api-config-data" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.565629 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=3.565619659 podStartE2EDuration="3.565619659s" podCreationTimestamp="2026-01-21 11:27:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:27:32.539649634 +0000 UTC m=+994.832678926" watchObservedRunningTime="2026-01-21 11:27:32.565619659 +0000 UTC m=+994.858648950" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.580673 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-5dc8c6bf78-szsb6"] Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.604711 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f2ec0128-5881-4c67-8d78-d53d8be3a2d3-config\") pod \"dnsmasq-dns-7756b9d78c-ftf8g\" (UID: \"f2ec0128-5881-4c67-8d78-d53d8be3a2d3\") " pod="openstack/dnsmasq-dns-7756b9d78c-ftf8g" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.604739 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2aef2f5b-bb6b-41dd-a3eb-372e7e777122-config-data-custom\") pod \"heat-api-5dc8c6bf78-szsb6\" (UID: \"2aef2f5b-bb6b-41dd-a3eb-372e7e777122\") " pod="openstack/heat-api-5dc8c6bf78-szsb6" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.604765 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2aef2f5b-bb6b-41dd-a3eb-372e7e777122-combined-ca-bundle\") pod \"heat-api-5dc8c6bf78-szsb6\" (UID: \"2aef2f5b-bb6b-41dd-a3eb-372e7e777122\") " pod="openstack/heat-api-5dc8c6bf78-szsb6" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.604798 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2aef2f5b-bb6b-41dd-a3eb-372e7e777122-config-data\") pod \"heat-api-5dc8c6bf78-szsb6\" (UID: \"2aef2f5b-bb6b-41dd-a3eb-372e7e777122\") " pod="openstack/heat-api-5dc8c6bf78-szsb6" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.604827 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dkztw\" (UniqueName: \"kubernetes.io/projected/2aef2f5b-bb6b-41dd-a3eb-372e7e777122-kube-api-access-dkztw\") pod \"heat-api-5dc8c6bf78-szsb6\" (UID: \"2aef2f5b-bb6b-41dd-a3eb-372e7e777122\") " pod="openstack/heat-api-5dc8c6bf78-szsb6" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.604866 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hljdx\" (UniqueName: \"kubernetes.io/projected/f2ec0128-5881-4c67-8d78-d53d8be3a2d3-kube-api-access-hljdx\") pod \"dnsmasq-dns-7756b9d78c-ftf8g\" (UID: \"f2ec0128-5881-4c67-8d78-d53d8be3a2d3\") " pod="openstack/dnsmasq-dns-7756b9d78c-ftf8g" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.604890 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f2ec0128-5881-4c67-8d78-d53d8be3a2d3-ovsdbserver-sb\") pod \"dnsmasq-dns-7756b9d78c-ftf8g\" (UID: \"f2ec0128-5881-4c67-8d78-d53d8be3a2d3\") " pod="openstack/dnsmasq-dns-7756b9d78c-ftf8g" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.604947 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/25ee21eb-c0b8-4d49-8906-4e4144abf6fc-combined-ca-bundle\") pod \"heat-cfnapi-667b46bf4d-5h2cq\" (UID: \"25ee21eb-c0b8-4d49-8906-4e4144abf6fc\") " pod="openstack/heat-cfnapi-667b46bf4d-5h2cq" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.605014 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/25ee21eb-c0b8-4d49-8906-4e4144abf6fc-config-data\") pod \"heat-cfnapi-667b46bf4d-5h2cq\" (UID: \"25ee21eb-c0b8-4d49-8906-4e4144abf6fc\") " pod="openstack/heat-cfnapi-667b46bf4d-5h2cq" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.605036 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8d66537-20e1-4f87-9077-484e264bae81-combined-ca-bundle\") pod \"heat-engine-656b845fb4-qst2s\" (UID: \"d8d66537-20e1-4f87-9077-484e264bae81\") " pod="openstack/heat-engine-656b845fb4-qst2s" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.605173 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f2ec0128-5881-4c67-8d78-d53d8be3a2d3-dns-svc\") pod \"dnsmasq-dns-7756b9d78c-ftf8g\" (UID: \"f2ec0128-5881-4c67-8d78-d53d8be3a2d3\") " pod="openstack/dnsmasq-dns-7756b9d78c-ftf8g" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.605263 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kxlch\" (UniqueName: \"kubernetes.io/projected/25ee21eb-c0b8-4d49-8906-4e4144abf6fc-kube-api-access-kxlch\") pod \"heat-cfnapi-667b46bf4d-5h2cq\" (UID: \"25ee21eb-c0b8-4d49-8906-4e4144abf6fc\") " pod="openstack/heat-cfnapi-667b46bf4d-5h2cq" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.605321 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d8d66537-20e1-4f87-9077-484e264bae81-config-data\") pod \"heat-engine-656b845fb4-qst2s\" (UID: \"d8d66537-20e1-4f87-9077-484e264bae81\") " pod="openstack/heat-engine-656b845fb4-qst2s" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.605356 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f2ec0128-5881-4c67-8d78-d53d8be3a2d3-ovsdbserver-nb\") pod \"dnsmasq-dns-7756b9d78c-ftf8g\" (UID: \"f2ec0128-5881-4c67-8d78-d53d8be3a2d3\") " pod="openstack/dnsmasq-dns-7756b9d78c-ftf8g" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.605471 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f2ec0128-5881-4c67-8d78-d53d8be3a2d3-dns-swift-storage-0\") pod \"dnsmasq-dns-7756b9d78c-ftf8g\" (UID: \"f2ec0128-5881-4c67-8d78-d53d8be3a2d3\") " pod="openstack/dnsmasq-dns-7756b9d78c-ftf8g" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.605493 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/25ee21eb-c0b8-4d49-8906-4e4144abf6fc-config-data-custom\") pod \"heat-cfnapi-667b46bf4d-5h2cq\" (UID: \"25ee21eb-c0b8-4d49-8906-4e4144abf6fc\") " pod="openstack/heat-cfnapi-667b46bf4d-5h2cq" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.605523 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rhf6r\" (UniqueName: \"kubernetes.io/projected/d8d66537-20e1-4f87-9077-484e264bae81-kube-api-access-rhf6r\") pod \"heat-engine-656b845fb4-qst2s\" (UID: \"d8d66537-20e1-4f87-9077-484e264bae81\") " pod="openstack/heat-engine-656b845fb4-qst2s" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.605577 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d8d66537-20e1-4f87-9077-484e264bae81-config-data-custom\") pod \"heat-engine-656b845fb4-qst2s\" (UID: \"d8d66537-20e1-4f87-9077-484e264bae81\") " pod="openstack/heat-engine-656b845fb4-qst2s" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.624693 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d8d66537-20e1-4f87-9077-484e264bae81-config-data-custom\") pod \"heat-engine-656b845fb4-qst2s\" (UID: \"d8d66537-20e1-4f87-9077-484e264bae81\") " pod="openstack/heat-engine-656b845fb4-qst2s" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.633565 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/25ee21eb-c0b8-4d49-8906-4e4144abf6fc-config-data-custom\") pod \"heat-cfnapi-667b46bf4d-5h2cq\" (UID: \"25ee21eb-c0b8-4d49-8906-4e4144abf6fc\") " pod="openstack/heat-cfnapi-667b46bf4d-5h2cq" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.640619 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/25ee21eb-c0b8-4d49-8906-4e4144abf6fc-combined-ca-bundle\") pod \"heat-cfnapi-667b46bf4d-5h2cq\" (UID: \"25ee21eb-c0b8-4d49-8906-4e4144abf6fc\") " pod="openstack/heat-cfnapi-667b46bf4d-5h2cq" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.641012 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d8d66537-20e1-4f87-9077-484e264bae81-config-data\") pod \"heat-engine-656b845fb4-qst2s\" (UID: \"d8d66537-20e1-4f87-9077-484e264bae81\") " pod="openstack/heat-engine-656b845fb4-qst2s" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.641240 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kxlch\" (UniqueName: \"kubernetes.io/projected/25ee21eb-c0b8-4d49-8906-4e4144abf6fc-kube-api-access-kxlch\") pod \"heat-cfnapi-667b46bf4d-5h2cq\" (UID: \"25ee21eb-c0b8-4d49-8906-4e4144abf6fc\") " pod="openstack/heat-cfnapi-667b46bf4d-5h2cq" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.644586 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8d66537-20e1-4f87-9077-484e264bae81-combined-ca-bundle\") pod \"heat-engine-656b845fb4-qst2s\" (UID: \"d8d66537-20e1-4f87-9077-484e264bae81\") " pod="openstack/heat-engine-656b845fb4-qst2s" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.646533 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rhf6r\" (UniqueName: \"kubernetes.io/projected/d8d66537-20e1-4f87-9077-484e264bae81-kube-api-access-rhf6r\") pod \"heat-engine-656b845fb4-qst2s\" (UID: \"d8d66537-20e1-4f87-9077-484e264bae81\") " pod="openstack/heat-engine-656b845fb4-qst2s" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.674796 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/25ee21eb-c0b8-4d49-8906-4e4144abf6fc-config-data\") pod \"heat-cfnapi-667b46bf4d-5h2cq\" (UID: \"25ee21eb-c0b8-4d49-8906-4e4144abf6fc\") " pod="openstack/heat-cfnapi-667b46bf4d-5h2cq" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.685371 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-656b845fb4-qst2s" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.710015 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f2ec0128-5881-4c67-8d78-d53d8be3a2d3-config\") pod \"dnsmasq-dns-7756b9d78c-ftf8g\" (UID: \"f2ec0128-5881-4c67-8d78-d53d8be3a2d3\") " pod="openstack/dnsmasq-dns-7756b9d78c-ftf8g" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.710056 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2aef2f5b-bb6b-41dd-a3eb-372e7e777122-config-data-custom\") pod \"heat-api-5dc8c6bf78-szsb6\" (UID: \"2aef2f5b-bb6b-41dd-a3eb-372e7e777122\") " pod="openstack/heat-api-5dc8c6bf78-szsb6" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.710072 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2aef2f5b-bb6b-41dd-a3eb-372e7e777122-combined-ca-bundle\") pod \"heat-api-5dc8c6bf78-szsb6\" (UID: \"2aef2f5b-bb6b-41dd-a3eb-372e7e777122\") " pod="openstack/heat-api-5dc8c6bf78-szsb6" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.710114 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2aef2f5b-bb6b-41dd-a3eb-372e7e777122-config-data\") pod \"heat-api-5dc8c6bf78-szsb6\" (UID: \"2aef2f5b-bb6b-41dd-a3eb-372e7e777122\") " pod="openstack/heat-api-5dc8c6bf78-szsb6" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.710156 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dkztw\" (UniqueName: \"kubernetes.io/projected/2aef2f5b-bb6b-41dd-a3eb-372e7e777122-kube-api-access-dkztw\") pod \"heat-api-5dc8c6bf78-szsb6\" (UID: \"2aef2f5b-bb6b-41dd-a3eb-372e7e777122\") " pod="openstack/heat-api-5dc8c6bf78-szsb6" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.710191 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hljdx\" (UniqueName: \"kubernetes.io/projected/f2ec0128-5881-4c67-8d78-d53d8be3a2d3-kube-api-access-hljdx\") pod \"dnsmasq-dns-7756b9d78c-ftf8g\" (UID: \"f2ec0128-5881-4c67-8d78-d53d8be3a2d3\") " pod="openstack/dnsmasq-dns-7756b9d78c-ftf8g" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.710209 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f2ec0128-5881-4c67-8d78-d53d8be3a2d3-ovsdbserver-sb\") pod \"dnsmasq-dns-7756b9d78c-ftf8g\" (UID: \"f2ec0128-5881-4c67-8d78-d53d8be3a2d3\") " pod="openstack/dnsmasq-dns-7756b9d78c-ftf8g" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.710310 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f2ec0128-5881-4c67-8d78-d53d8be3a2d3-dns-svc\") pod \"dnsmasq-dns-7756b9d78c-ftf8g\" (UID: \"f2ec0128-5881-4c67-8d78-d53d8be3a2d3\") " pod="openstack/dnsmasq-dns-7756b9d78c-ftf8g" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.710356 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f2ec0128-5881-4c67-8d78-d53d8be3a2d3-ovsdbserver-nb\") pod \"dnsmasq-dns-7756b9d78c-ftf8g\" (UID: \"f2ec0128-5881-4c67-8d78-d53d8be3a2d3\") " pod="openstack/dnsmasq-dns-7756b9d78c-ftf8g" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.710405 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f2ec0128-5881-4c67-8d78-d53d8be3a2d3-dns-swift-storage-0\") pod \"dnsmasq-dns-7756b9d78c-ftf8g\" (UID: \"f2ec0128-5881-4c67-8d78-d53d8be3a2d3\") " pod="openstack/dnsmasq-dns-7756b9d78c-ftf8g" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.710933 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f2ec0128-5881-4c67-8d78-d53d8be3a2d3-config\") pod \"dnsmasq-dns-7756b9d78c-ftf8g\" (UID: \"f2ec0128-5881-4c67-8d78-d53d8be3a2d3\") " pod="openstack/dnsmasq-dns-7756b9d78c-ftf8g" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.711136 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f2ec0128-5881-4c67-8d78-d53d8be3a2d3-dns-swift-storage-0\") pod \"dnsmasq-dns-7756b9d78c-ftf8g\" (UID: \"f2ec0128-5881-4c67-8d78-d53d8be3a2d3\") " pod="openstack/dnsmasq-dns-7756b9d78c-ftf8g" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.711844 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f2ec0128-5881-4c67-8d78-d53d8be3a2d3-ovsdbserver-sb\") pod \"dnsmasq-dns-7756b9d78c-ftf8g\" (UID: \"f2ec0128-5881-4c67-8d78-d53d8be3a2d3\") " pod="openstack/dnsmasq-dns-7756b9d78c-ftf8g" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.712379 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f2ec0128-5881-4c67-8d78-d53d8be3a2d3-dns-svc\") pod \"dnsmasq-dns-7756b9d78c-ftf8g\" (UID: \"f2ec0128-5881-4c67-8d78-d53d8be3a2d3\") " pod="openstack/dnsmasq-dns-7756b9d78c-ftf8g" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.712571 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f2ec0128-5881-4c67-8d78-d53d8be3a2d3-ovsdbserver-nb\") pod \"dnsmasq-dns-7756b9d78c-ftf8g\" (UID: \"f2ec0128-5881-4c67-8d78-d53d8be3a2d3\") " pod="openstack/dnsmasq-dns-7756b9d78c-ftf8g" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.716679 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2aef2f5b-bb6b-41dd-a3eb-372e7e777122-combined-ca-bundle\") pod \"heat-api-5dc8c6bf78-szsb6\" (UID: \"2aef2f5b-bb6b-41dd-a3eb-372e7e777122\") " pod="openstack/heat-api-5dc8c6bf78-szsb6" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.716843 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2aef2f5b-bb6b-41dd-a3eb-372e7e777122-config-data-custom\") pod \"heat-api-5dc8c6bf78-szsb6\" (UID: \"2aef2f5b-bb6b-41dd-a3eb-372e7e777122\") " pod="openstack/heat-api-5dc8c6bf78-szsb6" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.719309 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2aef2f5b-bb6b-41dd-a3eb-372e7e777122-config-data\") pod \"heat-api-5dc8c6bf78-szsb6\" (UID: \"2aef2f5b-bb6b-41dd-a3eb-372e7e777122\") " pod="openstack/heat-api-5dc8c6bf78-szsb6" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.732090 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hljdx\" (UniqueName: \"kubernetes.io/projected/f2ec0128-5881-4c67-8d78-d53d8be3a2d3-kube-api-access-hljdx\") pod \"dnsmasq-dns-7756b9d78c-ftf8g\" (UID: \"f2ec0128-5881-4c67-8d78-d53d8be3a2d3\") " pod="openstack/dnsmasq-dns-7756b9d78c-ftf8g" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.737118 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dkztw\" (UniqueName: \"kubernetes.io/projected/2aef2f5b-bb6b-41dd-a3eb-372e7e777122-kube-api-access-dkztw\") pod \"heat-api-5dc8c6bf78-szsb6\" (UID: \"2aef2f5b-bb6b-41dd-a3eb-372e7e777122\") " pod="openstack/heat-api-5dc8c6bf78-szsb6" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.739930 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-667b46bf4d-5h2cq" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.906382 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7756b9d78c-ftf8g" Jan 21 11:27:32 crc kubenswrapper[4824]: I0121 11:27:32.911306 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-5dc8c6bf78-szsb6" Jan 21 11:27:33 crc kubenswrapper[4824]: I0121 11:27:33.163438 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-656b845fb4-qst2s"] Jan 21 11:27:33 crc kubenswrapper[4824]: W0121 11:27:33.170199 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd8d66537_20e1_4f87_9077_484e264bae81.slice/crio-6515b7c77086397376de7620cef207d703b735c2b9d2c10ad1793aedbc3eb3b1 WatchSource:0}: Error finding container 6515b7c77086397376de7620cef207d703b735c2b9d2c10ad1793aedbc3eb3b1: Status 404 returned error can't find the container with id 6515b7c77086397376de7620cef207d703b735c2b9d2c10ad1793aedbc3eb3b1 Jan 21 11:27:33 crc kubenswrapper[4824]: W0121 11:27:33.305167 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod25ee21eb_c0b8_4d49_8906_4e4144abf6fc.slice/crio-3cb46ceb8d95d8862082f8003345a2758b9fe77a19473ef416fddbd583e50625 WatchSource:0}: Error finding container 3cb46ceb8d95d8862082f8003345a2758b9fe77a19473ef416fddbd583e50625: Status 404 returned error can't find the container with id 3cb46ceb8d95d8862082f8003345a2758b9fe77a19473ef416fddbd583e50625 Jan 21 11:27:33 crc kubenswrapper[4824]: I0121 11:27:33.305193 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-667b46bf4d-5h2cq"] Jan 21 11:27:33 crc kubenswrapper[4824]: I0121 11:27:33.428701 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-5dc8c6bf78-szsb6"] Jan 21 11:27:33 crc kubenswrapper[4824]: W0121 11:27:33.431985 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2aef2f5b_bb6b_41dd_a3eb_372e7e777122.slice/crio-986ea68f679368ee6e505b19ac9e641262bd881fd3d310847834bbf4f66d444f WatchSource:0}: Error finding container 986ea68f679368ee6e505b19ac9e641262bd881fd3d310847834bbf4f66d444f: Status 404 returned error can't find the container with id 986ea68f679368ee6e505b19ac9e641262bd881fd3d310847834bbf4f66d444f Jan 21 11:27:33 crc kubenswrapper[4824]: I0121 11:27:33.445607 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7756b9d78c-ftf8g"] Jan 21 11:27:33 crc kubenswrapper[4824]: W0121 11:27:33.448854 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf2ec0128_5881_4c67_8d78_d53d8be3a2d3.slice/crio-76a9ebfca2b2e541f5b20e22a77a9c3eda4db32f3628a468b328bf434a816a25 WatchSource:0}: Error finding container 76a9ebfca2b2e541f5b20e22a77a9c3eda4db32f3628a468b328bf434a816a25: Status 404 returned error can't find the container with id 76a9ebfca2b2e541f5b20e22a77a9c3eda4db32f3628a468b328bf434a816a25 Jan 21 11:27:33 crc kubenswrapper[4824]: I0121 11:27:33.533422 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-5dc8c6bf78-szsb6" event={"ID":"2aef2f5b-bb6b-41dd-a3eb-372e7e777122","Type":"ContainerStarted","Data":"986ea68f679368ee6e505b19ac9e641262bd881fd3d310847834bbf4f66d444f"} Jan 21 11:27:33 crc kubenswrapper[4824]: I0121 11:27:33.535180 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7756b9d78c-ftf8g" event={"ID":"f2ec0128-5881-4c67-8d78-d53d8be3a2d3","Type":"ContainerStarted","Data":"76a9ebfca2b2e541f5b20e22a77a9c3eda4db32f3628a468b328bf434a816a25"} Jan 21 11:27:33 crc kubenswrapper[4824]: I0121 11:27:33.536934 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a89ff227-6009-4377-8f9f-34ac4078646f","Type":"ContainerStarted","Data":"af5d7d53713185f666f7039b6b0241d8d3bcf7b749f6427dd3855134753e06ca"} Jan 21 11:27:33 crc kubenswrapper[4824]: I0121 11:27:33.553092 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-656b845fb4-qst2s" event={"ID":"d8d66537-20e1-4f87-9077-484e264bae81","Type":"ContainerStarted","Data":"05efb240856120694f40b5baf0ccbaf59da69510aa66da3d1ac87094a30bb9ea"} Jan 21 11:27:33 crc kubenswrapper[4824]: I0121 11:27:33.553145 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-engine-656b845fb4-qst2s" Jan 21 11:27:33 crc kubenswrapper[4824]: I0121 11:27:33.553164 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-656b845fb4-qst2s" event={"ID":"d8d66537-20e1-4f87-9077-484e264bae81","Type":"ContainerStarted","Data":"6515b7c77086397376de7620cef207d703b735c2b9d2c10ad1793aedbc3eb3b1"} Jan 21 11:27:33 crc kubenswrapper[4824]: I0121 11:27:33.553197 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-667b46bf4d-5h2cq" event={"ID":"25ee21eb-c0b8-4d49-8906-4e4144abf6fc","Type":"ContainerStarted","Data":"3cb46ceb8d95d8862082f8003345a2758b9fe77a19473ef416fddbd583e50625"} Jan 21 11:27:33 crc kubenswrapper[4824]: I0121 11:27:33.560991 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-engine-656b845fb4-qst2s" podStartSLOduration=1.5609775209999999 podStartE2EDuration="1.560977521s" podCreationTimestamp="2026-01-21 11:27:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:27:33.557485154 +0000 UTC m=+995.850514446" watchObservedRunningTime="2026-01-21 11:27:33.560977521 +0000 UTC m=+995.854006813" Jan 21 11:27:34 crc kubenswrapper[4824]: I0121 11:27:34.110666 4824 pod_container_manager_linux.go:210] "Failed to delete cgroup paths" cgroupName=["kubepods","besteffort","pod62025fca-0943-48d2-b139-c55b0c7df358"] err="unable to destroy cgroup paths for cgroup [kubepods besteffort pod62025fca-0943-48d2-b139-c55b0c7df358] : Timed out while waiting for systemd to remove kubepods-besteffort-pod62025fca_0943_48d2_b139_c55b0c7df358.slice" Jan 21 11:27:34 crc kubenswrapper[4824]: I0121 11:27:34.553462 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a89ff227-6009-4377-8f9f-34ac4078646f","Type":"ContainerStarted","Data":"184b945bcbd88dab106672c9b74811f2027be88ca7953a946187d4d4ab09c6b1"} Jan 21 11:27:34 crc kubenswrapper[4824]: I0121 11:27:34.553530 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Jan 21 11:27:34 crc kubenswrapper[4824]: I0121 11:27:34.556409 4824 generic.go:334] "Generic (PLEG): container finished" podID="f2ec0128-5881-4c67-8d78-d53d8be3a2d3" containerID="77516fe4a72158d203b645cac966a850920a66f17eb3fb68fcdc002cb9e2b607" exitCode=0 Jan 21 11:27:34 crc kubenswrapper[4824]: I0121 11:27:34.557261 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7756b9d78c-ftf8g" event={"ID":"f2ec0128-5881-4c67-8d78-d53d8be3a2d3","Type":"ContainerDied","Data":"77516fe4a72158d203b645cac966a850920a66f17eb3fb68fcdc002cb9e2b607"} Jan 21 11:27:34 crc kubenswrapper[4824]: I0121 11:27:34.568744 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.178635741 podStartE2EDuration="5.568731996s" podCreationTimestamp="2026-01-21 11:27:29 +0000 UTC" firstStartedPulling="2026-01-21 11:27:30.623235249 +0000 UTC m=+992.916264541" lastFinishedPulling="2026-01-21 11:27:34.013331504 +0000 UTC m=+996.306360796" observedRunningTime="2026-01-21 11:27:34.567056976 +0000 UTC m=+996.860086268" watchObservedRunningTime="2026-01-21 11:27:34.568731996 +0000 UTC m=+996.861761288" Jan 21 11:27:35 crc kubenswrapper[4824]: I0121 11:27:35.541653 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Jan 21 11:27:35 crc kubenswrapper[4824]: I0121 11:27:35.565021 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-667b46bf4d-5h2cq" event={"ID":"25ee21eb-c0b8-4d49-8906-4e4144abf6fc","Type":"ContainerStarted","Data":"a355cd0ecd36eae3a40739bece4826924a58c416690883d3d7de3207672df97c"} Jan 21 11:27:35 crc kubenswrapper[4824]: I0121 11:27:35.566134 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-cfnapi-667b46bf4d-5h2cq" Jan 21 11:27:35 crc kubenswrapper[4824]: I0121 11:27:35.570326 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-5dc8c6bf78-szsb6" event={"ID":"2aef2f5b-bb6b-41dd-a3eb-372e7e777122","Type":"ContainerStarted","Data":"494c1e9ccd78616e73e2a0a9317748e295b50e43bcd2bc2b2bf79a00562dbea9"} Jan 21 11:27:35 crc kubenswrapper[4824]: I0121 11:27:35.570403 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-api-5dc8c6bf78-szsb6" Jan 21 11:27:35 crc kubenswrapper[4824]: I0121 11:27:35.571903 4824 generic.go:334] "Generic (PLEG): container finished" podID="12776f5e-c372-41bd-91fe-ba7d0915050a" containerID="c66cf375354a975633a38f62d2bcee1e315728b454ed407911dcd2b8fbd32686" exitCode=0 Jan 21 11:27:35 crc kubenswrapper[4824]: I0121 11:27:35.571947 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Jan 21 11:27:35 crc kubenswrapper[4824]: I0121 11:27:35.571984 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"12776f5e-c372-41bd-91fe-ba7d0915050a","Type":"ContainerDied","Data":"c66cf375354a975633a38f62d2bcee1e315728b454ed407911dcd2b8fbd32686"} Jan 21 11:27:35 crc kubenswrapper[4824]: I0121 11:27:35.572006 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"12776f5e-c372-41bd-91fe-ba7d0915050a","Type":"ContainerDied","Data":"eaaf72feb767aa6ebf839a27a99ea70de0682afc00ba0ae8e8e10dabfcaf1864"} Jan 21 11:27:35 crc kubenswrapper[4824]: I0121 11:27:35.572023 4824 scope.go:117] "RemoveContainer" containerID="c66cf375354a975633a38f62d2bcee1e315728b454ed407911dcd2b8fbd32686" Jan 21 11:27:35 crc kubenswrapper[4824]: I0121 11:27:35.577472 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7756b9d78c-ftf8g" event={"ID":"f2ec0128-5881-4c67-8d78-d53d8be3a2d3","Type":"ContainerStarted","Data":"18927e4fff1f2c120dbfd5a267daf57c4c84dbc311e5fd297d4153df2a6f21f4"} Jan 21 11:27:35 crc kubenswrapper[4824]: I0121 11:27:35.577584 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7756b9d78c-ftf8g" Jan 21 11:27:35 crc kubenswrapper[4824]: I0121 11:27:35.595134 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-cfnapi-667b46bf4d-5h2cq" podStartSLOduration=1.795267887 podStartE2EDuration="3.595116211s" podCreationTimestamp="2026-01-21 11:27:32 +0000 UTC" firstStartedPulling="2026-01-21 11:27:33.307449048 +0000 UTC m=+995.600478341" lastFinishedPulling="2026-01-21 11:27:35.107297373 +0000 UTC m=+997.400326665" observedRunningTime="2026-01-21 11:27:35.584158753 +0000 UTC m=+997.877188044" watchObservedRunningTime="2026-01-21 11:27:35.595116211 +0000 UTC m=+997.888145503" Jan 21 11:27:35 crc kubenswrapper[4824]: I0121 11:27:35.596495 4824 scope.go:117] "RemoveContainer" containerID="9f11951e86971107f0a8d485c5dead64b960ce2e124caba92b7e241dae731aca" Jan 21 11:27:35 crc kubenswrapper[4824]: I0121 11:27:35.614560 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-api-5dc8c6bf78-szsb6" podStartSLOduration=1.941548805 podStartE2EDuration="3.614545218s" podCreationTimestamp="2026-01-21 11:27:32 +0000 UTC" firstStartedPulling="2026-01-21 11:27:33.433808221 +0000 UTC m=+995.726837512" lastFinishedPulling="2026-01-21 11:27:35.106804634 +0000 UTC m=+997.399833925" observedRunningTime="2026-01-21 11:27:35.610138366 +0000 UTC m=+997.903167658" watchObservedRunningTime="2026-01-21 11:27:35.614545218 +0000 UTC m=+997.907574510" Jan 21 11:27:35 crc kubenswrapper[4824]: I0121 11:27:35.641008 4824 scope.go:117] "RemoveContainer" containerID="c66cf375354a975633a38f62d2bcee1e315728b454ed407911dcd2b8fbd32686" Jan 21 11:27:35 crc kubenswrapper[4824]: E0121 11:27:35.642537 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c66cf375354a975633a38f62d2bcee1e315728b454ed407911dcd2b8fbd32686\": container with ID starting with c66cf375354a975633a38f62d2bcee1e315728b454ed407911dcd2b8fbd32686 not found: ID does not exist" containerID="c66cf375354a975633a38f62d2bcee1e315728b454ed407911dcd2b8fbd32686" Jan 21 11:27:35 crc kubenswrapper[4824]: I0121 11:27:35.642566 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c66cf375354a975633a38f62d2bcee1e315728b454ed407911dcd2b8fbd32686"} err="failed to get container status \"c66cf375354a975633a38f62d2bcee1e315728b454ed407911dcd2b8fbd32686\": rpc error: code = NotFound desc = could not find container \"c66cf375354a975633a38f62d2bcee1e315728b454ed407911dcd2b8fbd32686\": container with ID starting with c66cf375354a975633a38f62d2bcee1e315728b454ed407911dcd2b8fbd32686 not found: ID does not exist" Jan 21 11:27:35 crc kubenswrapper[4824]: I0121 11:27:35.642586 4824 scope.go:117] "RemoveContainer" containerID="9f11951e86971107f0a8d485c5dead64b960ce2e124caba92b7e241dae731aca" Jan 21 11:27:35 crc kubenswrapper[4824]: I0121 11:27:35.644793 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-k6zm8"] Jan 21 11:27:35 crc kubenswrapper[4824]: E0121 11:27:35.645150 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="12776f5e-c372-41bd-91fe-ba7d0915050a" containerName="glance-httpd" Jan 21 11:27:35 crc kubenswrapper[4824]: I0121 11:27:35.645162 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="12776f5e-c372-41bd-91fe-ba7d0915050a" containerName="glance-httpd" Jan 21 11:27:35 crc kubenswrapper[4824]: E0121 11:27:35.645176 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="12776f5e-c372-41bd-91fe-ba7d0915050a" containerName="glance-log" Jan 21 11:27:35 crc kubenswrapper[4824]: I0121 11:27:35.645181 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="12776f5e-c372-41bd-91fe-ba7d0915050a" containerName="glance-log" Jan 21 11:27:35 crc kubenswrapper[4824]: I0121 11:27:35.645361 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="12776f5e-c372-41bd-91fe-ba7d0915050a" containerName="glance-log" Jan 21 11:27:35 crc kubenswrapper[4824]: I0121 11:27:35.645382 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="12776f5e-c372-41bd-91fe-ba7d0915050a" containerName="glance-httpd" Jan 21 11:27:35 crc kubenswrapper[4824]: I0121 11:27:35.645466 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7756b9d78c-ftf8g" podStartSLOduration=3.645457456 podStartE2EDuration="3.645457456s" podCreationTimestamp="2026-01-21 11:27:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:27:35.641485193 +0000 UTC m=+997.934514486" watchObservedRunningTime="2026-01-21 11:27:35.645457456 +0000 UTC m=+997.938486747" Jan 21 11:27:35 crc kubenswrapper[4824]: I0121 11:27:35.645858 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-k6zm8" Jan 21 11:27:35 crc kubenswrapper[4824]: E0121 11:27:35.646920 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9f11951e86971107f0a8d485c5dead64b960ce2e124caba92b7e241dae731aca\": container with ID starting with 9f11951e86971107f0a8d485c5dead64b960ce2e124caba92b7e241dae731aca not found: ID does not exist" containerID="9f11951e86971107f0a8d485c5dead64b960ce2e124caba92b7e241dae731aca" Jan 21 11:27:35 crc kubenswrapper[4824]: I0121 11:27:35.646944 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9f11951e86971107f0a8d485c5dead64b960ce2e124caba92b7e241dae731aca"} err="failed to get container status \"9f11951e86971107f0a8d485c5dead64b960ce2e124caba92b7e241dae731aca\": rpc error: code = NotFound desc = could not find container \"9f11951e86971107f0a8d485c5dead64b960ce2e124caba92b7e241dae731aca\": container with ID starting with 9f11951e86971107f0a8d485c5dead64b960ce2e124caba92b7e241dae731aca not found: ID does not exist" Jan 21 11:27:35 crc kubenswrapper[4824]: I0121 11:27:35.647628 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-gbh62" Jan 21 11:27:35 crc kubenswrapper[4824]: I0121 11:27:35.648276 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Jan 21 11:27:35 crc kubenswrapper[4824]: I0121 11:27:35.648302 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Jan 21 11:27:35 crc kubenswrapper[4824]: I0121 11:27:35.662940 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-k6zm8"] Jan 21 11:27:35 crc kubenswrapper[4824]: I0121 11:27:35.680020 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/12776f5e-c372-41bd-91fe-ba7d0915050a-scripts\") pod \"12776f5e-c372-41bd-91fe-ba7d0915050a\" (UID: \"12776f5e-c372-41bd-91fe-ba7d0915050a\") " Jan 21 11:27:35 crc kubenswrapper[4824]: I0121 11:27:35.680053 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7qwlh\" (UniqueName: \"kubernetes.io/projected/12776f5e-c372-41bd-91fe-ba7d0915050a-kube-api-access-7qwlh\") pod \"12776f5e-c372-41bd-91fe-ba7d0915050a\" (UID: \"12776f5e-c372-41bd-91fe-ba7d0915050a\") " Jan 21 11:27:35 crc kubenswrapper[4824]: I0121 11:27:35.680097 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/12776f5e-c372-41bd-91fe-ba7d0915050a-public-tls-certs\") pod \"12776f5e-c372-41bd-91fe-ba7d0915050a\" (UID: \"12776f5e-c372-41bd-91fe-ba7d0915050a\") " Jan 21 11:27:35 crc kubenswrapper[4824]: I0121 11:27:35.680111 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"12776f5e-c372-41bd-91fe-ba7d0915050a\" (UID: \"12776f5e-c372-41bd-91fe-ba7d0915050a\") " Jan 21 11:27:35 crc kubenswrapper[4824]: I0121 11:27:35.680160 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12776f5e-c372-41bd-91fe-ba7d0915050a-combined-ca-bundle\") pod \"12776f5e-c372-41bd-91fe-ba7d0915050a\" (UID: \"12776f5e-c372-41bd-91fe-ba7d0915050a\") " Jan 21 11:27:35 crc kubenswrapper[4824]: I0121 11:27:35.680208 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/12776f5e-c372-41bd-91fe-ba7d0915050a-logs\") pod \"12776f5e-c372-41bd-91fe-ba7d0915050a\" (UID: \"12776f5e-c372-41bd-91fe-ba7d0915050a\") " Jan 21 11:27:35 crc kubenswrapper[4824]: I0121 11:27:35.680230 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/12776f5e-c372-41bd-91fe-ba7d0915050a-config-data\") pod \"12776f5e-c372-41bd-91fe-ba7d0915050a\" (UID: \"12776f5e-c372-41bd-91fe-ba7d0915050a\") " Jan 21 11:27:35 crc kubenswrapper[4824]: I0121 11:27:35.680267 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/12776f5e-c372-41bd-91fe-ba7d0915050a-httpd-run\") pod \"12776f5e-c372-41bd-91fe-ba7d0915050a\" (UID: \"12776f5e-c372-41bd-91fe-ba7d0915050a\") " Jan 21 11:27:35 crc kubenswrapper[4824]: I0121 11:27:35.685921 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/12776f5e-c372-41bd-91fe-ba7d0915050a-scripts" (OuterVolumeSpecName: "scripts") pod "12776f5e-c372-41bd-91fe-ba7d0915050a" (UID: "12776f5e-c372-41bd-91fe-ba7d0915050a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:27:35 crc kubenswrapper[4824]: I0121 11:27:35.686073 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/12776f5e-c372-41bd-91fe-ba7d0915050a-kube-api-access-7qwlh" (OuterVolumeSpecName: "kube-api-access-7qwlh") pod "12776f5e-c372-41bd-91fe-ba7d0915050a" (UID: "12776f5e-c372-41bd-91fe-ba7d0915050a"). InnerVolumeSpecName "kube-api-access-7qwlh". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:27:35 crc kubenswrapper[4824]: I0121 11:27:35.686609 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/12776f5e-c372-41bd-91fe-ba7d0915050a-logs" (OuterVolumeSpecName: "logs") pod "12776f5e-c372-41bd-91fe-ba7d0915050a" (UID: "12776f5e-c372-41bd-91fe-ba7d0915050a"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:27:35 crc kubenswrapper[4824]: I0121 11:27:35.687186 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/12776f5e-c372-41bd-91fe-ba7d0915050a-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "12776f5e-c372-41bd-91fe-ba7d0915050a" (UID: "12776f5e-c372-41bd-91fe-ba7d0915050a"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:27:35 crc kubenswrapper[4824]: I0121 11:27:35.696059 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage06-crc" (OuterVolumeSpecName: "glance") pod "12776f5e-c372-41bd-91fe-ba7d0915050a" (UID: "12776f5e-c372-41bd-91fe-ba7d0915050a"). InnerVolumeSpecName "local-storage06-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Jan 21 11:27:35 crc kubenswrapper[4824]: I0121 11:27:35.751098 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/12776f5e-c372-41bd-91fe-ba7d0915050a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "12776f5e-c372-41bd-91fe-ba7d0915050a" (UID: "12776f5e-c372-41bd-91fe-ba7d0915050a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:27:35 crc kubenswrapper[4824]: I0121 11:27:35.781799 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c73eda8-d376-4e18-9b33-f5ac4d10d091-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-k6zm8\" (UID: \"9c73eda8-d376-4e18-9b33-f5ac4d10d091\") " pod="openstack/nova-cell0-conductor-db-sync-k6zm8" Jan 21 11:27:35 crc kubenswrapper[4824]: I0121 11:27:35.781856 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5rs4n\" (UniqueName: \"kubernetes.io/projected/9c73eda8-d376-4e18-9b33-f5ac4d10d091-kube-api-access-5rs4n\") pod \"nova-cell0-conductor-db-sync-k6zm8\" (UID: \"9c73eda8-d376-4e18-9b33-f5ac4d10d091\") " pod="openstack/nova-cell0-conductor-db-sync-k6zm8" Jan 21 11:27:35 crc kubenswrapper[4824]: I0121 11:27:35.781985 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c73eda8-d376-4e18-9b33-f5ac4d10d091-config-data\") pod \"nova-cell0-conductor-db-sync-k6zm8\" (UID: \"9c73eda8-d376-4e18-9b33-f5ac4d10d091\") " pod="openstack/nova-cell0-conductor-db-sync-k6zm8" Jan 21 11:27:35 crc kubenswrapper[4824]: I0121 11:27:35.782144 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9c73eda8-d376-4e18-9b33-f5ac4d10d091-scripts\") pod \"nova-cell0-conductor-db-sync-k6zm8\" (UID: \"9c73eda8-d376-4e18-9b33-f5ac4d10d091\") " pod="openstack/nova-cell0-conductor-db-sync-k6zm8" Jan 21 11:27:35 crc kubenswrapper[4824]: I0121 11:27:35.782205 4824 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/12776f5e-c372-41bd-91fe-ba7d0915050a-scripts\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:35 crc kubenswrapper[4824]: I0121 11:27:35.782216 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7qwlh\" (UniqueName: \"kubernetes.io/projected/12776f5e-c372-41bd-91fe-ba7d0915050a-kube-api-access-7qwlh\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:35 crc kubenswrapper[4824]: I0121 11:27:35.782233 4824 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" " Jan 21 11:27:35 crc kubenswrapper[4824]: I0121 11:27:35.782242 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12776f5e-c372-41bd-91fe-ba7d0915050a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:35 crc kubenswrapper[4824]: I0121 11:27:35.782251 4824 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/12776f5e-c372-41bd-91fe-ba7d0915050a-logs\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:35 crc kubenswrapper[4824]: I0121 11:27:35.782259 4824 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/12776f5e-c372-41bd-91fe-ba7d0915050a-httpd-run\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:35 crc kubenswrapper[4824]: I0121 11:27:35.801788 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/12776f5e-c372-41bd-91fe-ba7d0915050a-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "12776f5e-c372-41bd-91fe-ba7d0915050a" (UID: "12776f5e-c372-41bd-91fe-ba7d0915050a"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:27:35 crc kubenswrapper[4824]: I0121 11:27:35.821683 4824 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage06-crc" (UniqueName: "kubernetes.io/local-volume/local-storage06-crc") on node "crc" Jan 21 11:27:35 crc kubenswrapper[4824]: I0121 11:27:35.840114 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/12776f5e-c372-41bd-91fe-ba7d0915050a-config-data" (OuterVolumeSpecName: "config-data") pod "12776f5e-c372-41bd-91fe-ba7d0915050a" (UID: "12776f5e-c372-41bd-91fe-ba7d0915050a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:27:35 crc kubenswrapper[4824]: I0121 11:27:35.887862 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c73eda8-d376-4e18-9b33-f5ac4d10d091-config-data\") pod \"nova-cell0-conductor-db-sync-k6zm8\" (UID: \"9c73eda8-d376-4e18-9b33-f5ac4d10d091\") " pod="openstack/nova-cell0-conductor-db-sync-k6zm8" Jan 21 11:27:35 crc kubenswrapper[4824]: I0121 11:27:35.888240 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9c73eda8-d376-4e18-9b33-f5ac4d10d091-scripts\") pod \"nova-cell0-conductor-db-sync-k6zm8\" (UID: \"9c73eda8-d376-4e18-9b33-f5ac4d10d091\") " pod="openstack/nova-cell0-conductor-db-sync-k6zm8" Jan 21 11:27:35 crc kubenswrapper[4824]: I0121 11:27:35.888338 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c73eda8-d376-4e18-9b33-f5ac4d10d091-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-k6zm8\" (UID: \"9c73eda8-d376-4e18-9b33-f5ac4d10d091\") " pod="openstack/nova-cell0-conductor-db-sync-k6zm8" Jan 21 11:27:35 crc kubenswrapper[4824]: I0121 11:27:35.888428 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5rs4n\" (UniqueName: \"kubernetes.io/projected/9c73eda8-d376-4e18-9b33-f5ac4d10d091-kube-api-access-5rs4n\") pod \"nova-cell0-conductor-db-sync-k6zm8\" (UID: \"9c73eda8-d376-4e18-9b33-f5ac4d10d091\") " pod="openstack/nova-cell0-conductor-db-sync-k6zm8" Jan 21 11:27:35 crc kubenswrapper[4824]: I0121 11:27:35.888617 4824 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/12776f5e-c372-41bd-91fe-ba7d0915050a-public-tls-certs\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:35 crc kubenswrapper[4824]: I0121 11:27:35.888686 4824 reconciler_common.go:293] "Volume detached for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:35 crc kubenswrapper[4824]: I0121 11:27:35.888749 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/12776f5e-c372-41bd-91fe-ba7d0915050a-config-data\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:35 crc kubenswrapper[4824]: I0121 11:27:35.892925 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c73eda8-d376-4e18-9b33-f5ac4d10d091-config-data\") pod \"nova-cell0-conductor-db-sync-k6zm8\" (UID: \"9c73eda8-d376-4e18-9b33-f5ac4d10d091\") " pod="openstack/nova-cell0-conductor-db-sync-k6zm8" Jan 21 11:27:35 crc kubenswrapper[4824]: I0121 11:27:35.900087 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9c73eda8-d376-4e18-9b33-f5ac4d10d091-scripts\") pod \"nova-cell0-conductor-db-sync-k6zm8\" (UID: \"9c73eda8-d376-4e18-9b33-f5ac4d10d091\") " pod="openstack/nova-cell0-conductor-db-sync-k6zm8" Jan 21 11:27:35 crc kubenswrapper[4824]: I0121 11:27:35.900367 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c73eda8-d376-4e18-9b33-f5ac4d10d091-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-k6zm8\" (UID: \"9c73eda8-d376-4e18-9b33-f5ac4d10d091\") " pod="openstack/nova-cell0-conductor-db-sync-k6zm8" Jan 21 11:27:35 crc kubenswrapper[4824]: I0121 11:27:35.922542 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5rs4n\" (UniqueName: \"kubernetes.io/projected/9c73eda8-d376-4e18-9b33-f5ac4d10d091-kube-api-access-5rs4n\") pod \"nova-cell0-conductor-db-sync-k6zm8\" (UID: \"9c73eda8-d376-4e18-9b33-f5ac4d10d091\") " pod="openstack/nova-cell0-conductor-db-sync-k6zm8" Jan 21 11:27:35 crc kubenswrapper[4824]: I0121 11:27:35.934448 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Jan 21 11:27:35 crc kubenswrapper[4824]: I0121 11:27:35.944422 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Jan 21 11:27:35 crc kubenswrapper[4824]: I0121 11:27:35.955731 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Jan 21 11:27:35 crc kubenswrapper[4824]: I0121 11:27:35.957053 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Jan 21 11:27:35 crc kubenswrapper[4824]: I0121 11:27:35.959437 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Jan 21 11:27:35 crc kubenswrapper[4824]: I0121 11:27:35.960149 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Jan 21 11:27:35 crc kubenswrapper[4824]: I0121 11:27:35.974909 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-79f4c948b7-4hz4p" Jan 21 11:27:35 crc kubenswrapper[4824]: I0121 11:27:35.975977 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-79f4c948b7-4hz4p" Jan 21 11:27:35 crc kubenswrapper[4824]: I0121 11:27:35.978011 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Jan 21 11:27:36 crc kubenswrapper[4824]: I0121 11:27:36.064023 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="12776f5e-c372-41bd-91fe-ba7d0915050a" path="/var/lib/kubelet/pods/12776f5e-c372-41bd-91fe-ba7d0915050a/volumes" Jan 21 11:27:36 crc kubenswrapper[4824]: I0121 11:27:36.091213 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/219725d0-9bef-4e42-8b78-cbc74a8ac9db-scripts\") pod \"glance-default-external-api-0\" (UID: \"219725d0-9bef-4e42-8b78-cbc74a8ac9db\") " pod="openstack/glance-default-external-api-0" Jan 21 11:27:36 crc kubenswrapper[4824]: I0121 11:27:36.091328 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/219725d0-9bef-4e42-8b78-cbc74a8ac9db-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"219725d0-9bef-4e42-8b78-cbc74a8ac9db\") " pod="openstack/glance-default-external-api-0" Jan 21 11:27:36 crc kubenswrapper[4824]: I0121 11:27:36.091369 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"219725d0-9bef-4e42-8b78-cbc74a8ac9db\") " pod="openstack/glance-default-external-api-0" Jan 21 11:27:36 crc kubenswrapper[4824]: I0121 11:27:36.091402 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/219725d0-9bef-4e42-8b78-cbc74a8ac9db-config-data\") pod \"glance-default-external-api-0\" (UID: \"219725d0-9bef-4e42-8b78-cbc74a8ac9db\") " pod="openstack/glance-default-external-api-0" Jan 21 11:27:36 crc kubenswrapper[4824]: I0121 11:27:36.091429 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ndtmv\" (UniqueName: \"kubernetes.io/projected/219725d0-9bef-4e42-8b78-cbc74a8ac9db-kube-api-access-ndtmv\") pod \"glance-default-external-api-0\" (UID: \"219725d0-9bef-4e42-8b78-cbc74a8ac9db\") " pod="openstack/glance-default-external-api-0" Jan 21 11:27:36 crc kubenswrapper[4824]: I0121 11:27:36.091486 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/219725d0-9bef-4e42-8b78-cbc74a8ac9db-logs\") pod \"glance-default-external-api-0\" (UID: \"219725d0-9bef-4e42-8b78-cbc74a8ac9db\") " pod="openstack/glance-default-external-api-0" Jan 21 11:27:36 crc kubenswrapper[4824]: I0121 11:27:36.091546 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/219725d0-9bef-4e42-8b78-cbc74a8ac9db-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"219725d0-9bef-4e42-8b78-cbc74a8ac9db\") " pod="openstack/glance-default-external-api-0" Jan 21 11:27:36 crc kubenswrapper[4824]: I0121 11:27:36.091562 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/219725d0-9bef-4e42-8b78-cbc74a8ac9db-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"219725d0-9bef-4e42-8b78-cbc74a8ac9db\") " pod="openstack/glance-default-external-api-0" Jan 21 11:27:36 crc kubenswrapper[4824]: I0121 11:27:36.138776 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-k6zm8" Jan 21 11:27:36 crc kubenswrapper[4824]: I0121 11:27:36.193900 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"219725d0-9bef-4e42-8b78-cbc74a8ac9db\") " pod="openstack/glance-default-external-api-0" Jan 21 11:27:36 crc kubenswrapper[4824]: I0121 11:27:36.193997 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/219725d0-9bef-4e42-8b78-cbc74a8ac9db-config-data\") pod \"glance-default-external-api-0\" (UID: \"219725d0-9bef-4e42-8b78-cbc74a8ac9db\") " pod="openstack/glance-default-external-api-0" Jan 21 11:27:36 crc kubenswrapper[4824]: I0121 11:27:36.194034 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ndtmv\" (UniqueName: \"kubernetes.io/projected/219725d0-9bef-4e42-8b78-cbc74a8ac9db-kube-api-access-ndtmv\") pod \"glance-default-external-api-0\" (UID: \"219725d0-9bef-4e42-8b78-cbc74a8ac9db\") " pod="openstack/glance-default-external-api-0" Jan 21 11:27:36 crc kubenswrapper[4824]: I0121 11:27:36.194096 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/219725d0-9bef-4e42-8b78-cbc74a8ac9db-logs\") pod \"glance-default-external-api-0\" (UID: \"219725d0-9bef-4e42-8b78-cbc74a8ac9db\") " pod="openstack/glance-default-external-api-0" Jan 21 11:27:36 crc kubenswrapper[4824]: I0121 11:27:36.194142 4824 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"219725d0-9bef-4e42-8b78-cbc74a8ac9db\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/glance-default-external-api-0" Jan 21 11:27:36 crc kubenswrapper[4824]: I0121 11:27:36.194152 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/219725d0-9bef-4e42-8b78-cbc74a8ac9db-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"219725d0-9bef-4e42-8b78-cbc74a8ac9db\") " pod="openstack/glance-default-external-api-0" Jan 21 11:27:36 crc kubenswrapper[4824]: I0121 11:27:36.197373 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/219725d0-9bef-4e42-8b78-cbc74a8ac9db-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"219725d0-9bef-4e42-8b78-cbc74a8ac9db\") " pod="openstack/glance-default-external-api-0" Jan 21 11:27:36 crc kubenswrapper[4824]: I0121 11:27:36.197487 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/219725d0-9bef-4e42-8b78-cbc74a8ac9db-scripts\") pod \"glance-default-external-api-0\" (UID: \"219725d0-9bef-4e42-8b78-cbc74a8ac9db\") " pod="openstack/glance-default-external-api-0" Jan 21 11:27:36 crc kubenswrapper[4824]: I0121 11:27:36.197615 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/219725d0-9bef-4e42-8b78-cbc74a8ac9db-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"219725d0-9bef-4e42-8b78-cbc74a8ac9db\") " pod="openstack/glance-default-external-api-0" Jan 21 11:27:36 crc kubenswrapper[4824]: I0121 11:27:36.198835 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/219725d0-9bef-4e42-8b78-cbc74a8ac9db-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"219725d0-9bef-4e42-8b78-cbc74a8ac9db\") " pod="openstack/glance-default-external-api-0" Jan 21 11:27:36 crc kubenswrapper[4824]: I0121 11:27:36.199280 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/219725d0-9bef-4e42-8b78-cbc74a8ac9db-logs\") pod \"glance-default-external-api-0\" (UID: \"219725d0-9bef-4e42-8b78-cbc74a8ac9db\") " pod="openstack/glance-default-external-api-0" Jan 21 11:27:36 crc kubenswrapper[4824]: I0121 11:27:36.201739 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/219725d0-9bef-4e42-8b78-cbc74a8ac9db-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"219725d0-9bef-4e42-8b78-cbc74a8ac9db\") " pod="openstack/glance-default-external-api-0" Jan 21 11:27:36 crc kubenswrapper[4824]: I0121 11:27:36.204783 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/219725d0-9bef-4e42-8b78-cbc74a8ac9db-scripts\") pod \"glance-default-external-api-0\" (UID: \"219725d0-9bef-4e42-8b78-cbc74a8ac9db\") " pod="openstack/glance-default-external-api-0" Jan 21 11:27:36 crc kubenswrapper[4824]: I0121 11:27:36.207004 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/219725d0-9bef-4e42-8b78-cbc74a8ac9db-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"219725d0-9bef-4e42-8b78-cbc74a8ac9db\") " pod="openstack/glance-default-external-api-0" Jan 21 11:27:36 crc kubenswrapper[4824]: I0121 11:27:36.204391 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/219725d0-9bef-4e42-8b78-cbc74a8ac9db-config-data\") pod \"glance-default-external-api-0\" (UID: \"219725d0-9bef-4e42-8b78-cbc74a8ac9db\") " pod="openstack/glance-default-external-api-0" Jan 21 11:27:36 crc kubenswrapper[4824]: I0121 11:27:36.216933 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ndtmv\" (UniqueName: \"kubernetes.io/projected/219725d0-9bef-4e42-8b78-cbc74a8ac9db-kube-api-access-ndtmv\") pod \"glance-default-external-api-0\" (UID: \"219725d0-9bef-4e42-8b78-cbc74a8ac9db\") " pod="openstack/glance-default-external-api-0" Jan 21 11:27:36 crc kubenswrapper[4824]: I0121 11:27:36.225911 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"219725d0-9bef-4e42-8b78-cbc74a8ac9db\") " pod="openstack/glance-default-external-api-0" Jan 21 11:27:36 crc kubenswrapper[4824]: I0121 11:27:36.275660 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Jan 21 11:27:36 crc kubenswrapper[4824]: I0121 11:27:36.538140 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-k6zm8"] Jan 21 11:27:36 crc kubenswrapper[4824]: I0121 11:27:36.586803 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-k6zm8" event={"ID":"9c73eda8-d376-4e18-9b33-f5ac4d10d091","Type":"ContainerStarted","Data":"bec80828734b592cc51fee125cf162b17f1a9431a491c60a24c542b64ae9e046"} Jan 21 11:27:36 crc kubenswrapper[4824]: I0121 11:27:36.716408 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Jan 21 11:27:37 crc kubenswrapper[4824]: I0121 11:27:37.595984 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"219725d0-9bef-4e42-8b78-cbc74a8ac9db","Type":"ContainerStarted","Data":"7445af8ee234b9314f9199fd51a796bd220ddda47357119a6f2a23774de972d8"} Jan 21 11:27:37 crc kubenswrapper[4824]: I0121 11:27:37.596439 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"219725d0-9bef-4e42-8b78-cbc74a8ac9db","Type":"ContainerStarted","Data":"6ba7b3bf3ee8993c98598c253f5273552426d07f3b2484dda5a710dacf8f30ac"} Jan 21 11:27:37 crc kubenswrapper[4824]: I0121 11:27:37.596454 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"219725d0-9bef-4e42-8b78-cbc74a8ac9db","Type":"ContainerStarted","Data":"3e89604df6712be074b7a1e1d8001bbe51252f5826bd96b032e5c1523e48ea5d"} Jan 21 11:27:37 crc kubenswrapper[4824]: I0121 11:27:37.611319 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=2.611303981 podStartE2EDuration="2.611303981s" podCreationTimestamp="2026-01-21 11:27:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:27:37.609299842 +0000 UTC m=+999.902329134" watchObservedRunningTime="2026-01-21 11:27:37.611303981 +0000 UTC m=+999.904333274" Jan 21 11:27:38 crc kubenswrapper[4824]: I0121 11:27:38.166890 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Jan 21 11:27:38 crc kubenswrapper[4824]: I0121 11:27:38.167138 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a89ff227-6009-4377-8f9f-34ac4078646f" containerName="ceilometer-central-agent" containerID="cri-o://02e341630b226a52243314bc26a342a8ea843da778071cfc554f827999082ef2" gracePeriod=30 Jan 21 11:27:38 crc kubenswrapper[4824]: I0121 11:27:38.167179 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a89ff227-6009-4377-8f9f-34ac4078646f" containerName="proxy-httpd" containerID="cri-o://184b945bcbd88dab106672c9b74811f2027be88ca7953a946187d4d4ab09c6b1" gracePeriod=30 Jan 21 11:27:38 crc kubenswrapper[4824]: I0121 11:27:38.167205 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a89ff227-6009-4377-8f9f-34ac4078646f" containerName="sg-core" containerID="cri-o://af5d7d53713185f666f7039b6b0241d8d3bcf7b749f6427dd3855134753e06ca" gracePeriod=30 Jan 21 11:27:38 crc kubenswrapper[4824]: I0121 11:27:38.167197 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a89ff227-6009-4377-8f9f-34ac4078646f" containerName="ceilometer-notification-agent" containerID="cri-o://01663fbfa71a8a22bc17e94621a84e2c11185ae78db507b95bd3152b51310836" gracePeriod=30 Jan 21 11:27:38 crc kubenswrapper[4824]: E0121 11:27:38.424602 4824 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda89ff227_6009_4377_8f9f_34ac4078646f.slice/crio-184b945bcbd88dab106672c9b74811f2027be88ca7953a946187d4d4ab09c6b1.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda89ff227_6009_4377_8f9f_34ac4078646f.slice/crio-conmon-184b945bcbd88dab106672c9b74811f2027be88ca7953a946187d4d4ab09c6b1.scope\": RecentStats: unable to find data in memory cache]" Jan 21 11:27:38 crc kubenswrapper[4824]: I0121 11:27:38.606472 4824 generic.go:334] "Generic (PLEG): container finished" podID="a89ff227-6009-4377-8f9f-34ac4078646f" containerID="184b945bcbd88dab106672c9b74811f2027be88ca7953a946187d4d4ab09c6b1" exitCode=0 Jan 21 11:27:38 crc kubenswrapper[4824]: I0121 11:27:38.606501 4824 generic.go:334] "Generic (PLEG): container finished" podID="a89ff227-6009-4377-8f9f-34ac4078646f" containerID="af5d7d53713185f666f7039b6b0241d8d3bcf7b749f6427dd3855134753e06ca" exitCode=2 Jan 21 11:27:38 crc kubenswrapper[4824]: I0121 11:27:38.606554 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a89ff227-6009-4377-8f9f-34ac4078646f","Type":"ContainerDied","Data":"184b945bcbd88dab106672c9b74811f2027be88ca7953a946187d4d4ab09c6b1"} Jan 21 11:27:38 crc kubenswrapper[4824]: I0121 11:27:38.606591 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a89ff227-6009-4377-8f9f-34ac4078646f","Type":"ContainerDied","Data":"af5d7d53713185f666f7039b6b0241d8d3bcf7b749f6427dd3855134753e06ca"} Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.016298 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.067857 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a89ff227-6009-4377-8f9f-34ac4078646f-config-data\") pod \"a89ff227-6009-4377-8f9f-34ac4078646f\" (UID: \"a89ff227-6009-4377-8f9f-34ac4078646f\") " Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.067897 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a89ff227-6009-4377-8f9f-34ac4078646f-sg-core-conf-yaml\") pod \"a89ff227-6009-4377-8f9f-34ac4078646f\" (UID: \"a89ff227-6009-4377-8f9f-34ac4078646f\") " Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.067914 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a89ff227-6009-4377-8f9f-34ac4078646f-combined-ca-bundle\") pod \"a89ff227-6009-4377-8f9f-34ac4078646f\" (UID: \"a89ff227-6009-4377-8f9f-34ac4078646f\") " Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.068034 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a89ff227-6009-4377-8f9f-34ac4078646f-run-httpd\") pod \"a89ff227-6009-4377-8f9f-34ac4078646f\" (UID: \"a89ff227-6009-4377-8f9f-34ac4078646f\") " Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.068064 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-898h7\" (UniqueName: \"kubernetes.io/projected/a89ff227-6009-4377-8f9f-34ac4078646f-kube-api-access-898h7\") pod \"a89ff227-6009-4377-8f9f-34ac4078646f\" (UID: \"a89ff227-6009-4377-8f9f-34ac4078646f\") " Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.068097 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a89ff227-6009-4377-8f9f-34ac4078646f-scripts\") pod \"a89ff227-6009-4377-8f9f-34ac4078646f\" (UID: \"a89ff227-6009-4377-8f9f-34ac4078646f\") " Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.068153 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a89ff227-6009-4377-8f9f-34ac4078646f-log-httpd\") pod \"a89ff227-6009-4377-8f9f-34ac4078646f\" (UID: \"a89ff227-6009-4377-8f9f-34ac4078646f\") " Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.069116 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a89ff227-6009-4377-8f9f-34ac4078646f-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "a89ff227-6009-4377-8f9f-34ac4078646f" (UID: "a89ff227-6009-4377-8f9f-34ac4078646f"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.069147 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a89ff227-6009-4377-8f9f-34ac4078646f-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "a89ff227-6009-4377-8f9f-34ac4078646f" (UID: "a89ff227-6009-4377-8f9f-34ac4078646f"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.072628 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a89ff227-6009-4377-8f9f-34ac4078646f-scripts" (OuterVolumeSpecName: "scripts") pod "a89ff227-6009-4377-8f9f-34ac4078646f" (UID: "a89ff227-6009-4377-8f9f-34ac4078646f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.088549 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a89ff227-6009-4377-8f9f-34ac4078646f-kube-api-access-898h7" (OuterVolumeSpecName: "kube-api-access-898h7") pod "a89ff227-6009-4377-8f9f-34ac4078646f" (UID: "a89ff227-6009-4377-8f9f-34ac4078646f"). InnerVolumeSpecName "kube-api-access-898h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.100739 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a89ff227-6009-4377-8f9f-34ac4078646f-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "a89ff227-6009-4377-8f9f-34ac4078646f" (UID: "a89ff227-6009-4377-8f9f-34ac4078646f"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.128615 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a89ff227-6009-4377-8f9f-34ac4078646f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a89ff227-6009-4377-8f9f-34ac4078646f" (UID: "a89ff227-6009-4377-8f9f-34ac4078646f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.143437 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a89ff227-6009-4377-8f9f-34ac4078646f-config-data" (OuterVolumeSpecName: "config-data") pod "a89ff227-6009-4377-8f9f-34ac4078646f" (UID: "a89ff227-6009-4377-8f9f-34ac4078646f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.170739 4824 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a89ff227-6009-4377-8f9f-34ac4078646f-run-httpd\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.170797 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-898h7\" (UniqueName: \"kubernetes.io/projected/a89ff227-6009-4377-8f9f-34ac4078646f-kube-api-access-898h7\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.170808 4824 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a89ff227-6009-4377-8f9f-34ac4078646f-scripts\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.170816 4824 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a89ff227-6009-4377-8f9f-34ac4078646f-log-httpd\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.170824 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a89ff227-6009-4377-8f9f-34ac4078646f-config-data\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.170831 4824 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a89ff227-6009-4377-8f9f-34ac4078646f-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.170840 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a89ff227-6009-4377-8f9f-34ac4078646f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.366584 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-engine-5bc6fc99dd-nv45v"] Jan 21 11:27:39 crc kubenswrapper[4824]: E0121 11:27:39.366903 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a89ff227-6009-4377-8f9f-34ac4078646f" containerName="ceilometer-central-agent" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.366921 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="a89ff227-6009-4377-8f9f-34ac4078646f" containerName="ceilometer-central-agent" Jan 21 11:27:39 crc kubenswrapper[4824]: E0121 11:27:39.366940 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a89ff227-6009-4377-8f9f-34ac4078646f" containerName="sg-core" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.366946 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="a89ff227-6009-4377-8f9f-34ac4078646f" containerName="sg-core" Jan 21 11:27:39 crc kubenswrapper[4824]: E0121 11:27:39.366986 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a89ff227-6009-4377-8f9f-34ac4078646f" containerName="ceilometer-notification-agent" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.366992 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="a89ff227-6009-4377-8f9f-34ac4078646f" containerName="ceilometer-notification-agent" Jan 21 11:27:39 crc kubenswrapper[4824]: E0121 11:27:39.367004 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a89ff227-6009-4377-8f9f-34ac4078646f" containerName="proxy-httpd" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.367009 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="a89ff227-6009-4377-8f9f-34ac4078646f" containerName="proxy-httpd" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.367151 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="a89ff227-6009-4377-8f9f-34ac4078646f" containerName="proxy-httpd" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.367163 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="a89ff227-6009-4377-8f9f-34ac4078646f" containerName="sg-core" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.367173 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="a89ff227-6009-4377-8f9f-34ac4078646f" containerName="ceilometer-notification-agent" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.367188 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="a89ff227-6009-4377-8f9f-34ac4078646f" containerName="ceilometer-central-agent" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.367662 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-5bc6fc99dd-nv45v" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.382660 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-5bc6fc99dd-nv45v"] Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.401123 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-cfnapi-58f47bdb5-gk4p7"] Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.402180 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-58f47bdb5-gk4p7" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.421003 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-api-7b744b9d97-zkcv4"] Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.427575 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-7b744b9d97-zkcv4" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.451377 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-58f47bdb5-gk4p7"] Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.462265 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-7b744b9d97-zkcv4"] Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.474785 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f5046792-741c-487c-adb8-43502e9fdba1-config-data\") pod \"heat-api-7b744b9d97-zkcv4\" (UID: \"f5046792-741c-487c-adb8-43502e9fdba1\") " pod="openstack/heat-api-7b744b9d97-zkcv4" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.474832 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f5046792-741c-487c-adb8-43502e9fdba1-config-data-custom\") pod \"heat-api-7b744b9d97-zkcv4\" (UID: \"f5046792-741c-487c-adb8-43502e9fdba1\") " pod="openstack/heat-api-7b744b9d97-zkcv4" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.474891 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gsb2r\" (UniqueName: \"kubernetes.io/projected/305dccfd-067c-4062-b9d1-0200e105e8fc-kube-api-access-gsb2r\") pod \"heat-cfnapi-58f47bdb5-gk4p7\" (UID: \"305dccfd-067c-4062-b9d1-0200e105e8fc\") " pod="openstack/heat-cfnapi-58f47bdb5-gk4p7" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.474912 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5046792-741c-487c-adb8-43502e9fdba1-combined-ca-bundle\") pod \"heat-api-7b744b9d97-zkcv4\" (UID: \"f5046792-741c-487c-adb8-43502e9fdba1\") " pod="openstack/heat-api-7b744b9d97-zkcv4" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.474928 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/305dccfd-067c-4062-b9d1-0200e105e8fc-config-data-custom\") pod \"heat-cfnapi-58f47bdb5-gk4p7\" (UID: \"305dccfd-067c-4062-b9d1-0200e105e8fc\") " pod="openstack/heat-cfnapi-58f47bdb5-gk4p7" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.474980 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/305dccfd-067c-4062-b9d1-0200e105e8fc-config-data\") pod \"heat-cfnapi-58f47bdb5-gk4p7\" (UID: \"305dccfd-067c-4062-b9d1-0200e105e8fc\") " pod="openstack/heat-cfnapi-58f47bdb5-gk4p7" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.474998 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ca8530d7-bcfb-45fa-b0ff-a2702c081bda-config-data-custom\") pod \"heat-engine-5bc6fc99dd-nv45v\" (UID: \"ca8530d7-bcfb-45fa-b0ff-a2702c081bda\") " pod="openstack/heat-engine-5bc6fc99dd-nv45v" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.475053 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x4787\" (UniqueName: \"kubernetes.io/projected/f5046792-741c-487c-adb8-43502e9fdba1-kube-api-access-x4787\") pod \"heat-api-7b744b9d97-zkcv4\" (UID: \"f5046792-741c-487c-adb8-43502e9fdba1\") " pod="openstack/heat-api-7b744b9d97-zkcv4" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.475075 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/305dccfd-067c-4062-b9d1-0200e105e8fc-combined-ca-bundle\") pod \"heat-cfnapi-58f47bdb5-gk4p7\" (UID: \"305dccfd-067c-4062-b9d1-0200e105e8fc\") " pod="openstack/heat-cfnapi-58f47bdb5-gk4p7" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.475154 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bfwhs\" (UniqueName: \"kubernetes.io/projected/ca8530d7-bcfb-45fa-b0ff-a2702c081bda-kube-api-access-bfwhs\") pod \"heat-engine-5bc6fc99dd-nv45v\" (UID: \"ca8530d7-bcfb-45fa-b0ff-a2702c081bda\") " pod="openstack/heat-engine-5bc6fc99dd-nv45v" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.475169 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca8530d7-bcfb-45fa-b0ff-a2702c081bda-combined-ca-bundle\") pod \"heat-engine-5bc6fc99dd-nv45v\" (UID: \"ca8530d7-bcfb-45fa-b0ff-a2702c081bda\") " pod="openstack/heat-engine-5bc6fc99dd-nv45v" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.475185 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca8530d7-bcfb-45fa-b0ff-a2702c081bda-config-data\") pod \"heat-engine-5bc6fc99dd-nv45v\" (UID: \"ca8530d7-bcfb-45fa-b0ff-a2702c081bda\") " pod="openstack/heat-engine-5bc6fc99dd-nv45v" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.576585 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bfwhs\" (UniqueName: \"kubernetes.io/projected/ca8530d7-bcfb-45fa-b0ff-a2702c081bda-kube-api-access-bfwhs\") pod \"heat-engine-5bc6fc99dd-nv45v\" (UID: \"ca8530d7-bcfb-45fa-b0ff-a2702c081bda\") " pod="openstack/heat-engine-5bc6fc99dd-nv45v" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.576632 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca8530d7-bcfb-45fa-b0ff-a2702c081bda-combined-ca-bundle\") pod \"heat-engine-5bc6fc99dd-nv45v\" (UID: \"ca8530d7-bcfb-45fa-b0ff-a2702c081bda\") " pod="openstack/heat-engine-5bc6fc99dd-nv45v" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.576652 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca8530d7-bcfb-45fa-b0ff-a2702c081bda-config-data\") pod \"heat-engine-5bc6fc99dd-nv45v\" (UID: \"ca8530d7-bcfb-45fa-b0ff-a2702c081bda\") " pod="openstack/heat-engine-5bc6fc99dd-nv45v" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.576699 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f5046792-741c-487c-adb8-43502e9fdba1-config-data\") pod \"heat-api-7b744b9d97-zkcv4\" (UID: \"f5046792-741c-487c-adb8-43502e9fdba1\") " pod="openstack/heat-api-7b744b9d97-zkcv4" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.576726 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f5046792-741c-487c-adb8-43502e9fdba1-config-data-custom\") pod \"heat-api-7b744b9d97-zkcv4\" (UID: \"f5046792-741c-487c-adb8-43502e9fdba1\") " pod="openstack/heat-api-7b744b9d97-zkcv4" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.576799 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gsb2r\" (UniqueName: \"kubernetes.io/projected/305dccfd-067c-4062-b9d1-0200e105e8fc-kube-api-access-gsb2r\") pod \"heat-cfnapi-58f47bdb5-gk4p7\" (UID: \"305dccfd-067c-4062-b9d1-0200e105e8fc\") " pod="openstack/heat-cfnapi-58f47bdb5-gk4p7" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.576820 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5046792-741c-487c-adb8-43502e9fdba1-combined-ca-bundle\") pod \"heat-api-7b744b9d97-zkcv4\" (UID: \"f5046792-741c-487c-adb8-43502e9fdba1\") " pod="openstack/heat-api-7b744b9d97-zkcv4" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.576838 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/305dccfd-067c-4062-b9d1-0200e105e8fc-config-data-custom\") pod \"heat-cfnapi-58f47bdb5-gk4p7\" (UID: \"305dccfd-067c-4062-b9d1-0200e105e8fc\") " pod="openstack/heat-cfnapi-58f47bdb5-gk4p7" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.576878 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/305dccfd-067c-4062-b9d1-0200e105e8fc-config-data\") pod \"heat-cfnapi-58f47bdb5-gk4p7\" (UID: \"305dccfd-067c-4062-b9d1-0200e105e8fc\") " pod="openstack/heat-cfnapi-58f47bdb5-gk4p7" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.576893 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ca8530d7-bcfb-45fa-b0ff-a2702c081bda-config-data-custom\") pod \"heat-engine-5bc6fc99dd-nv45v\" (UID: \"ca8530d7-bcfb-45fa-b0ff-a2702c081bda\") " pod="openstack/heat-engine-5bc6fc99dd-nv45v" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.576951 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x4787\" (UniqueName: \"kubernetes.io/projected/f5046792-741c-487c-adb8-43502e9fdba1-kube-api-access-x4787\") pod \"heat-api-7b744b9d97-zkcv4\" (UID: \"f5046792-741c-487c-adb8-43502e9fdba1\") " pod="openstack/heat-api-7b744b9d97-zkcv4" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.576989 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/305dccfd-067c-4062-b9d1-0200e105e8fc-combined-ca-bundle\") pod \"heat-cfnapi-58f47bdb5-gk4p7\" (UID: \"305dccfd-067c-4062-b9d1-0200e105e8fc\") " pod="openstack/heat-cfnapi-58f47bdb5-gk4p7" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.580841 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ca8530d7-bcfb-45fa-b0ff-a2702c081bda-config-data-custom\") pod \"heat-engine-5bc6fc99dd-nv45v\" (UID: \"ca8530d7-bcfb-45fa-b0ff-a2702c081bda\") " pod="openstack/heat-engine-5bc6fc99dd-nv45v" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.581549 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/305dccfd-067c-4062-b9d1-0200e105e8fc-combined-ca-bundle\") pod \"heat-cfnapi-58f47bdb5-gk4p7\" (UID: \"305dccfd-067c-4062-b9d1-0200e105e8fc\") " pod="openstack/heat-cfnapi-58f47bdb5-gk4p7" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.582075 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca8530d7-bcfb-45fa-b0ff-a2702c081bda-config-data\") pod \"heat-engine-5bc6fc99dd-nv45v\" (UID: \"ca8530d7-bcfb-45fa-b0ff-a2702c081bda\") " pod="openstack/heat-engine-5bc6fc99dd-nv45v" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.582142 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f5046792-741c-487c-adb8-43502e9fdba1-config-data\") pod \"heat-api-7b744b9d97-zkcv4\" (UID: \"f5046792-741c-487c-adb8-43502e9fdba1\") " pod="openstack/heat-api-7b744b9d97-zkcv4" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.585603 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/305dccfd-067c-4062-b9d1-0200e105e8fc-config-data-custom\") pod \"heat-cfnapi-58f47bdb5-gk4p7\" (UID: \"305dccfd-067c-4062-b9d1-0200e105e8fc\") " pod="openstack/heat-cfnapi-58f47bdb5-gk4p7" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.588565 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f5046792-741c-487c-adb8-43502e9fdba1-config-data-custom\") pod \"heat-api-7b744b9d97-zkcv4\" (UID: \"f5046792-741c-487c-adb8-43502e9fdba1\") " pod="openstack/heat-api-7b744b9d97-zkcv4" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.588658 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca8530d7-bcfb-45fa-b0ff-a2702c081bda-combined-ca-bundle\") pod \"heat-engine-5bc6fc99dd-nv45v\" (UID: \"ca8530d7-bcfb-45fa-b0ff-a2702c081bda\") " pod="openstack/heat-engine-5bc6fc99dd-nv45v" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.589100 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5046792-741c-487c-adb8-43502e9fdba1-combined-ca-bundle\") pod \"heat-api-7b744b9d97-zkcv4\" (UID: \"f5046792-741c-487c-adb8-43502e9fdba1\") " pod="openstack/heat-api-7b744b9d97-zkcv4" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.589370 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/305dccfd-067c-4062-b9d1-0200e105e8fc-config-data\") pod \"heat-cfnapi-58f47bdb5-gk4p7\" (UID: \"305dccfd-067c-4062-b9d1-0200e105e8fc\") " pod="openstack/heat-cfnapi-58f47bdb5-gk4p7" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.590043 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gsb2r\" (UniqueName: \"kubernetes.io/projected/305dccfd-067c-4062-b9d1-0200e105e8fc-kube-api-access-gsb2r\") pod \"heat-cfnapi-58f47bdb5-gk4p7\" (UID: \"305dccfd-067c-4062-b9d1-0200e105e8fc\") " pod="openstack/heat-cfnapi-58f47bdb5-gk4p7" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.590511 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bfwhs\" (UniqueName: \"kubernetes.io/projected/ca8530d7-bcfb-45fa-b0ff-a2702c081bda-kube-api-access-bfwhs\") pod \"heat-engine-5bc6fc99dd-nv45v\" (UID: \"ca8530d7-bcfb-45fa-b0ff-a2702c081bda\") " pod="openstack/heat-engine-5bc6fc99dd-nv45v" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.590665 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x4787\" (UniqueName: \"kubernetes.io/projected/f5046792-741c-487c-adb8-43502e9fdba1-kube-api-access-x4787\") pod \"heat-api-7b744b9d97-zkcv4\" (UID: \"f5046792-741c-487c-adb8-43502e9fdba1\") " pod="openstack/heat-api-7b744b9d97-zkcv4" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.615894 4824 generic.go:334] "Generic (PLEG): container finished" podID="a89ff227-6009-4377-8f9f-34ac4078646f" containerID="01663fbfa71a8a22bc17e94621a84e2c11185ae78db507b95bd3152b51310836" exitCode=0 Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.615919 4824 generic.go:334] "Generic (PLEG): container finished" podID="a89ff227-6009-4377-8f9f-34ac4078646f" containerID="02e341630b226a52243314bc26a342a8ea843da778071cfc554f827999082ef2" exitCode=0 Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.615936 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a89ff227-6009-4377-8f9f-34ac4078646f","Type":"ContainerDied","Data":"01663fbfa71a8a22bc17e94621a84e2c11185ae78db507b95bd3152b51310836"} Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.615947 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.615971 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a89ff227-6009-4377-8f9f-34ac4078646f","Type":"ContainerDied","Data":"02e341630b226a52243314bc26a342a8ea843da778071cfc554f827999082ef2"} Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.615982 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a89ff227-6009-4377-8f9f-34ac4078646f","Type":"ContainerDied","Data":"24216a6df3f372a52006263ca19c690f67c9d1bbdb8896d6d5fbff2e54961470"} Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.615996 4824 scope.go:117] "RemoveContainer" containerID="184b945bcbd88dab106672c9b74811f2027be88ca7953a946187d4d4ab09c6b1" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.668737 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.681856 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.686933 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.697483 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-5bc6fc99dd-nv45v" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.718933 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.718948 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.722829 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.723007 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.723417 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-58f47bdb5-gk4p7" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.748834 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-7b744b9d97-zkcv4" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.780190 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/964436a3-7963-4a72-bcb6-b7bab7e96574-log-httpd\") pod \"ceilometer-0\" (UID: \"964436a3-7963-4a72-bcb6-b7bab7e96574\") " pod="openstack/ceilometer-0" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.780241 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-928nq\" (UniqueName: \"kubernetes.io/projected/964436a3-7963-4a72-bcb6-b7bab7e96574-kube-api-access-928nq\") pod \"ceilometer-0\" (UID: \"964436a3-7963-4a72-bcb6-b7bab7e96574\") " pod="openstack/ceilometer-0" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.780396 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/964436a3-7963-4a72-bcb6-b7bab7e96574-run-httpd\") pod \"ceilometer-0\" (UID: \"964436a3-7963-4a72-bcb6-b7bab7e96574\") " pod="openstack/ceilometer-0" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.780437 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/964436a3-7963-4a72-bcb6-b7bab7e96574-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"964436a3-7963-4a72-bcb6-b7bab7e96574\") " pod="openstack/ceilometer-0" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.780607 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/964436a3-7963-4a72-bcb6-b7bab7e96574-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"964436a3-7963-4a72-bcb6-b7bab7e96574\") " pod="openstack/ceilometer-0" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.780721 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/964436a3-7963-4a72-bcb6-b7bab7e96574-config-data\") pod \"ceilometer-0\" (UID: \"964436a3-7963-4a72-bcb6-b7bab7e96574\") " pod="openstack/ceilometer-0" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.780766 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/964436a3-7963-4a72-bcb6-b7bab7e96574-scripts\") pod \"ceilometer-0\" (UID: \"964436a3-7963-4a72-bcb6-b7bab7e96574\") " pod="openstack/ceilometer-0" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.882353 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/964436a3-7963-4a72-bcb6-b7bab7e96574-config-data\") pod \"ceilometer-0\" (UID: \"964436a3-7963-4a72-bcb6-b7bab7e96574\") " pod="openstack/ceilometer-0" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.882407 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/964436a3-7963-4a72-bcb6-b7bab7e96574-scripts\") pod \"ceilometer-0\" (UID: \"964436a3-7963-4a72-bcb6-b7bab7e96574\") " pod="openstack/ceilometer-0" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.882449 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/964436a3-7963-4a72-bcb6-b7bab7e96574-log-httpd\") pod \"ceilometer-0\" (UID: \"964436a3-7963-4a72-bcb6-b7bab7e96574\") " pod="openstack/ceilometer-0" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.882472 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-928nq\" (UniqueName: \"kubernetes.io/projected/964436a3-7963-4a72-bcb6-b7bab7e96574-kube-api-access-928nq\") pod \"ceilometer-0\" (UID: \"964436a3-7963-4a72-bcb6-b7bab7e96574\") " pod="openstack/ceilometer-0" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.882537 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/964436a3-7963-4a72-bcb6-b7bab7e96574-run-httpd\") pod \"ceilometer-0\" (UID: \"964436a3-7963-4a72-bcb6-b7bab7e96574\") " pod="openstack/ceilometer-0" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.882554 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/964436a3-7963-4a72-bcb6-b7bab7e96574-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"964436a3-7963-4a72-bcb6-b7bab7e96574\") " pod="openstack/ceilometer-0" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.882653 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/964436a3-7963-4a72-bcb6-b7bab7e96574-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"964436a3-7963-4a72-bcb6-b7bab7e96574\") " pod="openstack/ceilometer-0" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.883535 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/964436a3-7963-4a72-bcb6-b7bab7e96574-log-httpd\") pod \"ceilometer-0\" (UID: \"964436a3-7963-4a72-bcb6-b7bab7e96574\") " pod="openstack/ceilometer-0" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.883742 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/964436a3-7963-4a72-bcb6-b7bab7e96574-run-httpd\") pod \"ceilometer-0\" (UID: \"964436a3-7963-4a72-bcb6-b7bab7e96574\") " pod="openstack/ceilometer-0" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.888231 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/964436a3-7963-4a72-bcb6-b7bab7e96574-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"964436a3-7963-4a72-bcb6-b7bab7e96574\") " pod="openstack/ceilometer-0" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.889359 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/964436a3-7963-4a72-bcb6-b7bab7e96574-scripts\") pod \"ceilometer-0\" (UID: \"964436a3-7963-4a72-bcb6-b7bab7e96574\") " pod="openstack/ceilometer-0" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.890836 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/964436a3-7963-4a72-bcb6-b7bab7e96574-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"964436a3-7963-4a72-bcb6-b7bab7e96574\") " pod="openstack/ceilometer-0" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.894375 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/964436a3-7963-4a72-bcb6-b7bab7e96574-config-data\") pod \"ceilometer-0\" (UID: \"964436a3-7963-4a72-bcb6-b7bab7e96574\") " pod="openstack/ceilometer-0" Jan 21 11:27:39 crc kubenswrapper[4824]: I0121 11:27:39.895802 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-928nq\" (UniqueName: \"kubernetes.io/projected/964436a3-7963-4a72-bcb6-b7bab7e96574-kube-api-access-928nq\") pod \"ceilometer-0\" (UID: \"964436a3-7963-4a72-bcb6-b7bab7e96574\") " pod="openstack/ceilometer-0" Jan 21 11:27:40 crc kubenswrapper[4824]: I0121 11:27:40.048336 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 21 11:27:40 crc kubenswrapper[4824]: I0121 11:27:40.058944 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a89ff227-6009-4377-8f9f-34ac4078646f" path="/var/lib/kubelet/pods/a89ff227-6009-4377-8f9f-34ac4078646f/volumes" Jan 21 11:27:40 crc kubenswrapper[4824]: I0121 11:27:40.177508 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Jan 21 11:27:40 crc kubenswrapper[4824]: I0121 11:27:40.177555 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Jan 21 11:27:40 crc kubenswrapper[4824]: I0121 11:27:40.207467 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Jan 21 11:27:40 crc kubenswrapper[4824]: I0121 11:27:40.228273 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Jan 21 11:27:40 crc kubenswrapper[4824]: I0121 11:27:40.444824 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-667b46bf4d-5h2cq"] Jan 21 11:27:40 crc kubenswrapper[4824]: I0121 11:27:40.445075 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/heat-cfnapi-667b46bf4d-5h2cq" podUID="25ee21eb-c0b8-4d49-8906-4e4144abf6fc" containerName="heat-cfnapi" containerID="cri-o://a355cd0ecd36eae3a40739bece4826924a58c416690883d3d7de3207672df97c" gracePeriod=60 Jan 21 11:27:40 crc kubenswrapper[4824]: I0121 11:27:40.453624 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-5dc8c6bf78-szsb6"] Jan 21 11:27:40 crc kubenswrapper[4824]: I0121 11:27:40.453949 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/heat-api-5dc8c6bf78-szsb6" podUID="2aef2f5b-bb6b-41dd-a3eb-372e7e777122" containerName="heat-api" containerID="cri-o://494c1e9ccd78616e73e2a0a9317748e295b50e43bcd2bc2b2bf79a00562dbea9" gracePeriod=60 Jan 21 11:27:40 crc kubenswrapper[4824]: I0121 11:27:40.457351 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/heat-cfnapi-667b46bf4d-5h2cq" podUID="25ee21eb-c0b8-4d49-8906-4e4144abf6fc" containerName="heat-cfnapi" probeResult="failure" output="Get \"http://10.217.0.178:8000/healthcheck\": EOF" Jan 21 11:27:40 crc kubenswrapper[4824]: I0121 11:27:40.462636 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/heat-api-5dc8c6bf78-szsb6" podUID="2aef2f5b-bb6b-41dd-a3eb-372e7e777122" containerName="heat-api" probeResult="failure" output="Get \"http://10.217.0.180:8004/healthcheck\": EOF" Jan 21 11:27:40 crc kubenswrapper[4824]: I0121 11:27:40.468234 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-cfnapi-68b8cf894c-ws5zd"] Jan 21 11:27:40 crc kubenswrapper[4824]: I0121 11:27:40.469168 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-68b8cf894c-ws5zd" Jan 21 11:27:40 crc kubenswrapper[4824]: I0121 11:27:40.471542 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-heat-cfnapi-internal-svc" Jan 21 11:27:40 crc kubenswrapper[4824]: I0121 11:27:40.471678 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-heat-cfnapi-public-svc" Jan 21 11:27:40 crc kubenswrapper[4824]: I0121 11:27:40.492951 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-68b8cf894c-ws5zd"] Jan 21 11:27:40 crc kubenswrapper[4824]: I0121 11:27:40.505022 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-api-549445f5fc-fwxlz"] Jan 21 11:27:40 crc kubenswrapper[4824]: I0121 11:27:40.506132 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-549445f5fc-fwxlz" Jan 21 11:27:40 crc kubenswrapper[4824]: I0121 11:27:40.508133 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-heat-api-public-svc" Jan 21 11:27:40 crc kubenswrapper[4824]: I0121 11:27:40.508133 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-heat-api-internal-svc" Jan 21 11:27:40 crc kubenswrapper[4824]: I0121 11:27:40.526445 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-549445f5fc-fwxlz"] Jan 21 11:27:40 crc kubenswrapper[4824]: I0121 11:27:40.595910 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1aefb21-aac8-4f8f-a402-16006c1f336a-combined-ca-bundle\") pod \"heat-cfnapi-68b8cf894c-ws5zd\" (UID: \"a1aefb21-aac8-4f8f-a402-16006c1f336a\") " pod="openstack/heat-cfnapi-68b8cf894c-ws5zd" Jan 21 11:27:40 crc kubenswrapper[4824]: I0121 11:27:40.595975 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a1aefb21-aac8-4f8f-a402-16006c1f336a-public-tls-certs\") pod \"heat-cfnapi-68b8cf894c-ws5zd\" (UID: \"a1aefb21-aac8-4f8f-a402-16006c1f336a\") " pod="openstack/heat-cfnapi-68b8cf894c-ws5zd" Jan 21 11:27:40 crc kubenswrapper[4824]: I0121 11:27:40.595996 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8655cad5-2680-4cf6-b50a-0745adeeb469-internal-tls-certs\") pod \"heat-api-549445f5fc-fwxlz\" (UID: \"8655cad5-2680-4cf6-b50a-0745adeeb469\") " pod="openstack/heat-api-549445f5fc-fwxlz" Jan 21 11:27:40 crc kubenswrapper[4824]: I0121 11:27:40.596173 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8655cad5-2680-4cf6-b50a-0745adeeb469-config-data-custom\") pod \"heat-api-549445f5fc-fwxlz\" (UID: \"8655cad5-2680-4cf6-b50a-0745adeeb469\") " pod="openstack/heat-api-549445f5fc-fwxlz" Jan 21 11:27:40 crc kubenswrapper[4824]: I0121 11:27:40.596198 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zpkl5\" (UniqueName: \"kubernetes.io/projected/a1aefb21-aac8-4f8f-a402-16006c1f336a-kube-api-access-zpkl5\") pod \"heat-cfnapi-68b8cf894c-ws5zd\" (UID: \"a1aefb21-aac8-4f8f-a402-16006c1f336a\") " pod="openstack/heat-cfnapi-68b8cf894c-ws5zd" Jan 21 11:27:40 crc kubenswrapper[4824]: I0121 11:27:40.596219 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a1aefb21-aac8-4f8f-a402-16006c1f336a-config-data\") pod \"heat-cfnapi-68b8cf894c-ws5zd\" (UID: \"a1aefb21-aac8-4f8f-a402-16006c1f336a\") " pod="openstack/heat-cfnapi-68b8cf894c-ws5zd" Jan 21 11:27:40 crc kubenswrapper[4824]: I0121 11:27:40.596237 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a1aefb21-aac8-4f8f-a402-16006c1f336a-config-data-custom\") pod \"heat-cfnapi-68b8cf894c-ws5zd\" (UID: \"a1aefb21-aac8-4f8f-a402-16006c1f336a\") " pod="openstack/heat-cfnapi-68b8cf894c-ws5zd" Jan 21 11:27:40 crc kubenswrapper[4824]: I0121 11:27:40.596265 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z5sg2\" (UniqueName: \"kubernetes.io/projected/8655cad5-2680-4cf6-b50a-0745adeeb469-kube-api-access-z5sg2\") pod \"heat-api-549445f5fc-fwxlz\" (UID: \"8655cad5-2680-4cf6-b50a-0745adeeb469\") " pod="openstack/heat-api-549445f5fc-fwxlz" Jan 21 11:27:40 crc kubenswrapper[4824]: I0121 11:27:40.596289 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8655cad5-2680-4cf6-b50a-0745adeeb469-combined-ca-bundle\") pod \"heat-api-549445f5fc-fwxlz\" (UID: \"8655cad5-2680-4cf6-b50a-0745adeeb469\") " pod="openstack/heat-api-549445f5fc-fwxlz" Jan 21 11:27:40 crc kubenswrapper[4824]: I0121 11:27:40.596322 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a1aefb21-aac8-4f8f-a402-16006c1f336a-internal-tls-certs\") pod \"heat-cfnapi-68b8cf894c-ws5zd\" (UID: \"a1aefb21-aac8-4f8f-a402-16006c1f336a\") " pod="openstack/heat-cfnapi-68b8cf894c-ws5zd" Jan 21 11:27:40 crc kubenswrapper[4824]: I0121 11:27:40.596365 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8655cad5-2680-4cf6-b50a-0745adeeb469-public-tls-certs\") pod \"heat-api-549445f5fc-fwxlz\" (UID: \"8655cad5-2680-4cf6-b50a-0745adeeb469\") " pod="openstack/heat-api-549445f5fc-fwxlz" Jan 21 11:27:40 crc kubenswrapper[4824]: I0121 11:27:40.596380 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8655cad5-2680-4cf6-b50a-0745adeeb469-config-data\") pod \"heat-api-549445f5fc-fwxlz\" (UID: \"8655cad5-2680-4cf6-b50a-0745adeeb469\") " pod="openstack/heat-api-549445f5fc-fwxlz" Jan 21 11:27:40 crc kubenswrapper[4824]: I0121 11:27:40.625966 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Jan 21 11:27:40 crc kubenswrapper[4824]: I0121 11:27:40.625995 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Jan 21 11:27:40 crc kubenswrapper[4824]: I0121 11:27:40.697566 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8655cad5-2680-4cf6-b50a-0745adeeb469-config-data-custom\") pod \"heat-api-549445f5fc-fwxlz\" (UID: \"8655cad5-2680-4cf6-b50a-0745adeeb469\") " pod="openstack/heat-api-549445f5fc-fwxlz" Jan 21 11:27:40 crc kubenswrapper[4824]: I0121 11:27:40.697605 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zpkl5\" (UniqueName: \"kubernetes.io/projected/a1aefb21-aac8-4f8f-a402-16006c1f336a-kube-api-access-zpkl5\") pod \"heat-cfnapi-68b8cf894c-ws5zd\" (UID: \"a1aefb21-aac8-4f8f-a402-16006c1f336a\") " pod="openstack/heat-cfnapi-68b8cf894c-ws5zd" Jan 21 11:27:40 crc kubenswrapper[4824]: I0121 11:27:40.697629 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a1aefb21-aac8-4f8f-a402-16006c1f336a-config-data\") pod \"heat-cfnapi-68b8cf894c-ws5zd\" (UID: \"a1aefb21-aac8-4f8f-a402-16006c1f336a\") " pod="openstack/heat-cfnapi-68b8cf894c-ws5zd" Jan 21 11:27:40 crc kubenswrapper[4824]: I0121 11:27:40.697646 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a1aefb21-aac8-4f8f-a402-16006c1f336a-config-data-custom\") pod \"heat-cfnapi-68b8cf894c-ws5zd\" (UID: \"a1aefb21-aac8-4f8f-a402-16006c1f336a\") " pod="openstack/heat-cfnapi-68b8cf894c-ws5zd" Jan 21 11:27:40 crc kubenswrapper[4824]: I0121 11:27:40.697674 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z5sg2\" (UniqueName: \"kubernetes.io/projected/8655cad5-2680-4cf6-b50a-0745adeeb469-kube-api-access-z5sg2\") pod \"heat-api-549445f5fc-fwxlz\" (UID: \"8655cad5-2680-4cf6-b50a-0745adeeb469\") " pod="openstack/heat-api-549445f5fc-fwxlz" Jan 21 11:27:40 crc kubenswrapper[4824]: I0121 11:27:40.697702 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8655cad5-2680-4cf6-b50a-0745adeeb469-combined-ca-bundle\") pod \"heat-api-549445f5fc-fwxlz\" (UID: \"8655cad5-2680-4cf6-b50a-0745adeeb469\") " pod="openstack/heat-api-549445f5fc-fwxlz" Jan 21 11:27:40 crc kubenswrapper[4824]: I0121 11:27:40.697770 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a1aefb21-aac8-4f8f-a402-16006c1f336a-internal-tls-certs\") pod \"heat-cfnapi-68b8cf894c-ws5zd\" (UID: \"a1aefb21-aac8-4f8f-a402-16006c1f336a\") " pod="openstack/heat-cfnapi-68b8cf894c-ws5zd" Jan 21 11:27:40 crc kubenswrapper[4824]: I0121 11:27:40.697815 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8655cad5-2680-4cf6-b50a-0745adeeb469-public-tls-certs\") pod \"heat-api-549445f5fc-fwxlz\" (UID: \"8655cad5-2680-4cf6-b50a-0745adeeb469\") " pod="openstack/heat-api-549445f5fc-fwxlz" Jan 21 11:27:40 crc kubenswrapper[4824]: I0121 11:27:40.697833 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8655cad5-2680-4cf6-b50a-0745adeeb469-config-data\") pod \"heat-api-549445f5fc-fwxlz\" (UID: \"8655cad5-2680-4cf6-b50a-0745adeeb469\") " pod="openstack/heat-api-549445f5fc-fwxlz" Jan 21 11:27:40 crc kubenswrapper[4824]: I0121 11:27:40.697901 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1aefb21-aac8-4f8f-a402-16006c1f336a-combined-ca-bundle\") pod \"heat-cfnapi-68b8cf894c-ws5zd\" (UID: \"a1aefb21-aac8-4f8f-a402-16006c1f336a\") " pod="openstack/heat-cfnapi-68b8cf894c-ws5zd" Jan 21 11:27:40 crc kubenswrapper[4824]: I0121 11:27:40.698454 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a1aefb21-aac8-4f8f-a402-16006c1f336a-public-tls-certs\") pod \"heat-cfnapi-68b8cf894c-ws5zd\" (UID: \"a1aefb21-aac8-4f8f-a402-16006c1f336a\") " pod="openstack/heat-cfnapi-68b8cf894c-ws5zd" Jan 21 11:27:40 crc kubenswrapper[4824]: I0121 11:27:40.699050 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8655cad5-2680-4cf6-b50a-0745adeeb469-internal-tls-certs\") pod \"heat-api-549445f5fc-fwxlz\" (UID: \"8655cad5-2680-4cf6-b50a-0745adeeb469\") " pod="openstack/heat-api-549445f5fc-fwxlz" Jan 21 11:27:40 crc kubenswrapper[4824]: I0121 11:27:40.702552 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8655cad5-2680-4cf6-b50a-0745adeeb469-combined-ca-bundle\") pod \"heat-api-549445f5fc-fwxlz\" (UID: \"8655cad5-2680-4cf6-b50a-0745adeeb469\") " pod="openstack/heat-api-549445f5fc-fwxlz" Jan 21 11:27:40 crc kubenswrapper[4824]: I0121 11:27:40.707995 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a1aefb21-aac8-4f8f-a402-16006c1f336a-config-data-custom\") pod \"heat-cfnapi-68b8cf894c-ws5zd\" (UID: \"a1aefb21-aac8-4f8f-a402-16006c1f336a\") " pod="openstack/heat-cfnapi-68b8cf894c-ws5zd" Jan 21 11:27:40 crc kubenswrapper[4824]: I0121 11:27:40.708093 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1aefb21-aac8-4f8f-a402-16006c1f336a-combined-ca-bundle\") pod \"heat-cfnapi-68b8cf894c-ws5zd\" (UID: \"a1aefb21-aac8-4f8f-a402-16006c1f336a\") " pod="openstack/heat-cfnapi-68b8cf894c-ws5zd" Jan 21 11:27:40 crc kubenswrapper[4824]: I0121 11:27:40.708479 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a1aefb21-aac8-4f8f-a402-16006c1f336a-config-data\") pod \"heat-cfnapi-68b8cf894c-ws5zd\" (UID: \"a1aefb21-aac8-4f8f-a402-16006c1f336a\") " pod="openstack/heat-cfnapi-68b8cf894c-ws5zd" Jan 21 11:27:40 crc kubenswrapper[4824]: I0121 11:27:40.708921 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8655cad5-2680-4cf6-b50a-0745adeeb469-public-tls-certs\") pod \"heat-api-549445f5fc-fwxlz\" (UID: \"8655cad5-2680-4cf6-b50a-0745adeeb469\") " pod="openstack/heat-api-549445f5fc-fwxlz" Jan 21 11:27:40 crc kubenswrapper[4824]: I0121 11:27:40.709118 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8655cad5-2680-4cf6-b50a-0745adeeb469-config-data-custom\") pod \"heat-api-549445f5fc-fwxlz\" (UID: \"8655cad5-2680-4cf6-b50a-0745adeeb469\") " pod="openstack/heat-api-549445f5fc-fwxlz" Jan 21 11:27:40 crc kubenswrapper[4824]: I0121 11:27:40.709531 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8655cad5-2680-4cf6-b50a-0745adeeb469-internal-tls-certs\") pod \"heat-api-549445f5fc-fwxlz\" (UID: \"8655cad5-2680-4cf6-b50a-0745adeeb469\") " pod="openstack/heat-api-549445f5fc-fwxlz" Jan 21 11:27:40 crc kubenswrapper[4824]: I0121 11:27:40.712215 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a1aefb21-aac8-4f8f-a402-16006c1f336a-internal-tls-certs\") pod \"heat-cfnapi-68b8cf894c-ws5zd\" (UID: \"a1aefb21-aac8-4f8f-a402-16006c1f336a\") " pod="openstack/heat-cfnapi-68b8cf894c-ws5zd" Jan 21 11:27:40 crc kubenswrapper[4824]: I0121 11:27:40.717411 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a1aefb21-aac8-4f8f-a402-16006c1f336a-public-tls-certs\") pod \"heat-cfnapi-68b8cf894c-ws5zd\" (UID: \"a1aefb21-aac8-4f8f-a402-16006c1f336a\") " pod="openstack/heat-cfnapi-68b8cf894c-ws5zd" Jan 21 11:27:40 crc kubenswrapper[4824]: I0121 11:27:40.720313 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z5sg2\" (UniqueName: \"kubernetes.io/projected/8655cad5-2680-4cf6-b50a-0745adeeb469-kube-api-access-z5sg2\") pod \"heat-api-549445f5fc-fwxlz\" (UID: \"8655cad5-2680-4cf6-b50a-0745adeeb469\") " pod="openstack/heat-api-549445f5fc-fwxlz" Jan 21 11:27:40 crc kubenswrapper[4824]: I0121 11:27:40.724334 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zpkl5\" (UniqueName: \"kubernetes.io/projected/a1aefb21-aac8-4f8f-a402-16006c1f336a-kube-api-access-zpkl5\") pod \"heat-cfnapi-68b8cf894c-ws5zd\" (UID: \"a1aefb21-aac8-4f8f-a402-16006c1f336a\") " pod="openstack/heat-cfnapi-68b8cf894c-ws5zd" Jan 21 11:27:40 crc kubenswrapper[4824]: I0121 11:27:40.724586 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8655cad5-2680-4cf6-b50a-0745adeeb469-config-data\") pod \"heat-api-549445f5fc-fwxlz\" (UID: \"8655cad5-2680-4cf6-b50a-0745adeeb469\") " pod="openstack/heat-api-549445f5fc-fwxlz" Jan 21 11:27:40 crc kubenswrapper[4824]: I0121 11:27:40.784522 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-68b8cf894c-ws5zd" Jan 21 11:27:40 crc kubenswrapper[4824]: I0121 11:27:40.823258 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-549445f5fc-fwxlz" Jan 21 11:27:42 crc kubenswrapper[4824]: I0121 11:27:42.333609 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Jan 21 11:27:42 crc kubenswrapper[4824]: I0121 11:27:42.393560 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Jan 21 11:27:42 crc kubenswrapper[4824]: I0121 11:27:42.685605 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Jan 21 11:27:42 crc kubenswrapper[4824]: I0121 11:27:42.908933 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7756b9d78c-ftf8g" Jan 21 11:27:42 crc kubenswrapper[4824]: I0121 11:27:42.954882 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-q2xgx"] Jan 21 11:27:42 crc kubenswrapper[4824]: I0121 11:27:42.955086 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5c9776ccc5-q2xgx" podUID="661c81f6-6854-416e-aea7-1344a635f2db" containerName="dnsmasq-dns" containerID="cri-o://4b21f7d92e0f61e8df8d11e4819f3e436ec4a36c57f8769d50008872b4efc836" gracePeriod=10 Jan 21 11:27:43 crc kubenswrapper[4824]: I0121 11:27:43.651924 4824 generic.go:334] "Generic (PLEG): container finished" podID="661c81f6-6854-416e-aea7-1344a635f2db" containerID="4b21f7d92e0f61e8df8d11e4819f3e436ec4a36c57f8769d50008872b4efc836" exitCode=0 Jan 21 11:27:43 crc kubenswrapper[4824]: I0121 11:27:43.652872 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-q2xgx" event={"ID":"661c81f6-6854-416e-aea7-1344a635f2db","Type":"ContainerDied","Data":"4b21f7d92e0f61e8df8d11e4819f3e436ec4a36c57f8769d50008872b4efc836"} Jan 21 11:27:43 crc kubenswrapper[4824]: I0121 11:27:43.731240 4824 scope.go:117] "RemoveContainer" containerID="af5d7d53713185f666f7039b6b0241d8d3bcf7b749f6427dd3855134753e06ca" Jan 21 11:27:43 crc kubenswrapper[4824]: I0121 11:27:43.770586 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-q2xgx" Jan 21 11:27:43 crc kubenswrapper[4824]: I0121 11:27:43.868171 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4ttl\" (UniqueName: \"kubernetes.io/projected/661c81f6-6854-416e-aea7-1344a635f2db-kube-api-access-s4ttl\") pod \"661c81f6-6854-416e-aea7-1344a635f2db\" (UID: \"661c81f6-6854-416e-aea7-1344a635f2db\") " Jan 21 11:27:43 crc kubenswrapper[4824]: I0121 11:27:43.868409 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/661c81f6-6854-416e-aea7-1344a635f2db-ovsdbserver-sb\") pod \"661c81f6-6854-416e-aea7-1344a635f2db\" (UID: \"661c81f6-6854-416e-aea7-1344a635f2db\") " Jan 21 11:27:43 crc kubenswrapper[4824]: I0121 11:27:43.868488 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/661c81f6-6854-416e-aea7-1344a635f2db-ovsdbserver-nb\") pod \"661c81f6-6854-416e-aea7-1344a635f2db\" (UID: \"661c81f6-6854-416e-aea7-1344a635f2db\") " Jan 21 11:27:43 crc kubenswrapper[4824]: I0121 11:27:43.868647 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/661c81f6-6854-416e-aea7-1344a635f2db-config\") pod \"661c81f6-6854-416e-aea7-1344a635f2db\" (UID: \"661c81f6-6854-416e-aea7-1344a635f2db\") " Jan 21 11:27:43 crc kubenswrapper[4824]: I0121 11:27:43.868680 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/661c81f6-6854-416e-aea7-1344a635f2db-dns-svc\") pod \"661c81f6-6854-416e-aea7-1344a635f2db\" (UID: \"661c81f6-6854-416e-aea7-1344a635f2db\") " Jan 21 11:27:43 crc kubenswrapper[4824]: I0121 11:27:43.868705 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/661c81f6-6854-416e-aea7-1344a635f2db-dns-swift-storage-0\") pod \"661c81f6-6854-416e-aea7-1344a635f2db\" (UID: \"661c81f6-6854-416e-aea7-1344a635f2db\") " Jan 21 11:27:43 crc kubenswrapper[4824]: I0121 11:27:43.886419 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/661c81f6-6854-416e-aea7-1344a635f2db-kube-api-access-s4ttl" (OuterVolumeSpecName: "kube-api-access-s4ttl") pod "661c81f6-6854-416e-aea7-1344a635f2db" (UID: "661c81f6-6854-416e-aea7-1344a635f2db"). InnerVolumeSpecName "kube-api-access-s4ttl". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:27:43 crc kubenswrapper[4824]: I0121 11:27:43.895223 4824 scope.go:117] "RemoveContainer" containerID="01663fbfa71a8a22bc17e94621a84e2c11185ae78db507b95bd3152b51310836" Jan 21 11:27:43 crc kubenswrapper[4824]: I0121 11:27:43.927686 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/661c81f6-6854-416e-aea7-1344a635f2db-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "661c81f6-6854-416e-aea7-1344a635f2db" (UID: "661c81f6-6854-416e-aea7-1344a635f2db"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:27:43 crc kubenswrapper[4824]: I0121 11:27:43.935276 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/661c81f6-6854-416e-aea7-1344a635f2db-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "661c81f6-6854-416e-aea7-1344a635f2db" (UID: "661c81f6-6854-416e-aea7-1344a635f2db"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:27:43 crc kubenswrapper[4824]: I0121 11:27:43.942252 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/661c81f6-6854-416e-aea7-1344a635f2db-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "661c81f6-6854-416e-aea7-1344a635f2db" (UID: "661c81f6-6854-416e-aea7-1344a635f2db"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:27:43 crc kubenswrapper[4824]: I0121 11:27:43.946397 4824 scope.go:117] "RemoveContainer" containerID="02e341630b226a52243314bc26a342a8ea843da778071cfc554f827999082ef2" Jan 21 11:27:43 crc kubenswrapper[4824]: I0121 11:27:43.950297 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/661c81f6-6854-416e-aea7-1344a635f2db-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "661c81f6-6854-416e-aea7-1344a635f2db" (UID: "661c81f6-6854-416e-aea7-1344a635f2db"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:27:43 crc kubenswrapper[4824]: I0121 11:27:43.955572 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/661c81f6-6854-416e-aea7-1344a635f2db-config" (OuterVolumeSpecName: "config") pod "661c81f6-6854-416e-aea7-1344a635f2db" (UID: "661c81f6-6854-416e-aea7-1344a635f2db"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:27:43 crc kubenswrapper[4824]: I0121 11:27:43.970792 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/661c81f6-6854-416e-aea7-1344a635f2db-config\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:43 crc kubenswrapper[4824]: I0121 11:27:43.970815 4824 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/661c81f6-6854-416e-aea7-1344a635f2db-dns-svc\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:43 crc kubenswrapper[4824]: I0121 11:27:43.970824 4824 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/661c81f6-6854-416e-aea7-1344a635f2db-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:43 crc kubenswrapper[4824]: I0121 11:27:43.970834 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4ttl\" (UniqueName: \"kubernetes.io/projected/661c81f6-6854-416e-aea7-1344a635f2db-kube-api-access-s4ttl\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:43 crc kubenswrapper[4824]: I0121 11:27:43.970842 4824 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/661c81f6-6854-416e-aea7-1344a635f2db-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:43 crc kubenswrapper[4824]: I0121 11:27:43.970849 4824 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/661c81f6-6854-416e-aea7-1344a635f2db-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:43 crc kubenswrapper[4824]: I0121 11:27:43.983080 4824 scope.go:117] "RemoveContainer" containerID="184b945bcbd88dab106672c9b74811f2027be88ca7953a946187d4d4ab09c6b1" Jan 21 11:27:43 crc kubenswrapper[4824]: E0121 11:27:43.983472 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"184b945bcbd88dab106672c9b74811f2027be88ca7953a946187d4d4ab09c6b1\": container with ID starting with 184b945bcbd88dab106672c9b74811f2027be88ca7953a946187d4d4ab09c6b1 not found: ID does not exist" containerID="184b945bcbd88dab106672c9b74811f2027be88ca7953a946187d4d4ab09c6b1" Jan 21 11:27:43 crc kubenswrapper[4824]: I0121 11:27:43.983507 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"184b945bcbd88dab106672c9b74811f2027be88ca7953a946187d4d4ab09c6b1"} err="failed to get container status \"184b945bcbd88dab106672c9b74811f2027be88ca7953a946187d4d4ab09c6b1\": rpc error: code = NotFound desc = could not find container \"184b945bcbd88dab106672c9b74811f2027be88ca7953a946187d4d4ab09c6b1\": container with ID starting with 184b945bcbd88dab106672c9b74811f2027be88ca7953a946187d4d4ab09c6b1 not found: ID does not exist" Jan 21 11:27:43 crc kubenswrapper[4824]: I0121 11:27:43.983532 4824 scope.go:117] "RemoveContainer" containerID="af5d7d53713185f666f7039b6b0241d8d3bcf7b749f6427dd3855134753e06ca" Jan 21 11:27:43 crc kubenswrapper[4824]: E0121 11:27:43.985631 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"af5d7d53713185f666f7039b6b0241d8d3bcf7b749f6427dd3855134753e06ca\": container with ID starting with af5d7d53713185f666f7039b6b0241d8d3bcf7b749f6427dd3855134753e06ca not found: ID does not exist" containerID="af5d7d53713185f666f7039b6b0241d8d3bcf7b749f6427dd3855134753e06ca" Jan 21 11:27:43 crc kubenswrapper[4824]: I0121 11:27:43.985660 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"af5d7d53713185f666f7039b6b0241d8d3bcf7b749f6427dd3855134753e06ca"} err="failed to get container status \"af5d7d53713185f666f7039b6b0241d8d3bcf7b749f6427dd3855134753e06ca\": rpc error: code = NotFound desc = could not find container \"af5d7d53713185f666f7039b6b0241d8d3bcf7b749f6427dd3855134753e06ca\": container with ID starting with af5d7d53713185f666f7039b6b0241d8d3bcf7b749f6427dd3855134753e06ca not found: ID does not exist" Jan 21 11:27:43 crc kubenswrapper[4824]: I0121 11:27:43.985681 4824 scope.go:117] "RemoveContainer" containerID="01663fbfa71a8a22bc17e94621a84e2c11185ae78db507b95bd3152b51310836" Jan 21 11:27:43 crc kubenswrapper[4824]: E0121 11:27:43.986072 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"01663fbfa71a8a22bc17e94621a84e2c11185ae78db507b95bd3152b51310836\": container with ID starting with 01663fbfa71a8a22bc17e94621a84e2c11185ae78db507b95bd3152b51310836 not found: ID does not exist" containerID="01663fbfa71a8a22bc17e94621a84e2c11185ae78db507b95bd3152b51310836" Jan 21 11:27:43 crc kubenswrapper[4824]: I0121 11:27:43.986120 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"01663fbfa71a8a22bc17e94621a84e2c11185ae78db507b95bd3152b51310836"} err="failed to get container status \"01663fbfa71a8a22bc17e94621a84e2c11185ae78db507b95bd3152b51310836\": rpc error: code = NotFound desc = could not find container \"01663fbfa71a8a22bc17e94621a84e2c11185ae78db507b95bd3152b51310836\": container with ID starting with 01663fbfa71a8a22bc17e94621a84e2c11185ae78db507b95bd3152b51310836 not found: ID does not exist" Jan 21 11:27:43 crc kubenswrapper[4824]: I0121 11:27:43.986143 4824 scope.go:117] "RemoveContainer" containerID="02e341630b226a52243314bc26a342a8ea843da778071cfc554f827999082ef2" Jan 21 11:27:43 crc kubenswrapper[4824]: E0121 11:27:43.986913 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"02e341630b226a52243314bc26a342a8ea843da778071cfc554f827999082ef2\": container with ID starting with 02e341630b226a52243314bc26a342a8ea843da778071cfc554f827999082ef2 not found: ID does not exist" containerID="02e341630b226a52243314bc26a342a8ea843da778071cfc554f827999082ef2" Jan 21 11:27:43 crc kubenswrapper[4824]: I0121 11:27:43.986939 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"02e341630b226a52243314bc26a342a8ea843da778071cfc554f827999082ef2"} err="failed to get container status \"02e341630b226a52243314bc26a342a8ea843da778071cfc554f827999082ef2\": rpc error: code = NotFound desc = could not find container \"02e341630b226a52243314bc26a342a8ea843da778071cfc554f827999082ef2\": container with ID starting with 02e341630b226a52243314bc26a342a8ea843da778071cfc554f827999082ef2 not found: ID does not exist" Jan 21 11:27:43 crc kubenswrapper[4824]: I0121 11:27:43.986975 4824 scope.go:117] "RemoveContainer" containerID="184b945bcbd88dab106672c9b74811f2027be88ca7953a946187d4d4ab09c6b1" Jan 21 11:27:43 crc kubenswrapper[4824]: I0121 11:27:43.988102 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"184b945bcbd88dab106672c9b74811f2027be88ca7953a946187d4d4ab09c6b1"} err="failed to get container status \"184b945bcbd88dab106672c9b74811f2027be88ca7953a946187d4d4ab09c6b1\": rpc error: code = NotFound desc = could not find container \"184b945bcbd88dab106672c9b74811f2027be88ca7953a946187d4d4ab09c6b1\": container with ID starting with 184b945bcbd88dab106672c9b74811f2027be88ca7953a946187d4d4ab09c6b1 not found: ID does not exist" Jan 21 11:27:43 crc kubenswrapper[4824]: I0121 11:27:43.988127 4824 scope.go:117] "RemoveContainer" containerID="af5d7d53713185f666f7039b6b0241d8d3bcf7b749f6427dd3855134753e06ca" Jan 21 11:27:43 crc kubenswrapper[4824]: I0121 11:27:43.988614 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"af5d7d53713185f666f7039b6b0241d8d3bcf7b749f6427dd3855134753e06ca"} err="failed to get container status \"af5d7d53713185f666f7039b6b0241d8d3bcf7b749f6427dd3855134753e06ca\": rpc error: code = NotFound desc = could not find container \"af5d7d53713185f666f7039b6b0241d8d3bcf7b749f6427dd3855134753e06ca\": container with ID starting with af5d7d53713185f666f7039b6b0241d8d3bcf7b749f6427dd3855134753e06ca not found: ID does not exist" Jan 21 11:27:43 crc kubenswrapper[4824]: I0121 11:27:43.988643 4824 scope.go:117] "RemoveContainer" containerID="01663fbfa71a8a22bc17e94621a84e2c11185ae78db507b95bd3152b51310836" Jan 21 11:27:43 crc kubenswrapper[4824]: I0121 11:27:43.989065 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"01663fbfa71a8a22bc17e94621a84e2c11185ae78db507b95bd3152b51310836"} err="failed to get container status \"01663fbfa71a8a22bc17e94621a84e2c11185ae78db507b95bd3152b51310836\": rpc error: code = NotFound desc = could not find container \"01663fbfa71a8a22bc17e94621a84e2c11185ae78db507b95bd3152b51310836\": container with ID starting with 01663fbfa71a8a22bc17e94621a84e2c11185ae78db507b95bd3152b51310836 not found: ID does not exist" Jan 21 11:27:43 crc kubenswrapper[4824]: I0121 11:27:43.989087 4824 scope.go:117] "RemoveContainer" containerID="02e341630b226a52243314bc26a342a8ea843da778071cfc554f827999082ef2" Jan 21 11:27:43 crc kubenswrapper[4824]: I0121 11:27:43.989426 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"02e341630b226a52243314bc26a342a8ea843da778071cfc554f827999082ef2"} err="failed to get container status \"02e341630b226a52243314bc26a342a8ea843da778071cfc554f827999082ef2\": rpc error: code = NotFound desc = could not find container \"02e341630b226a52243314bc26a342a8ea843da778071cfc554f827999082ef2\": container with ID starting with 02e341630b226a52243314bc26a342a8ea843da778071cfc554f827999082ef2 not found: ID does not exist" Jan 21 11:27:44 crc kubenswrapper[4824]: W0121 11:27:44.300853 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8655cad5_2680_4cf6_b50a_0745adeeb469.slice/crio-a44e730098ab7c3cdd3ba90cf4c5cbf78ade4baaed854eb0ab5682d139719e48 WatchSource:0}: Error finding container a44e730098ab7c3cdd3ba90cf4c5cbf78ade4baaed854eb0ab5682d139719e48: Status 404 returned error can't find the container with id a44e730098ab7c3cdd3ba90cf4c5cbf78ade4baaed854eb0ab5682d139719e48 Jan 21 11:27:44 crc kubenswrapper[4824]: I0121 11:27:44.302095 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-549445f5fc-fwxlz"] Jan 21 11:27:44 crc kubenswrapper[4824]: I0121 11:27:44.312626 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-7b744b9d97-zkcv4"] Jan 21 11:27:44 crc kubenswrapper[4824]: I0121 11:27:44.440981 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-58f47bdb5-gk4p7"] Jan 21 11:27:44 crc kubenswrapper[4824]: I0121 11:27:44.448586 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-5bc6fc99dd-nv45v"] Jan 21 11:27:44 crc kubenswrapper[4824]: I0121 11:27:44.456044 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-68b8cf894c-ws5zd"] Jan 21 11:27:44 crc kubenswrapper[4824]: I0121 11:27:44.468886 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Jan 21 11:27:44 crc kubenswrapper[4824]: I0121 11:27:44.662322 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-5bc6fc99dd-nv45v" event={"ID":"ca8530d7-bcfb-45fa-b0ff-a2702c081bda","Type":"ContainerStarted","Data":"5e735e3d39be9fd0c79178ef7f036c4cf4579854605131e05065dea5855f0562"} Jan 21 11:27:44 crc kubenswrapper[4824]: I0121 11:27:44.664164 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-7b744b9d97-zkcv4" event={"ID":"f5046792-741c-487c-adb8-43502e9fdba1","Type":"ContainerStarted","Data":"8b94a9402348e00a388234d7d2563fa253b97825d6e2edc2c72d291efaccf943"} Jan 21 11:27:44 crc kubenswrapper[4824]: I0121 11:27:44.664683 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-api-7b744b9d97-zkcv4" Jan 21 11:27:44 crc kubenswrapper[4824]: I0121 11:27:44.666944 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-68b8cf894c-ws5zd" event={"ID":"a1aefb21-aac8-4f8f-a402-16006c1f336a","Type":"ContainerStarted","Data":"d74e4c157d87bd34241fbac801d56eecd0141e60287193222fe70e309b760b68"} Jan 21 11:27:44 crc kubenswrapper[4824]: I0121 11:27:44.670800 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-q2xgx" event={"ID":"661c81f6-6854-416e-aea7-1344a635f2db","Type":"ContainerDied","Data":"e5f93c719de6d721a85fa8dea128d5a96c33c0da5f5e176c82969aadb4af41ac"} Jan 21 11:27:44 crc kubenswrapper[4824]: I0121 11:27:44.670821 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-q2xgx" Jan 21 11:27:44 crc kubenswrapper[4824]: I0121 11:27:44.670832 4824 scope.go:117] "RemoveContainer" containerID="4b21f7d92e0f61e8df8d11e4819f3e436ec4a36c57f8769d50008872b4efc836" Jan 21 11:27:44 crc kubenswrapper[4824]: I0121 11:27:44.672069 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-549445f5fc-fwxlz" event={"ID":"8655cad5-2680-4cf6-b50a-0745adeeb469","Type":"ContainerStarted","Data":"a44e730098ab7c3cdd3ba90cf4c5cbf78ade4baaed854eb0ab5682d139719e48"} Jan 21 11:27:44 crc kubenswrapper[4824]: I0121 11:27:44.679996 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-api-7b744b9d97-zkcv4" podStartSLOduration=5.679984606 podStartE2EDuration="5.679984606s" podCreationTimestamp="2026-01-21 11:27:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:27:44.674308249 +0000 UTC m=+1006.967337541" watchObservedRunningTime="2026-01-21 11:27:44.679984606 +0000 UTC m=+1006.973013898" Jan 21 11:27:44 crc kubenswrapper[4824]: I0121 11:27:44.680312 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-k6zm8" event={"ID":"9c73eda8-d376-4e18-9b33-f5ac4d10d091","Type":"ContainerStarted","Data":"c2bb2e93e558c6a9e9ce5d9ae43e9812275d36c293fd644336fd7e9fcc3ba6c0"} Jan 21 11:27:44 crc kubenswrapper[4824]: I0121 11:27:44.686287 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"964436a3-7963-4a72-bcb6-b7bab7e96574","Type":"ContainerStarted","Data":"cd894bcfd6b429a651ad2e896089b4e35850d9690bc597b40bbe1fa904affe57"} Jan 21 11:27:44 crc kubenswrapper[4824]: I0121 11:27:44.689668 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-58f47bdb5-gk4p7" event={"ID":"305dccfd-067c-4062-b9d1-0200e105e8fc","Type":"ContainerStarted","Data":"88223e1727fde2b87a86e2d66dfde6e4b257cc1ef3f6ba548c34ddd1445f80ea"} Jan 21 11:27:44 crc kubenswrapper[4824]: I0121 11:27:44.707644 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-q2xgx"] Jan 21 11:27:44 crc kubenswrapper[4824]: I0121 11:27:44.715488 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-q2xgx"] Jan 21 11:27:44 crc kubenswrapper[4824]: I0121 11:27:44.719471 4824 scope.go:117] "RemoveContainer" containerID="18cc2896c05b5182d3e266af8c0e8d6f107da5daa9b81a2e4d5ef8e0cf371e65" Jan 21 11:27:44 crc kubenswrapper[4824]: I0121 11:27:44.720026 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-k6zm8" podStartSLOduration=2.366957178 podStartE2EDuration="9.720011316s" podCreationTimestamp="2026-01-21 11:27:35 +0000 UTC" firstStartedPulling="2026-01-21 11:27:36.542165652 +0000 UTC m=+998.835194944" lastFinishedPulling="2026-01-21 11:27:43.895219791 +0000 UTC m=+1006.188249082" observedRunningTime="2026-01-21 11:27:44.697166845 +0000 UTC m=+1006.990196137" watchObservedRunningTime="2026-01-21 11:27:44.720011316 +0000 UTC m=+1007.013040608" Jan 21 11:27:44 crc kubenswrapper[4824]: I0121 11:27:44.843180 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/heat-cfnapi-667b46bf4d-5h2cq" podUID="25ee21eb-c0b8-4d49-8906-4e4144abf6fc" containerName="heat-cfnapi" probeResult="failure" output="Get \"http://10.217.0.178:8000/healthcheck\": read tcp 10.217.0.2:58164->10.217.0.178:8000: read: connection reset by peer" Jan 21 11:27:44 crc kubenswrapper[4824]: I0121 11:27:44.847374 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/heat-api-5dc8c6bf78-szsb6" podUID="2aef2f5b-bb6b-41dd-a3eb-372e7e777122" containerName="heat-api" probeResult="failure" output="Get \"http://10.217.0.180:8004/healthcheck\": read tcp 10.217.0.2:48836->10.217.0.180:8004: read: connection reset by peer" Jan 21 11:27:45 crc kubenswrapper[4824]: I0121 11:27:45.140879 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-5dc8c6bf78-szsb6" Jan 21 11:27:45 crc kubenswrapper[4824]: I0121 11:27:45.197467 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dkztw\" (UniqueName: \"kubernetes.io/projected/2aef2f5b-bb6b-41dd-a3eb-372e7e777122-kube-api-access-dkztw\") pod \"2aef2f5b-bb6b-41dd-a3eb-372e7e777122\" (UID: \"2aef2f5b-bb6b-41dd-a3eb-372e7e777122\") " Jan 21 11:27:45 crc kubenswrapper[4824]: I0121 11:27:45.197814 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2aef2f5b-bb6b-41dd-a3eb-372e7e777122-combined-ca-bundle\") pod \"2aef2f5b-bb6b-41dd-a3eb-372e7e777122\" (UID: \"2aef2f5b-bb6b-41dd-a3eb-372e7e777122\") " Jan 21 11:27:45 crc kubenswrapper[4824]: I0121 11:27:45.198078 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2aef2f5b-bb6b-41dd-a3eb-372e7e777122-config-data\") pod \"2aef2f5b-bb6b-41dd-a3eb-372e7e777122\" (UID: \"2aef2f5b-bb6b-41dd-a3eb-372e7e777122\") " Jan 21 11:27:45 crc kubenswrapper[4824]: I0121 11:27:45.198200 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2aef2f5b-bb6b-41dd-a3eb-372e7e777122-config-data-custom\") pod \"2aef2f5b-bb6b-41dd-a3eb-372e7e777122\" (UID: \"2aef2f5b-bb6b-41dd-a3eb-372e7e777122\") " Jan 21 11:27:45 crc kubenswrapper[4824]: I0121 11:27:45.204044 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2aef2f5b-bb6b-41dd-a3eb-372e7e777122-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "2aef2f5b-bb6b-41dd-a3eb-372e7e777122" (UID: "2aef2f5b-bb6b-41dd-a3eb-372e7e777122"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:27:45 crc kubenswrapper[4824]: I0121 11:27:45.222078 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2aef2f5b-bb6b-41dd-a3eb-372e7e777122-kube-api-access-dkztw" (OuterVolumeSpecName: "kube-api-access-dkztw") pod "2aef2f5b-bb6b-41dd-a3eb-372e7e777122" (UID: "2aef2f5b-bb6b-41dd-a3eb-372e7e777122"). InnerVolumeSpecName "kube-api-access-dkztw". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:27:45 crc kubenswrapper[4824]: I0121 11:27:45.271945 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2aef2f5b-bb6b-41dd-a3eb-372e7e777122-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2aef2f5b-bb6b-41dd-a3eb-372e7e777122" (UID: "2aef2f5b-bb6b-41dd-a3eb-372e7e777122"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:27:45 crc kubenswrapper[4824]: I0121 11:27:45.279225 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2aef2f5b-bb6b-41dd-a3eb-372e7e777122-config-data" (OuterVolumeSpecName: "config-data") pod "2aef2f5b-bb6b-41dd-a3eb-372e7e777122" (UID: "2aef2f5b-bb6b-41dd-a3eb-372e7e777122"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:27:45 crc kubenswrapper[4824]: I0121 11:27:45.301858 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2aef2f5b-bb6b-41dd-a3eb-372e7e777122-config-data\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:45 crc kubenswrapper[4824]: I0121 11:27:45.301967 4824 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2aef2f5b-bb6b-41dd-a3eb-372e7e777122-config-data-custom\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:45 crc kubenswrapper[4824]: I0121 11:27:45.302038 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dkztw\" (UniqueName: \"kubernetes.io/projected/2aef2f5b-bb6b-41dd-a3eb-372e7e777122-kube-api-access-dkztw\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:45 crc kubenswrapper[4824]: I0121 11:27:45.302098 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2aef2f5b-bb6b-41dd-a3eb-372e7e777122-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:45 crc kubenswrapper[4824]: I0121 11:27:45.310470 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-667b46bf4d-5h2cq" Jan 21 11:27:45 crc kubenswrapper[4824]: I0121 11:27:45.403214 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/25ee21eb-c0b8-4d49-8906-4e4144abf6fc-config-data-custom\") pod \"25ee21eb-c0b8-4d49-8906-4e4144abf6fc\" (UID: \"25ee21eb-c0b8-4d49-8906-4e4144abf6fc\") " Jan 21 11:27:45 crc kubenswrapper[4824]: I0121 11:27:45.403334 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/25ee21eb-c0b8-4d49-8906-4e4144abf6fc-config-data\") pod \"25ee21eb-c0b8-4d49-8906-4e4144abf6fc\" (UID: \"25ee21eb-c0b8-4d49-8906-4e4144abf6fc\") " Jan 21 11:27:45 crc kubenswrapper[4824]: I0121 11:27:45.403444 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kxlch\" (UniqueName: \"kubernetes.io/projected/25ee21eb-c0b8-4d49-8906-4e4144abf6fc-kube-api-access-kxlch\") pod \"25ee21eb-c0b8-4d49-8906-4e4144abf6fc\" (UID: \"25ee21eb-c0b8-4d49-8906-4e4144abf6fc\") " Jan 21 11:27:45 crc kubenswrapper[4824]: I0121 11:27:45.403472 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/25ee21eb-c0b8-4d49-8906-4e4144abf6fc-combined-ca-bundle\") pod \"25ee21eb-c0b8-4d49-8906-4e4144abf6fc\" (UID: \"25ee21eb-c0b8-4d49-8906-4e4144abf6fc\") " Jan 21 11:27:45 crc kubenswrapper[4824]: I0121 11:27:45.406305 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25ee21eb-c0b8-4d49-8906-4e4144abf6fc-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "25ee21eb-c0b8-4d49-8906-4e4144abf6fc" (UID: "25ee21eb-c0b8-4d49-8906-4e4144abf6fc"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:27:45 crc kubenswrapper[4824]: I0121 11:27:45.407167 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25ee21eb-c0b8-4d49-8906-4e4144abf6fc-kube-api-access-kxlch" (OuterVolumeSpecName: "kube-api-access-kxlch") pod "25ee21eb-c0b8-4d49-8906-4e4144abf6fc" (UID: "25ee21eb-c0b8-4d49-8906-4e4144abf6fc"). InnerVolumeSpecName "kube-api-access-kxlch". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:27:45 crc kubenswrapper[4824]: I0121 11:27:45.427233 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25ee21eb-c0b8-4d49-8906-4e4144abf6fc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "25ee21eb-c0b8-4d49-8906-4e4144abf6fc" (UID: "25ee21eb-c0b8-4d49-8906-4e4144abf6fc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:27:45 crc kubenswrapper[4824]: I0121 11:27:45.442584 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25ee21eb-c0b8-4d49-8906-4e4144abf6fc-config-data" (OuterVolumeSpecName: "config-data") pod "25ee21eb-c0b8-4d49-8906-4e4144abf6fc" (UID: "25ee21eb-c0b8-4d49-8906-4e4144abf6fc"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:27:45 crc kubenswrapper[4824]: I0121 11:27:45.505613 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kxlch\" (UniqueName: \"kubernetes.io/projected/25ee21eb-c0b8-4d49-8906-4e4144abf6fc-kube-api-access-kxlch\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:45 crc kubenswrapper[4824]: I0121 11:27:45.505738 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/25ee21eb-c0b8-4d49-8906-4e4144abf6fc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:45 crc kubenswrapper[4824]: I0121 11:27:45.505808 4824 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/25ee21eb-c0b8-4d49-8906-4e4144abf6fc-config-data-custom\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:45 crc kubenswrapper[4824]: I0121 11:27:45.505859 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/25ee21eb-c0b8-4d49-8906-4e4144abf6fc-config-data\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:45 crc kubenswrapper[4824]: I0121 11:27:45.711610 4824 generic.go:334] "Generic (PLEG): container finished" podID="2aef2f5b-bb6b-41dd-a3eb-372e7e777122" containerID="494c1e9ccd78616e73e2a0a9317748e295b50e43bcd2bc2b2bf79a00562dbea9" exitCode=0 Jan 21 11:27:45 crc kubenswrapper[4824]: I0121 11:27:45.712598 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-5dc8c6bf78-szsb6" event={"ID":"2aef2f5b-bb6b-41dd-a3eb-372e7e777122","Type":"ContainerDied","Data":"494c1e9ccd78616e73e2a0a9317748e295b50e43bcd2bc2b2bf79a00562dbea9"} Jan 21 11:27:45 crc kubenswrapper[4824]: I0121 11:27:45.712705 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-5dc8c6bf78-szsb6" event={"ID":"2aef2f5b-bb6b-41dd-a3eb-372e7e777122","Type":"ContainerDied","Data":"986ea68f679368ee6e505b19ac9e641262bd881fd3d310847834bbf4f66d444f"} Jan 21 11:27:45 crc kubenswrapper[4824]: I0121 11:27:45.712800 4824 scope.go:117] "RemoveContainer" containerID="494c1e9ccd78616e73e2a0a9317748e295b50e43bcd2bc2b2bf79a00562dbea9" Jan 21 11:27:45 crc kubenswrapper[4824]: I0121 11:27:45.712965 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-5dc8c6bf78-szsb6" Jan 21 11:27:45 crc kubenswrapper[4824]: I0121 11:27:45.722940 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-549445f5fc-fwxlz" event={"ID":"8655cad5-2680-4cf6-b50a-0745adeeb469","Type":"ContainerStarted","Data":"7ca3aa4524f443c234ea4b2cf034ce17a404f31b8799b420c3731316971fe010"} Jan 21 11:27:45 crc kubenswrapper[4824]: I0121 11:27:45.723863 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-api-549445f5fc-fwxlz" Jan 21 11:27:45 crc kubenswrapper[4824]: I0121 11:27:45.725169 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-5bc6fc99dd-nv45v" event={"ID":"ca8530d7-bcfb-45fa-b0ff-a2702c081bda","Type":"ContainerStarted","Data":"cb9f39deb21448da8b4f6bd8fbe7c0a85276c9c2b132f84f1777f124f04d67b2"} Jan 21 11:27:45 crc kubenswrapper[4824]: I0121 11:27:45.725785 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-engine-5bc6fc99dd-nv45v" Jan 21 11:27:45 crc kubenswrapper[4824]: I0121 11:27:45.731559 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"964436a3-7963-4a72-bcb6-b7bab7e96574","Type":"ContainerStarted","Data":"1746442b941230e3b55e2a919cace736ddde3df751e7bdf5c201a6243b55e8fc"} Jan 21 11:27:45 crc kubenswrapper[4824]: I0121 11:27:45.745188 4824 generic.go:334] "Generic (PLEG): container finished" podID="25ee21eb-c0b8-4d49-8906-4e4144abf6fc" containerID="a355cd0ecd36eae3a40739bece4826924a58c416690883d3d7de3207672df97c" exitCode=0 Jan 21 11:27:45 crc kubenswrapper[4824]: I0121 11:27:45.745303 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-667b46bf4d-5h2cq" event={"ID":"25ee21eb-c0b8-4d49-8906-4e4144abf6fc","Type":"ContainerDied","Data":"a355cd0ecd36eae3a40739bece4826924a58c416690883d3d7de3207672df97c"} Jan 21 11:27:45 crc kubenswrapper[4824]: I0121 11:27:45.745390 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-667b46bf4d-5h2cq" event={"ID":"25ee21eb-c0b8-4d49-8906-4e4144abf6fc","Type":"ContainerDied","Data":"3cb46ceb8d95d8862082f8003345a2758b9fe77a19473ef416fddbd583e50625"} Jan 21 11:27:45 crc kubenswrapper[4824]: I0121 11:27:45.745490 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-667b46bf4d-5h2cq" Jan 21 11:27:45 crc kubenswrapper[4824]: I0121 11:27:45.772181 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-68b8cf894c-ws5zd" event={"ID":"a1aefb21-aac8-4f8f-a402-16006c1f336a","Type":"ContainerStarted","Data":"54cc72cdba4177c799b1f21f80543fa4853b99e45fa1181b099b6665e7615077"} Jan 21 11:27:45 crc kubenswrapper[4824]: I0121 11:27:45.772862 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-cfnapi-68b8cf894c-ws5zd" Jan 21 11:27:45 crc kubenswrapper[4824]: I0121 11:27:45.774351 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-api-549445f5fc-fwxlz" podStartSLOduration=5.774335331 podStartE2EDuration="5.774335331s" podCreationTimestamp="2026-01-21 11:27:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:27:45.760189208 +0000 UTC m=+1008.053218500" watchObservedRunningTime="2026-01-21 11:27:45.774335331 +0000 UTC m=+1008.067364623" Jan 21 11:27:45 crc kubenswrapper[4824]: I0121 11:27:45.777056 4824 scope.go:117] "RemoveContainer" containerID="494c1e9ccd78616e73e2a0a9317748e295b50e43bcd2bc2b2bf79a00562dbea9" Jan 21 11:27:45 crc kubenswrapper[4824]: E0121 11:27:45.779132 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"494c1e9ccd78616e73e2a0a9317748e295b50e43bcd2bc2b2bf79a00562dbea9\": container with ID starting with 494c1e9ccd78616e73e2a0a9317748e295b50e43bcd2bc2b2bf79a00562dbea9 not found: ID does not exist" containerID="494c1e9ccd78616e73e2a0a9317748e295b50e43bcd2bc2b2bf79a00562dbea9" Jan 21 11:27:45 crc kubenswrapper[4824]: I0121 11:27:45.779162 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"494c1e9ccd78616e73e2a0a9317748e295b50e43bcd2bc2b2bf79a00562dbea9"} err="failed to get container status \"494c1e9ccd78616e73e2a0a9317748e295b50e43bcd2bc2b2bf79a00562dbea9\": rpc error: code = NotFound desc = could not find container \"494c1e9ccd78616e73e2a0a9317748e295b50e43bcd2bc2b2bf79a00562dbea9\": container with ID starting with 494c1e9ccd78616e73e2a0a9317748e295b50e43bcd2bc2b2bf79a00562dbea9 not found: ID does not exist" Jan 21 11:27:45 crc kubenswrapper[4824]: I0121 11:27:45.779182 4824 scope.go:117] "RemoveContainer" containerID="a355cd0ecd36eae3a40739bece4826924a58c416690883d3d7de3207672df97c" Jan 21 11:27:45 crc kubenswrapper[4824]: I0121 11:27:45.782925 4824 generic.go:334] "Generic (PLEG): container finished" podID="305dccfd-067c-4062-b9d1-0200e105e8fc" containerID="439d46cbb2fda2e5e73bdff7822009c565ff5eca78f542d8aa2a37a671b5451c" exitCode=1 Jan 21 11:27:45 crc kubenswrapper[4824]: I0121 11:27:45.783028 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-58f47bdb5-gk4p7" event={"ID":"305dccfd-067c-4062-b9d1-0200e105e8fc","Type":"ContainerDied","Data":"439d46cbb2fda2e5e73bdff7822009c565ff5eca78f542d8aa2a37a671b5451c"} Jan 21 11:27:45 crc kubenswrapper[4824]: I0121 11:27:45.783610 4824 scope.go:117] "RemoveContainer" containerID="439d46cbb2fda2e5e73bdff7822009c565ff5eca78f542d8aa2a37a671b5451c" Jan 21 11:27:45 crc kubenswrapper[4824]: I0121 11:27:45.790388 4824 generic.go:334] "Generic (PLEG): container finished" podID="f5046792-741c-487c-adb8-43502e9fdba1" containerID="8eff8df136d30e860308b8a313f1f2d3a512dae3641e1576295eeeb32d6ca5d6" exitCode=1 Jan 21 11:27:45 crc kubenswrapper[4824]: I0121 11:27:45.790423 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-7b744b9d97-zkcv4" event={"ID":"f5046792-741c-487c-adb8-43502e9fdba1","Type":"ContainerDied","Data":"8eff8df136d30e860308b8a313f1f2d3a512dae3641e1576295eeeb32d6ca5d6"} Jan 21 11:27:45 crc kubenswrapper[4824]: I0121 11:27:45.790825 4824 scope.go:117] "RemoveContainer" containerID="8eff8df136d30e860308b8a313f1f2d3a512dae3641e1576295eeeb32d6ca5d6" Jan 21 11:27:45 crc kubenswrapper[4824]: I0121 11:27:45.793487 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-5dc8c6bf78-szsb6"] Jan 21 11:27:45 crc kubenswrapper[4824]: I0121 11:27:45.809166 4824 scope.go:117] "RemoveContainer" containerID="a355cd0ecd36eae3a40739bece4826924a58c416690883d3d7de3207672df97c" Jan 21 11:27:45 crc kubenswrapper[4824]: E0121 11:27:45.811105 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a355cd0ecd36eae3a40739bece4826924a58c416690883d3d7de3207672df97c\": container with ID starting with a355cd0ecd36eae3a40739bece4826924a58c416690883d3d7de3207672df97c not found: ID does not exist" containerID="a355cd0ecd36eae3a40739bece4826924a58c416690883d3d7de3207672df97c" Jan 21 11:27:45 crc kubenswrapper[4824]: I0121 11:27:45.811139 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a355cd0ecd36eae3a40739bece4826924a58c416690883d3d7de3207672df97c"} err="failed to get container status \"a355cd0ecd36eae3a40739bece4826924a58c416690883d3d7de3207672df97c\": rpc error: code = NotFound desc = could not find container \"a355cd0ecd36eae3a40739bece4826924a58c416690883d3d7de3207672df97c\": container with ID starting with a355cd0ecd36eae3a40739bece4826924a58c416690883d3d7de3207672df97c not found: ID does not exist" Jan 21 11:27:45 crc kubenswrapper[4824]: I0121 11:27:45.822047 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-api-5dc8c6bf78-szsb6"] Jan 21 11:27:45 crc kubenswrapper[4824]: I0121 11:27:45.829560 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-engine-5bc6fc99dd-nv45v" podStartSLOduration=6.829544829 podStartE2EDuration="6.829544829s" podCreationTimestamp="2026-01-21 11:27:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:27:45.797452105 +0000 UTC m=+1008.090481397" watchObservedRunningTime="2026-01-21 11:27:45.829544829 +0000 UTC m=+1008.122574121" Jan 21 11:27:45 crc kubenswrapper[4824]: I0121 11:27:45.839381 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-cfnapi-68b8cf894c-ws5zd" podStartSLOduration=5.839370202 podStartE2EDuration="5.839370202s" podCreationTimestamp="2026-01-21 11:27:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:27:45.814554643 +0000 UTC m=+1008.107583936" watchObservedRunningTime="2026-01-21 11:27:45.839370202 +0000 UTC m=+1008.132399494" Jan 21 11:27:45 crc kubenswrapper[4824]: I0121 11:27:45.849911 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-667b46bf4d-5h2cq"] Jan 21 11:27:45 crc kubenswrapper[4824]: I0121 11:27:45.857155 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-cfnapi-667b46bf4d-5h2cq"] Jan 21 11:27:46 crc kubenswrapper[4824]: I0121 11:27:46.057843 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25ee21eb-c0b8-4d49-8906-4e4144abf6fc" path="/var/lib/kubelet/pods/25ee21eb-c0b8-4d49-8906-4e4144abf6fc/volumes" Jan 21 11:27:46 crc kubenswrapper[4824]: I0121 11:27:46.058431 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2aef2f5b-bb6b-41dd-a3eb-372e7e777122" path="/var/lib/kubelet/pods/2aef2f5b-bb6b-41dd-a3eb-372e7e777122/volumes" Jan 21 11:27:46 crc kubenswrapper[4824]: I0121 11:27:46.058908 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="661c81f6-6854-416e-aea7-1344a635f2db" path="/var/lib/kubelet/pods/661c81f6-6854-416e-aea7-1344a635f2db/volumes" Jan 21 11:27:46 crc kubenswrapper[4824]: I0121 11:27:46.277369 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Jan 21 11:27:46 crc kubenswrapper[4824]: I0121 11:27:46.277430 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Jan 21 11:27:46 crc kubenswrapper[4824]: I0121 11:27:46.300749 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Jan 21 11:27:46 crc kubenswrapper[4824]: I0121 11:27:46.311739 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Jan 21 11:27:46 crc kubenswrapper[4824]: I0121 11:27:46.799900 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"964436a3-7963-4a72-bcb6-b7bab7e96574","Type":"ContainerStarted","Data":"046096c43cf9529b49344992f75e1cb1734cdd1c3f1019adb2b092d0c61ebda2"} Jan 21 11:27:46 crc kubenswrapper[4824]: I0121 11:27:46.800163 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"964436a3-7963-4a72-bcb6-b7bab7e96574","Type":"ContainerStarted","Data":"2dc1ee5999daf0317130af25e5ebcd932c53ba5e258d3f19c1936078e8e8c494"} Jan 21 11:27:46 crc kubenswrapper[4824]: I0121 11:27:46.803661 4824 generic.go:334] "Generic (PLEG): container finished" podID="305dccfd-067c-4062-b9d1-0200e105e8fc" containerID="215f40ded6896e576cc20a6a9c66190a3349ec78629b2c8a46b0029532467862" exitCode=1 Jan 21 11:27:46 crc kubenswrapper[4824]: I0121 11:27:46.803728 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-58f47bdb5-gk4p7" event={"ID":"305dccfd-067c-4062-b9d1-0200e105e8fc","Type":"ContainerDied","Data":"215f40ded6896e576cc20a6a9c66190a3349ec78629b2c8a46b0029532467862"} Jan 21 11:27:46 crc kubenswrapper[4824]: I0121 11:27:46.803773 4824 scope.go:117] "RemoveContainer" containerID="439d46cbb2fda2e5e73bdff7822009c565ff5eca78f542d8aa2a37a671b5451c" Jan 21 11:27:46 crc kubenswrapper[4824]: I0121 11:27:46.804211 4824 scope.go:117] "RemoveContainer" containerID="215f40ded6896e576cc20a6a9c66190a3349ec78629b2c8a46b0029532467862" Jan 21 11:27:46 crc kubenswrapper[4824]: E0121 11:27:46.804483 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-cfnapi\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-cfnapi pod=heat-cfnapi-58f47bdb5-gk4p7_openstack(305dccfd-067c-4062-b9d1-0200e105e8fc)\"" pod="openstack/heat-cfnapi-58f47bdb5-gk4p7" podUID="305dccfd-067c-4062-b9d1-0200e105e8fc" Jan 21 11:27:46 crc kubenswrapper[4824]: I0121 11:27:46.813454 4824 generic.go:334] "Generic (PLEG): container finished" podID="f5046792-741c-487c-adb8-43502e9fdba1" containerID="e58cf2bf1212538c2e3f23913c9a7e61c93782cc4b46b33a029ec428d40554cc" exitCode=1 Jan 21 11:27:46 crc kubenswrapper[4824]: I0121 11:27:46.822784 4824 scope.go:117] "RemoveContainer" containerID="e58cf2bf1212538c2e3f23913c9a7e61c93782cc4b46b33a029ec428d40554cc" Jan 21 11:27:46 crc kubenswrapper[4824]: E0121 11:27:46.823093 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-api\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-api pod=heat-api-7b744b9d97-zkcv4_openstack(f5046792-741c-487c-adb8-43502e9fdba1)\"" pod="openstack/heat-api-7b744b9d97-zkcv4" podUID="f5046792-741c-487c-adb8-43502e9fdba1" Jan 21 11:27:46 crc kubenswrapper[4824]: I0121 11:27:46.824085 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-7b744b9d97-zkcv4" event={"ID":"f5046792-741c-487c-adb8-43502e9fdba1","Type":"ContainerDied","Data":"e58cf2bf1212538c2e3f23913c9a7e61c93782cc4b46b33a029ec428d40554cc"} Jan 21 11:27:46 crc kubenswrapper[4824]: I0121 11:27:46.824119 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Jan 21 11:27:46 crc kubenswrapper[4824]: I0121 11:27:46.824266 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Jan 21 11:27:46 crc kubenswrapper[4824]: I0121 11:27:46.869948 4824 scope.go:117] "RemoveContainer" containerID="8eff8df136d30e860308b8a313f1f2d3a512dae3641e1576295eeeb32d6ca5d6" Jan 21 11:27:47 crc kubenswrapper[4824]: I0121 11:27:47.822949 4824 scope.go:117] "RemoveContainer" containerID="e58cf2bf1212538c2e3f23913c9a7e61c93782cc4b46b33a029ec428d40554cc" Jan 21 11:27:47 crc kubenswrapper[4824]: I0121 11:27:47.824164 4824 scope.go:117] "RemoveContainer" containerID="215f40ded6896e576cc20a6a9c66190a3349ec78629b2c8a46b0029532467862" Jan 21 11:27:47 crc kubenswrapper[4824]: E0121 11:27:47.824360 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-cfnapi\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-cfnapi pod=heat-cfnapi-58f47bdb5-gk4p7_openstack(305dccfd-067c-4062-b9d1-0200e105e8fc)\"" pod="openstack/heat-cfnapi-58f47bdb5-gk4p7" podUID="305dccfd-067c-4062-b9d1-0200e105e8fc" Jan 21 11:27:47 crc kubenswrapper[4824]: E0121 11:27:47.824642 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-api\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-api pod=heat-api-7b744b9d97-zkcv4_openstack(f5046792-741c-487c-adb8-43502e9fdba1)\"" pod="openstack/heat-api-7b744b9d97-zkcv4" podUID="f5046792-741c-487c-adb8-43502e9fdba1" Jan 21 11:27:48 crc kubenswrapper[4824]: I0121 11:27:48.455883 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Jan 21 11:27:48 crc kubenswrapper[4824]: I0121 11:27:48.462255 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Jan 21 11:27:48 crc kubenswrapper[4824]: I0121 11:27:48.832867 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"964436a3-7963-4a72-bcb6-b7bab7e96574","Type":"ContainerStarted","Data":"526650c893e6734287566bc28f16c206ab2aa5ea8b01189c5e98ecfd8b1c17d6"} Jan 21 11:27:48 crc kubenswrapper[4824]: I0121 11:27:48.833137 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="964436a3-7963-4a72-bcb6-b7bab7e96574" containerName="ceilometer-central-agent" containerID="cri-o://1746442b941230e3b55e2a919cace736ddde3df751e7bdf5c201a6243b55e8fc" gracePeriod=30 Jan 21 11:27:48 crc kubenswrapper[4824]: I0121 11:27:48.833223 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="964436a3-7963-4a72-bcb6-b7bab7e96574" containerName="proxy-httpd" containerID="cri-o://526650c893e6734287566bc28f16c206ab2aa5ea8b01189c5e98ecfd8b1c17d6" gracePeriod=30 Jan 21 11:27:48 crc kubenswrapper[4824]: I0121 11:27:48.833265 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="964436a3-7963-4a72-bcb6-b7bab7e96574" containerName="sg-core" containerID="cri-o://046096c43cf9529b49344992f75e1cb1734cdd1c3f1019adb2b092d0c61ebda2" gracePeriod=30 Jan 21 11:27:48 crc kubenswrapper[4824]: I0121 11:27:48.833294 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="964436a3-7963-4a72-bcb6-b7bab7e96574" containerName="ceilometer-notification-agent" containerID="cri-o://2dc1ee5999daf0317130af25e5ebcd932c53ba5e258d3f19c1936078e8e8c494" gracePeriod=30 Jan 21 11:27:48 crc kubenswrapper[4824]: I0121 11:27:48.851998 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=6.112202346 podStartE2EDuration="9.851951037s" podCreationTimestamp="2026-01-21 11:27:39 +0000 UTC" firstStartedPulling="2026-01-21 11:27:44.470505931 +0000 UTC m=+1006.763535223" lastFinishedPulling="2026-01-21 11:27:48.210254621 +0000 UTC m=+1010.503283914" observedRunningTime="2026-01-21 11:27:48.847988472 +0000 UTC m=+1011.141017765" watchObservedRunningTime="2026-01-21 11:27:48.851951037 +0000 UTC m=+1011.144980328" Jan 21 11:27:49 crc kubenswrapper[4824]: I0121 11:27:49.724077 4824 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/heat-cfnapi-58f47bdb5-gk4p7" Jan 21 11:27:49 crc kubenswrapper[4824]: I0121 11:27:49.724302 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-cfnapi-58f47bdb5-gk4p7" Jan 21 11:27:49 crc kubenswrapper[4824]: I0121 11:27:49.724912 4824 scope.go:117] "RemoveContainer" containerID="215f40ded6896e576cc20a6a9c66190a3349ec78629b2c8a46b0029532467862" Jan 21 11:27:49 crc kubenswrapper[4824]: E0121 11:27:49.725257 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-cfnapi\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-cfnapi pod=heat-cfnapi-58f47bdb5-gk4p7_openstack(305dccfd-067c-4062-b9d1-0200e105e8fc)\"" pod="openstack/heat-cfnapi-58f47bdb5-gk4p7" podUID="305dccfd-067c-4062-b9d1-0200e105e8fc" Jan 21 11:27:49 crc kubenswrapper[4824]: I0121 11:27:49.749618 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-api-7b744b9d97-zkcv4" Jan 21 11:27:49 crc kubenswrapper[4824]: I0121 11:27:49.749666 4824 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/heat-api-7b744b9d97-zkcv4" Jan 21 11:27:49 crc kubenswrapper[4824]: I0121 11:27:49.750280 4824 scope.go:117] "RemoveContainer" containerID="e58cf2bf1212538c2e3f23913c9a7e61c93782cc4b46b33a029ec428d40554cc" Jan 21 11:27:49 crc kubenswrapper[4824]: E0121 11:27:49.750533 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-api\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-api pod=heat-api-7b744b9d97-zkcv4_openstack(f5046792-741c-487c-adb8-43502e9fdba1)\"" pod="openstack/heat-api-7b744b9d97-zkcv4" podUID="f5046792-741c-487c-adb8-43502e9fdba1" Jan 21 11:27:49 crc kubenswrapper[4824]: I0121 11:27:49.842722 4824 generic.go:334] "Generic (PLEG): container finished" podID="964436a3-7963-4a72-bcb6-b7bab7e96574" containerID="526650c893e6734287566bc28f16c206ab2aa5ea8b01189c5e98ecfd8b1c17d6" exitCode=0 Jan 21 11:27:49 crc kubenswrapper[4824]: I0121 11:27:49.842754 4824 generic.go:334] "Generic (PLEG): container finished" podID="964436a3-7963-4a72-bcb6-b7bab7e96574" containerID="046096c43cf9529b49344992f75e1cb1734cdd1c3f1019adb2b092d0c61ebda2" exitCode=2 Jan 21 11:27:49 crc kubenswrapper[4824]: I0121 11:27:49.842793 4824 generic.go:334] "Generic (PLEG): container finished" podID="964436a3-7963-4a72-bcb6-b7bab7e96574" containerID="2dc1ee5999daf0317130af25e5ebcd932c53ba5e258d3f19c1936078e8e8c494" exitCode=0 Jan 21 11:27:49 crc kubenswrapper[4824]: I0121 11:27:49.842782 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"964436a3-7963-4a72-bcb6-b7bab7e96574","Type":"ContainerDied","Data":"526650c893e6734287566bc28f16c206ab2aa5ea8b01189c5e98ecfd8b1c17d6"} Jan 21 11:27:49 crc kubenswrapper[4824]: I0121 11:27:49.842825 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"964436a3-7963-4a72-bcb6-b7bab7e96574","Type":"ContainerDied","Data":"046096c43cf9529b49344992f75e1cb1734cdd1c3f1019adb2b092d0c61ebda2"} Jan 21 11:27:49 crc kubenswrapper[4824]: I0121 11:27:49.842836 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"964436a3-7963-4a72-bcb6-b7bab7e96574","Type":"ContainerDied","Data":"2dc1ee5999daf0317130af25e5ebcd932c53ba5e258d3f19c1936078e8e8c494"} Jan 21 11:27:50 crc kubenswrapper[4824]: I0121 11:27:50.850861 4824 generic.go:334] "Generic (PLEG): container finished" podID="9c73eda8-d376-4e18-9b33-f5ac4d10d091" containerID="c2bb2e93e558c6a9e9ce5d9ae43e9812275d36c293fd644336fd7e9fcc3ba6c0" exitCode=0 Jan 21 11:27:50 crc kubenswrapper[4824]: I0121 11:27:50.850905 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-k6zm8" event={"ID":"9c73eda8-d376-4e18-9b33-f5ac4d10d091","Type":"ContainerDied","Data":"c2bb2e93e558c6a9e9ce5d9ae43e9812275d36c293fd644336fd7e9fcc3ba6c0"} Jan 21 11:27:51 crc kubenswrapper[4824]: I0121 11:27:51.904290 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-cfnapi-68b8cf894c-ws5zd" Jan 21 11:27:51 crc kubenswrapper[4824]: I0121 11:27:51.943596 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-api-549445f5fc-fwxlz" Jan 21 11:27:51 crc kubenswrapper[4824]: I0121 11:27:51.963429 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-58f47bdb5-gk4p7"] Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.009054 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-7b744b9d97-zkcv4"] Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.248714 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-k6zm8" Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.344490 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c73eda8-d376-4e18-9b33-f5ac4d10d091-combined-ca-bundle\") pod \"9c73eda8-d376-4e18-9b33-f5ac4d10d091\" (UID: \"9c73eda8-d376-4e18-9b33-f5ac4d10d091\") " Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.344561 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5rs4n\" (UniqueName: \"kubernetes.io/projected/9c73eda8-d376-4e18-9b33-f5ac4d10d091-kube-api-access-5rs4n\") pod \"9c73eda8-d376-4e18-9b33-f5ac4d10d091\" (UID: \"9c73eda8-d376-4e18-9b33-f5ac4d10d091\") " Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.344669 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c73eda8-d376-4e18-9b33-f5ac4d10d091-config-data\") pod \"9c73eda8-d376-4e18-9b33-f5ac4d10d091\" (UID: \"9c73eda8-d376-4e18-9b33-f5ac4d10d091\") " Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.344710 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9c73eda8-d376-4e18-9b33-f5ac4d10d091-scripts\") pod \"9c73eda8-d376-4e18-9b33-f5ac4d10d091\" (UID: \"9c73eda8-d376-4e18-9b33-f5ac4d10d091\") " Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.350566 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9c73eda8-d376-4e18-9b33-f5ac4d10d091-scripts" (OuterVolumeSpecName: "scripts") pod "9c73eda8-d376-4e18-9b33-f5ac4d10d091" (UID: "9c73eda8-d376-4e18-9b33-f5ac4d10d091"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.358809 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9c73eda8-d376-4e18-9b33-f5ac4d10d091-kube-api-access-5rs4n" (OuterVolumeSpecName: "kube-api-access-5rs4n") pod "9c73eda8-d376-4e18-9b33-f5ac4d10d091" (UID: "9c73eda8-d376-4e18-9b33-f5ac4d10d091"). InnerVolumeSpecName "kube-api-access-5rs4n". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.374555 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9c73eda8-d376-4e18-9b33-f5ac4d10d091-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9c73eda8-d376-4e18-9b33-f5ac4d10d091" (UID: "9c73eda8-d376-4e18-9b33-f5ac4d10d091"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.379511 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9c73eda8-d376-4e18-9b33-f5ac4d10d091-config-data" (OuterVolumeSpecName: "config-data") pod "9c73eda8-d376-4e18-9b33-f5ac4d10d091" (UID: "9c73eda8-d376-4e18-9b33-f5ac4d10d091"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.446380 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c73eda8-d376-4e18-9b33-f5ac4d10d091-config-data\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.446409 4824 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9c73eda8-d376-4e18-9b33-f5ac4d10d091-scripts\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.446427 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c73eda8-d376-4e18-9b33-f5ac4d10d091-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.446437 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5rs4n\" (UniqueName: \"kubernetes.io/projected/9c73eda8-d376-4e18-9b33-f5ac4d10d091-kube-api-access-5rs4n\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.510556 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-58f47bdb5-gk4p7" Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.515748 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-7b744b9d97-zkcv4" Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.649629 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gsb2r\" (UniqueName: \"kubernetes.io/projected/305dccfd-067c-4062-b9d1-0200e105e8fc-kube-api-access-gsb2r\") pod \"305dccfd-067c-4062-b9d1-0200e105e8fc\" (UID: \"305dccfd-067c-4062-b9d1-0200e105e8fc\") " Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.649665 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/305dccfd-067c-4062-b9d1-0200e105e8fc-config-data-custom\") pod \"305dccfd-067c-4062-b9d1-0200e105e8fc\" (UID: \"305dccfd-067c-4062-b9d1-0200e105e8fc\") " Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.649685 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4787\" (UniqueName: \"kubernetes.io/projected/f5046792-741c-487c-adb8-43502e9fdba1-kube-api-access-x4787\") pod \"f5046792-741c-487c-adb8-43502e9fdba1\" (UID: \"f5046792-741c-487c-adb8-43502e9fdba1\") " Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.649804 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/305dccfd-067c-4062-b9d1-0200e105e8fc-combined-ca-bundle\") pod \"305dccfd-067c-4062-b9d1-0200e105e8fc\" (UID: \"305dccfd-067c-4062-b9d1-0200e105e8fc\") " Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.649832 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/305dccfd-067c-4062-b9d1-0200e105e8fc-config-data\") pod \"305dccfd-067c-4062-b9d1-0200e105e8fc\" (UID: \"305dccfd-067c-4062-b9d1-0200e105e8fc\") " Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.649899 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5046792-741c-487c-adb8-43502e9fdba1-combined-ca-bundle\") pod \"f5046792-741c-487c-adb8-43502e9fdba1\" (UID: \"f5046792-741c-487c-adb8-43502e9fdba1\") " Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.649948 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f5046792-741c-487c-adb8-43502e9fdba1-config-data-custom\") pod \"f5046792-741c-487c-adb8-43502e9fdba1\" (UID: \"f5046792-741c-487c-adb8-43502e9fdba1\") " Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.650000 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f5046792-741c-487c-adb8-43502e9fdba1-config-data\") pod \"f5046792-741c-487c-adb8-43502e9fdba1\" (UID: \"f5046792-741c-487c-adb8-43502e9fdba1\") " Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.670092 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/305dccfd-067c-4062-b9d1-0200e105e8fc-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "305dccfd-067c-4062-b9d1-0200e105e8fc" (UID: "305dccfd-067c-4062-b9d1-0200e105e8fc"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.675079 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/305dccfd-067c-4062-b9d1-0200e105e8fc-kube-api-access-gsb2r" (OuterVolumeSpecName: "kube-api-access-gsb2r") pod "305dccfd-067c-4062-b9d1-0200e105e8fc" (UID: "305dccfd-067c-4062-b9d1-0200e105e8fc"). InnerVolumeSpecName "kube-api-access-gsb2r". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.675103 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f5046792-741c-487c-adb8-43502e9fdba1-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "f5046792-741c-487c-adb8-43502e9fdba1" (UID: "f5046792-741c-487c-adb8-43502e9fdba1"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.685821 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f5046792-741c-487c-adb8-43502e9fdba1-kube-api-access-x4787" (OuterVolumeSpecName: "kube-api-access-x4787") pod "f5046792-741c-487c-adb8-43502e9fdba1" (UID: "f5046792-741c-487c-adb8-43502e9fdba1"). InnerVolumeSpecName "kube-api-access-x4787". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.758207 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f5046792-741c-487c-adb8-43502e9fdba1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f5046792-741c-487c-adb8-43502e9fdba1" (UID: "f5046792-741c-487c-adb8-43502e9fdba1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.759396 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5046792-741c-487c-adb8-43502e9fdba1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.759416 4824 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f5046792-741c-487c-adb8-43502e9fdba1-config-data-custom\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.759426 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gsb2r\" (UniqueName: \"kubernetes.io/projected/305dccfd-067c-4062-b9d1-0200e105e8fc-kube-api-access-gsb2r\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.759435 4824 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/305dccfd-067c-4062-b9d1-0200e105e8fc-config-data-custom\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.759444 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4787\" (UniqueName: \"kubernetes.io/projected/f5046792-741c-487c-adb8-43502e9fdba1-kube-api-access-x4787\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.779107 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/305dccfd-067c-4062-b9d1-0200e105e8fc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "305dccfd-067c-4062-b9d1-0200e105e8fc" (UID: "305dccfd-067c-4062-b9d1-0200e105e8fc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.779177 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-engine-656b845fb4-qst2s" Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.785569 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f5046792-741c-487c-adb8-43502e9fdba1-config-data" (OuterVolumeSpecName: "config-data") pod "f5046792-741c-487c-adb8-43502e9fdba1" (UID: "f5046792-741c-487c-adb8-43502e9fdba1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.806292 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/305dccfd-067c-4062-b9d1-0200e105e8fc-config-data" (OuterVolumeSpecName: "config-data") pod "305dccfd-067c-4062-b9d1-0200e105e8fc" (UID: "305dccfd-067c-4062-b9d1-0200e105e8fc"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.834104 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.861683 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/305dccfd-067c-4062-b9d1-0200e105e8fc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.861709 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/305dccfd-067c-4062-b9d1-0200e105e8fc-config-data\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.861719 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f5046792-741c-487c-adb8-43502e9fdba1-config-data\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.877664 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-58f47bdb5-gk4p7" event={"ID":"305dccfd-067c-4062-b9d1-0200e105e8fc","Type":"ContainerDied","Data":"88223e1727fde2b87a86e2d66dfde6e4b257cc1ef3f6ba548c34ddd1445f80ea"} Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.878489 4824 scope.go:117] "RemoveContainer" containerID="215f40ded6896e576cc20a6a9c66190a3349ec78629b2c8a46b0029532467862" Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.878616 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-58f47bdb5-gk4p7" Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.882730 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-7b744b9d97-zkcv4" Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.884026 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-7b744b9d97-zkcv4" event={"ID":"f5046792-741c-487c-adb8-43502e9fdba1","Type":"ContainerDied","Data":"8b94a9402348e00a388234d7d2563fa253b97825d6e2edc2c72d291efaccf943"} Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.907248 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-k6zm8" event={"ID":"9c73eda8-d376-4e18-9b33-f5ac4d10d091","Type":"ContainerDied","Data":"bec80828734b592cc51fee125cf162b17f1a9431a491c60a24c542b64ae9e046"} Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.907280 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bec80828734b592cc51fee125cf162b17f1a9431a491c60a24c542b64ae9e046" Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.907289 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-k6zm8" Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.912349 4824 scope.go:117] "RemoveContainer" containerID="e58cf2bf1212538c2e3f23913c9a7e61c93782cc4b46b33a029ec428d40554cc" Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.913367 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"964436a3-7963-4a72-bcb6-b7bab7e96574","Type":"ContainerDied","Data":"1746442b941230e3b55e2a919cace736ddde3df751e7bdf5c201a6243b55e8fc"} Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.914753 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.922126 4824 generic.go:334] "Generic (PLEG): container finished" podID="964436a3-7963-4a72-bcb6-b7bab7e96574" containerID="1746442b941230e3b55e2a919cace736ddde3df751e7bdf5c201a6243b55e8fc" exitCode=0 Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.922193 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"964436a3-7963-4a72-bcb6-b7bab7e96574","Type":"ContainerDied","Data":"cd894bcfd6b429a651ad2e896089b4e35850d9690bc597b40bbe1fa904affe57"} Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.938106 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-58f47bdb5-gk4p7"] Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.949650 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-cfnapi-58f47bdb5-gk4p7"] Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.951867 4824 scope.go:117] "RemoveContainer" containerID="526650c893e6734287566bc28f16c206ab2aa5ea8b01189c5e98ecfd8b1c17d6" Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.963588 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-928nq\" (UniqueName: \"kubernetes.io/projected/964436a3-7963-4a72-bcb6-b7bab7e96574-kube-api-access-928nq\") pod \"964436a3-7963-4a72-bcb6-b7bab7e96574\" (UID: \"964436a3-7963-4a72-bcb6-b7bab7e96574\") " Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.963685 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/964436a3-7963-4a72-bcb6-b7bab7e96574-log-httpd\") pod \"964436a3-7963-4a72-bcb6-b7bab7e96574\" (UID: \"964436a3-7963-4a72-bcb6-b7bab7e96574\") " Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.963756 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/964436a3-7963-4a72-bcb6-b7bab7e96574-scripts\") pod \"964436a3-7963-4a72-bcb6-b7bab7e96574\" (UID: \"964436a3-7963-4a72-bcb6-b7bab7e96574\") " Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.963810 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/964436a3-7963-4a72-bcb6-b7bab7e96574-run-httpd\") pod \"964436a3-7963-4a72-bcb6-b7bab7e96574\" (UID: \"964436a3-7963-4a72-bcb6-b7bab7e96574\") " Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.963927 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/964436a3-7963-4a72-bcb6-b7bab7e96574-sg-core-conf-yaml\") pod \"964436a3-7963-4a72-bcb6-b7bab7e96574\" (UID: \"964436a3-7963-4a72-bcb6-b7bab7e96574\") " Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.964019 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/964436a3-7963-4a72-bcb6-b7bab7e96574-config-data\") pod \"964436a3-7963-4a72-bcb6-b7bab7e96574\" (UID: \"964436a3-7963-4a72-bcb6-b7bab7e96574\") " Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.964086 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/964436a3-7963-4a72-bcb6-b7bab7e96574-combined-ca-bundle\") pod \"964436a3-7963-4a72-bcb6-b7bab7e96574\" (UID: \"964436a3-7963-4a72-bcb6-b7bab7e96574\") " Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.964262 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-7b744b9d97-zkcv4"] Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.964314 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/964436a3-7963-4a72-bcb6-b7bab7e96574-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "964436a3-7963-4a72-bcb6-b7bab7e96574" (UID: "964436a3-7963-4a72-bcb6-b7bab7e96574"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.964540 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/964436a3-7963-4a72-bcb6-b7bab7e96574-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "964436a3-7963-4a72-bcb6-b7bab7e96574" (UID: "964436a3-7963-4a72-bcb6-b7bab7e96574"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.973121 4824 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/964436a3-7963-4a72-bcb6-b7bab7e96574-log-httpd\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.973142 4824 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/964436a3-7963-4a72-bcb6-b7bab7e96574-run-httpd\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.975103 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/964436a3-7963-4a72-bcb6-b7bab7e96574-scripts" (OuterVolumeSpecName: "scripts") pod "964436a3-7963-4a72-bcb6-b7bab7e96574" (UID: "964436a3-7963-4a72-bcb6-b7bab7e96574"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.977310 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-api-7b744b9d97-zkcv4"] Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.979168 4824 scope.go:117] "RemoveContainer" containerID="046096c43cf9529b49344992f75e1cb1734cdd1c3f1019adb2b092d0c61ebda2" Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.985846 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/964436a3-7963-4a72-bcb6-b7bab7e96574-kube-api-access-928nq" (OuterVolumeSpecName: "kube-api-access-928nq") pod "964436a3-7963-4a72-bcb6-b7bab7e96574" (UID: "964436a3-7963-4a72-bcb6-b7bab7e96574"). InnerVolumeSpecName "kube-api-access-928nq". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.988915 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Jan 21 11:27:52 crc kubenswrapper[4824]: E0121 11:27:52.989500 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="964436a3-7963-4a72-bcb6-b7bab7e96574" containerName="sg-core" Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.989513 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="964436a3-7963-4a72-bcb6-b7bab7e96574" containerName="sg-core" Jan 21 11:27:52 crc kubenswrapper[4824]: E0121 11:27:52.989525 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="964436a3-7963-4a72-bcb6-b7bab7e96574" containerName="ceilometer-notification-agent" Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.989532 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="964436a3-7963-4a72-bcb6-b7bab7e96574" containerName="ceilometer-notification-agent" Jan 21 11:27:52 crc kubenswrapper[4824]: E0121 11:27:52.989542 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="661c81f6-6854-416e-aea7-1344a635f2db" containerName="dnsmasq-dns" Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.989547 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="661c81f6-6854-416e-aea7-1344a635f2db" containerName="dnsmasq-dns" Jan 21 11:27:52 crc kubenswrapper[4824]: E0121 11:27:52.989554 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="661c81f6-6854-416e-aea7-1344a635f2db" containerName="init" Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.989559 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="661c81f6-6854-416e-aea7-1344a635f2db" containerName="init" Jan 21 11:27:52 crc kubenswrapper[4824]: E0121 11:27:52.989573 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="305dccfd-067c-4062-b9d1-0200e105e8fc" containerName="heat-cfnapi" Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.989578 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="305dccfd-067c-4062-b9d1-0200e105e8fc" containerName="heat-cfnapi" Jan 21 11:27:52 crc kubenswrapper[4824]: E0121 11:27:52.989585 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5046792-741c-487c-adb8-43502e9fdba1" containerName="heat-api" Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.989590 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5046792-741c-487c-adb8-43502e9fdba1" containerName="heat-api" Jan 21 11:27:52 crc kubenswrapper[4824]: E0121 11:27:52.989597 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="964436a3-7963-4a72-bcb6-b7bab7e96574" containerName="proxy-httpd" Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.989602 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="964436a3-7963-4a72-bcb6-b7bab7e96574" containerName="proxy-httpd" Jan 21 11:27:52 crc kubenswrapper[4824]: E0121 11:27:52.989611 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2aef2f5b-bb6b-41dd-a3eb-372e7e777122" containerName="heat-api" Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.989616 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="2aef2f5b-bb6b-41dd-a3eb-372e7e777122" containerName="heat-api" Jan 21 11:27:52 crc kubenswrapper[4824]: E0121 11:27:52.989626 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c73eda8-d376-4e18-9b33-f5ac4d10d091" containerName="nova-cell0-conductor-db-sync" Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.989631 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c73eda8-d376-4e18-9b33-f5ac4d10d091" containerName="nova-cell0-conductor-db-sync" Jan 21 11:27:52 crc kubenswrapper[4824]: E0121 11:27:52.989638 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5046792-741c-487c-adb8-43502e9fdba1" containerName="heat-api" Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.989643 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5046792-741c-487c-adb8-43502e9fdba1" containerName="heat-api" Jan 21 11:27:52 crc kubenswrapper[4824]: E0121 11:27:52.989653 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="964436a3-7963-4a72-bcb6-b7bab7e96574" containerName="ceilometer-central-agent" Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.989659 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="964436a3-7963-4a72-bcb6-b7bab7e96574" containerName="ceilometer-central-agent" Jan 21 11:27:52 crc kubenswrapper[4824]: E0121 11:27:52.989673 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="25ee21eb-c0b8-4d49-8906-4e4144abf6fc" containerName="heat-cfnapi" Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.989678 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="25ee21eb-c0b8-4d49-8906-4e4144abf6fc" containerName="heat-cfnapi" Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.989849 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="f5046792-741c-487c-adb8-43502e9fdba1" containerName="heat-api" Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.989859 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="964436a3-7963-4a72-bcb6-b7bab7e96574" containerName="proxy-httpd" Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.989866 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="964436a3-7963-4a72-bcb6-b7bab7e96574" containerName="sg-core" Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.989876 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="f5046792-741c-487c-adb8-43502e9fdba1" containerName="heat-api" Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.989888 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="9c73eda8-d376-4e18-9b33-f5ac4d10d091" containerName="nova-cell0-conductor-db-sync" Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.989897 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="25ee21eb-c0b8-4d49-8906-4e4144abf6fc" containerName="heat-cfnapi" Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.989908 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="305dccfd-067c-4062-b9d1-0200e105e8fc" containerName="heat-cfnapi" Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.989915 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="661c81f6-6854-416e-aea7-1344a635f2db" containerName="dnsmasq-dns" Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.989923 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="2aef2f5b-bb6b-41dd-a3eb-372e7e777122" containerName="heat-api" Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.989936 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="964436a3-7963-4a72-bcb6-b7bab7e96574" containerName="ceilometer-central-agent" Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.989945 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="964436a3-7963-4a72-bcb6-b7bab7e96574" containerName="ceilometer-notification-agent" Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.990491 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.992393 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.992777 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-gbh62" Jan 21 11:27:52 crc kubenswrapper[4824]: I0121 11:27:52.993419 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/964436a3-7963-4a72-bcb6-b7bab7e96574-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "964436a3-7963-4a72-bcb6-b7bab7e96574" (UID: "964436a3-7963-4a72-bcb6-b7bab7e96574"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:27:53 crc kubenswrapper[4824]: I0121 11:27:53.001626 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Jan 21 11:27:53 crc kubenswrapper[4824]: I0121 11:27:53.003375 4824 scope.go:117] "RemoveContainer" containerID="2dc1ee5999daf0317130af25e5ebcd932c53ba5e258d3f19c1936078e8e8c494" Jan 21 11:27:53 crc kubenswrapper[4824]: I0121 11:27:53.027931 4824 scope.go:117] "RemoveContainer" containerID="1746442b941230e3b55e2a919cace736ddde3df751e7bdf5c201a6243b55e8fc" Jan 21 11:27:53 crc kubenswrapper[4824]: I0121 11:27:53.048510 4824 scope.go:117] "RemoveContainer" containerID="526650c893e6734287566bc28f16c206ab2aa5ea8b01189c5e98ecfd8b1c17d6" Jan 21 11:27:53 crc kubenswrapper[4824]: E0121 11:27:53.049306 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"526650c893e6734287566bc28f16c206ab2aa5ea8b01189c5e98ecfd8b1c17d6\": container with ID starting with 526650c893e6734287566bc28f16c206ab2aa5ea8b01189c5e98ecfd8b1c17d6 not found: ID does not exist" containerID="526650c893e6734287566bc28f16c206ab2aa5ea8b01189c5e98ecfd8b1c17d6" Jan 21 11:27:53 crc kubenswrapper[4824]: I0121 11:27:53.049345 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"526650c893e6734287566bc28f16c206ab2aa5ea8b01189c5e98ecfd8b1c17d6"} err="failed to get container status \"526650c893e6734287566bc28f16c206ab2aa5ea8b01189c5e98ecfd8b1c17d6\": rpc error: code = NotFound desc = could not find container \"526650c893e6734287566bc28f16c206ab2aa5ea8b01189c5e98ecfd8b1c17d6\": container with ID starting with 526650c893e6734287566bc28f16c206ab2aa5ea8b01189c5e98ecfd8b1c17d6 not found: ID does not exist" Jan 21 11:27:53 crc kubenswrapper[4824]: I0121 11:27:53.049364 4824 scope.go:117] "RemoveContainer" containerID="046096c43cf9529b49344992f75e1cb1734cdd1c3f1019adb2b092d0c61ebda2" Jan 21 11:27:53 crc kubenswrapper[4824]: E0121 11:27:53.049592 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"046096c43cf9529b49344992f75e1cb1734cdd1c3f1019adb2b092d0c61ebda2\": container with ID starting with 046096c43cf9529b49344992f75e1cb1734cdd1c3f1019adb2b092d0c61ebda2 not found: ID does not exist" containerID="046096c43cf9529b49344992f75e1cb1734cdd1c3f1019adb2b092d0c61ebda2" Jan 21 11:27:53 crc kubenswrapper[4824]: I0121 11:27:53.049613 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"046096c43cf9529b49344992f75e1cb1734cdd1c3f1019adb2b092d0c61ebda2"} err="failed to get container status \"046096c43cf9529b49344992f75e1cb1734cdd1c3f1019adb2b092d0c61ebda2\": rpc error: code = NotFound desc = could not find container \"046096c43cf9529b49344992f75e1cb1734cdd1c3f1019adb2b092d0c61ebda2\": container with ID starting with 046096c43cf9529b49344992f75e1cb1734cdd1c3f1019adb2b092d0c61ebda2 not found: ID does not exist" Jan 21 11:27:53 crc kubenswrapper[4824]: I0121 11:27:53.049626 4824 scope.go:117] "RemoveContainer" containerID="2dc1ee5999daf0317130af25e5ebcd932c53ba5e258d3f19c1936078e8e8c494" Jan 21 11:27:53 crc kubenswrapper[4824]: E0121 11:27:53.049993 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2dc1ee5999daf0317130af25e5ebcd932c53ba5e258d3f19c1936078e8e8c494\": container with ID starting with 2dc1ee5999daf0317130af25e5ebcd932c53ba5e258d3f19c1936078e8e8c494 not found: ID does not exist" containerID="2dc1ee5999daf0317130af25e5ebcd932c53ba5e258d3f19c1936078e8e8c494" Jan 21 11:27:53 crc kubenswrapper[4824]: I0121 11:27:53.050028 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2dc1ee5999daf0317130af25e5ebcd932c53ba5e258d3f19c1936078e8e8c494"} err="failed to get container status \"2dc1ee5999daf0317130af25e5ebcd932c53ba5e258d3f19c1936078e8e8c494\": rpc error: code = NotFound desc = could not find container \"2dc1ee5999daf0317130af25e5ebcd932c53ba5e258d3f19c1936078e8e8c494\": container with ID starting with 2dc1ee5999daf0317130af25e5ebcd932c53ba5e258d3f19c1936078e8e8c494 not found: ID does not exist" Jan 21 11:27:53 crc kubenswrapper[4824]: I0121 11:27:53.050051 4824 scope.go:117] "RemoveContainer" containerID="1746442b941230e3b55e2a919cace736ddde3df751e7bdf5c201a6243b55e8fc" Jan 21 11:27:53 crc kubenswrapper[4824]: E0121 11:27:53.050323 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1746442b941230e3b55e2a919cace736ddde3df751e7bdf5c201a6243b55e8fc\": container with ID starting with 1746442b941230e3b55e2a919cace736ddde3df751e7bdf5c201a6243b55e8fc not found: ID does not exist" containerID="1746442b941230e3b55e2a919cace736ddde3df751e7bdf5c201a6243b55e8fc" Jan 21 11:27:53 crc kubenswrapper[4824]: I0121 11:27:53.050344 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1746442b941230e3b55e2a919cace736ddde3df751e7bdf5c201a6243b55e8fc"} err="failed to get container status \"1746442b941230e3b55e2a919cace736ddde3df751e7bdf5c201a6243b55e8fc\": rpc error: code = NotFound desc = could not find container \"1746442b941230e3b55e2a919cace736ddde3df751e7bdf5c201a6243b55e8fc\": container with ID starting with 1746442b941230e3b55e2a919cace736ddde3df751e7bdf5c201a6243b55e8fc not found: ID does not exist" Jan 21 11:27:53 crc kubenswrapper[4824]: I0121 11:27:53.052097 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/964436a3-7963-4a72-bcb6-b7bab7e96574-config-data" (OuterVolumeSpecName: "config-data") pod "964436a3-7963-4a72-bcb6-b7bab7e96574" (UID: "964436a3-7963-4a72-bcb6-b7bab7e96574"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:27:53 crc kubenswrapper[4824]: I0121 11:27:53.064926 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/964436a3-7963-4a72-bcb6-b7bab7e96574-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "964436a3-7963-4a72-bcb6-b7bab7e96574" (UID: "964436a3-7963-4a72-bcb6-b7bab7e96574"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:27:53 crc kubenswrapper[4824]: I0121 11:27:53.076609 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ddqrh\" (UniqueName: \"kubernetes.io/projected/5d989da6-218c-4d56-b121-ce2002e90616-kube-api-access-ddqrh\") pod \"nova-cell0-conductor-0\" (UID: \"5d989da6-218c-4d56-b121-ce2002e90616\") " pod="openstack/nova-cell0-conductor-0" Jan 21 11:27:53 crc kubenswrapper[4824]: I0121 11:27:53.076659 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5d989da6-218c-4d56-b121-ce2002e90616-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"5d989da6-218c-4d56-b121-ce2002e90616\") " pod="openstack/nova-cell0-conductor-0" Jan 21 11:27:53 crc kubenswrapper[4824]: I0121 11:27:53.076733 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d989da6-218c-4d56-b121-ce2002e90616-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"5d989da6-218c-4d56-b121-ce2002e90616\") " pod="openstack/nova-cell0-conductor-0" Jan 21 11:27:53 crc kubenswrapper[4824]: I0121 11:27:53.076893 4824 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/964436a3-7963-4a72-bcb6-b7bab7e96574-scripts\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:53 crc kubenswrapper[4824]: I0121 11:27:53.076909 4824 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/964436a3-7963-4a72-bcb6-b7bab7e96574-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:53 crc kubenswrapper[4824]: I0121 11:27:53.076918 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/964436a3-7963-4a72-bcb6-b7bab7e96574-config-data\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:53 crc kubenswrapper[4824]: I0121 11:27:53.076926 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/964436a3-7963-4a72-bcb6-b7bab7e96574-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:53 crc kubenswrapper[4824]: I0121 11:27:53.076934 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-928nq\" (UniqueName: \"kubernetes.io/projected/964436a3-7963-4a72-bcb6-b7bab7e96574-kube-api-access-928nq\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:53 crc kubenswrapper[4824]: I0121 11:27:53.178447 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ddqrh\" (UniqueName: \"kubernetes.io/projected/5d989da6-218c-4d56-b121-ce2002e90616-kube-api-access-ddqrh\") pod \"nova-cell0-conductor-0\" (UID: \"5d989da6-218c-4d56-b121-ce2002e90616\") " pod="openstack/nova-cell0-conductor-0" Jan 21 11:27:53 crc kubenswrapper[4824]: I0121 11:27:53.178501 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5d989da6-218c-4d56-b121-ce2002e90616-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"5d989da6-218c-4d56-b121-ce2002e90616\") " pod="openstack/nova-cell0-conductor-0" Jan 21 11:27:53 crc kubenswrapper[4824]: I0121 11:27:53.178594 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d989da6-218c-4d56-b121-ce2002e90616-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"5d989da6-218c-4d56-b121-ce2002e90616\") " pod="openstack/nova-cell0-conductor-0" Jan 21 11:27:53 crc kubenswrapper[4824]: I0121 11:27:53.184565 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d989da6-218c-4d56-b121-ce2002e90616-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"5d989da6-218c-4d56-b121-ce2002e90616\") " pod="openstack/nova-cell0-conductor-0" Jan 21 11:27:53 crc kubenswrapper[4824]: I0121 11:27:53.186309 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5d989da6-218c-4d56-b121-ce2002e90616-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"5d989da6-218c-4d56-b121-ce2002e90616\") " pod="openstack/nova-cell0-conductor-0" Jan 21 11:27:53 crc kubenswrapper[4824]: I0121 11:27:53.191884 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ddqrh\" (UniqueName: \"kubernetes.io/projected/5d989da6-218c-4d56-b121-ce2002e90616-kube-api-access-ddqrh\") pod \"nova-cell0-conductor-0\" (UID: \"5d989da6-218c-4d56-b121-ce2002e90616\") " pod="openstack/nova-cell0-conductor-0" Jan 21 11:27:53 crc kubenswrapper[4824]: I0121 11:27:53.258340 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Jan 21 11:27:53 crc kubenswrapper[4824]: I0121 11:27:53.264787 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Jan 21 11:27:53 crc kubenswrapper[4824]: I0121 11:27:53.275925 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Jan 21 11:27:53 crc kubenswrapper[4824]: E0121 11:27:53.276267 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="305dccfd-067c-4062-b9d1-0200e105e8fc" containerName="heat-cfnapi" Jan 21 11:27:53 crc kubenswrapper[4824]: I0121 11:27:53.276286 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="305dccfd-067c-4062-b9d1-0200e105e8fc" containerName="heat-cfnapi" Jan 21 11:27:53 crc kubenswrapper[4824]: I0121 11:27:53.276475 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="305dccfd-067c-4062-b9d1-0200e105e8fc" containerName="heat-cfnapi" Jan 21 11:27:53 crc kubenswrapper[4824]: I0121 11:27:53.277893 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 21 11:27:53 crc kubenswrapper[4824]: I0121 11:27:53.279323 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Jan 21 11:27:53 crc kubenswrapper[4824]: I0121 11:27:53.279364 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Jan 21 11:27:53 crc kubenswrapper[4824]: I0121 11:27:53.293148 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Jan 21 11:27:53 crc kubenswrapper[4824]: I0121 11:27:53.309247 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Jan 21 11:27:53 crc kubenswrapper[4824]: I0121 11:27:53.384658 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4229a37f-5aa8-46c7-8ec4-74dde8937047-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"4229a37f-5aa8-46c7-8ec4-74dde8937047\") " pod="openstack/ceilometer-0" Jan 21 11:27:53 crc kubenswrapper[4824]: I0121 11:27:53.384896 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6zwmx\" (UniqueName: \"kubernetes.io/projected/4229a37f-5aa8-46c7-8ec4-74dde8937047-kube-api-access-6zwmx\") pod \"ceilometer-0\" (UID: \"4229a37f-5aa8-46c7-8ec4-74dde8937047\") " pod="openstack/ceilometer-0" Jan 21 11:27:53 crc kubenswrapper[4824]: I0121 11:27:53.384940 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4229a37f-5aa8-46c7-8ec4-74dde8937047-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"4229a37f-5aa8-46c7-8ec4-74dde8937047\") " pod="openstack/ceilometer-0" Jan 21 11:27:53 crc kubenswrapper[4824]: I0121 11:27:53.384975 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4229a37f-5aa8-46c7-8ec4-74dde8937047-config-data\") pod \"ceilometer-0\" (UID: \"4229a37f-5aa8-46c7-8ec4-74dde8937047\") " pod="openstack/ceilometer-0" Jan 21 11:27:53 crc kubenswrapper[4824]: I0121 11:27:53.385043 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4229a37f-5aa8-46c7-8ec4-74dde8937047-log-httpd\") pod \"ceilometer-0\" (UID: \"4229a37f-5aa8-46c7-8ec4-74dde8937047\") " pod="openstack/ceilometer-0" Jan 21 11:27:53 crc kubenswrapper[4824]: I0121 11:27:53.385132 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4229a37f-5aa8-46c7-8ec4-74dde8937047-scripts\") pod \"ceilometer-0\" (UID: \"4229a37f-5aa8-46c7-8ec4-74dde8937047\") " pod="openstack/ceilometer-0" Jan 21 11:27:53 crc kubenswrapper[4824]: I0121 11:27:53.385159 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4229a37f-5aa8-46c7-8ec4-74dde8937047-run-httpd\") pod \"ceilometer-0\" (UID: \"4229a37f-5aa8-46c7-8ec4-74dde8937047\") " pod="openstack/ceilometer-0" Jan 21 11:27:53 crc kubenswrapper[4824]: I0121 11:27:53.487320 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4229a37f-5aa8-46c7-8ec4-74dde8937047-run-httpd\") pod \"ceilometer-0\" (UID: \"4229a37f-5aa8-46c7-8ec4-74dde8937047\") " pod="openstack/ceilometer-0" Jan 21 11:27:53 crc kubenswrapper[4824]: I0121 11:27:53.487450 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4229a37f-5aa8-46c7-8ec4-74dde8937047-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"4229a37f-5aa8-46c7-8ec4-74dde8937047\") " pod="openstack/ceilometer-0" Jan 21 11:27:53 crc kubenswrapper[4824]: I0121 11:27:53.487532 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6zwmx\" (UniqueName: \"kubernetes.io/projected/4229a37f-5aa8-46c7-8ec4-74dde8937047-kube-api-access-6zwmx\") pod \"ceilometer-0\" (UID: \"4229a37f-5aa8-46c7-8ec4-74dde8937047\") " pod="openstack/ceilometer-0" Jan 21 11:27:53 crc kubenswrapper[4824]: I0121 11:27:53.487564 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4229a37f-5aa8-46c7-8ec4-74dde8937047-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"4229a37f-5aa8-46c7-8ec4-74dde8937047\") " pod="openstack/ceilometer-0" Jan 21 11:27:53 crc kubenswrapper[4824]: I0121 11:27:53.487598 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4229a37f-5aa8-46c7-8ec4-74dde8937047-config-data\") pod \"ceilometer-0\" (UID: \"4229a37f-5aa8-46c7-8ec4-74dde8937047\") " pod="openstack/ceilometer-0" Jan 21 11:27:53 crc kubenswrapper[4824]: I0121 11:27:53.487667 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4229a37f-5aa8-46c7-8ec4-74dde8937047-log-httpd\") pod \"ceilometer-0\" (UID: \"4229a37f-5aa8-46c7-8ec4-74dde8937047\") " pod="openstack/ceilometer-0" Jan 21 11:27:53 crc kubenswrapper[4824]: I0121 11:27:53.487726 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4229a37f-5aa8-46c7-8ec4-74dde8937047-run-httpd\") pod \"ceilometer-0\" (UID: \"4229a37f-5aa8-46c7-8ec4-74dde8937047\") " pod="openstack/ceilometer-0" Jan 21 11:27:53 crc kubenswrapper[4824]: I0121 11:27:53.487756 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4229a37f-5aa8-46c7-8ec4-74dde8937047-scripts\") pod \"ceilometer-0\" (UID: \"4229a37f-5aa8-46c7-8ec4-74dde8937047\") " pod="openstack/ceilometer-0" Jan 21 11:27:53 crc kubenswrapper[4824]: I0121 11:27:53.488499 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4229a37f-5aa8-46c7-8ec4-74dde8937047-log-httpd\") pod \"ceilometer-0\" (UID: \"4229a37f-5aa8-46c7-8ec4-74dde8937047\") " pod="openstack/ceilometer-0" Jan 21 11:27:53 crc kubenswrapper[4824]: I0121 11:27:53.491946 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4229a37f-5aa8-46c7-8ec4-74dde8937047-scripts\") pod \"ceilometer-0\" (UID: \"4229a37f-5aa8-46c7-8ec4-74dde8937047\") " pod="openstack/ceilometer-0" Jan 21 11:27:53 crc kubenswrapper[4824]: I0121 11:27:53.492476 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4229a37f-5aa8-46c7-8ec4-74dde8937047-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"4229a37f-5aa8-46c7-8ec4-74dde8937047\") " pod="openstack/ceilometer-0" Jan 21 11:27:53 crc kubenswrapper[4824]: I0121 11:27:53.492923 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4229a37f-5aa8-46c7-8ec4-74dde8937047-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"4229a37f-5aa8-46c7-8ec4-74dde8937047\") " pod="openstack/ceilometer-0" Jan 21 11:27:53 crc kubenswrapper[4824]: I0121 11:27:53.502827 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4229a37f-5aa8-46c7-8ec4-74dde8937047-config-data\") pod \"ceilometer-0\" (UID: \"4229a37f-5aa8-46c7-8ec4-74dde8937047\") " pod="openstack/ceilometer-0" Jan 21 11:27:53 crc kubenswrapper[4824]: I0121 11:27:53.505102 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6zwmx\" (UniqueName: \"kubernetes.io/projected/4229a37f-5aa8-46c7-8ec4-74dde8937047-kube-api-access-6zwmx\") pod \"ceilometer-0\" (UID: \"4229a37f-5aa8-46c7-8ec4-74dde8937047\") " pod="openstack/ceilometer-0" Jan 21 11:27:53 crc kubenswrapper[4824]: I0121 11:27:53.590833 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 21 11:27:53 crc kubenswrapper[4824]: I0121 11:27:53.702866 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Jan 21 11:27:53 crc kubenswrapper[4824]: W0121 11:27:53.710068 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5d989da6_218c_4d56_b121_ce2002e90616.slice/crio-41bf3501b387352adac25f51221ca7a56a480e13956b916db174150a5458d50b WatchSource:0}: Error finding container 41bf3501b387352adac25f51221ca7a56a480e13956b916db174150a5458d50b: Status 404 returned error can't find the container with id 41bf3501b387352adac25f51221ca7a56a480e13956b916db174150a5458d50b Jan 21 11:27:53 crc kubenswrapper[4824]: I0121 11:27:53.932087 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"5d989da6-218c-4d56-b121-ce2002e90616","Type":"ContainerStarted","Data":"30df82afcb16527a31ea3e3ed18927eb1b094d788d010b009a7cc8f18ac7450c"} Jan 21 11:27:53 crc kubenswrapper[4824]: I0121 11:27:53.932301 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"5d989da6-218c-4d56-b121-ce2002e90616","Type":"ContainerStarted","Data":"41bf3501b387352adac25f51221ca7a56a480e13956b916db174150a5458d50b"} Jan 21 11:27:53 crc kubenswrapper[4824]: I0121 11:27:53.933138 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Jan 21 11:27:53 crc kubenswrapper[4824]: I0121 11:27:53.947550 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=1.947537316 podStartE2EDuration="1.947537316s" podCreationTimestamp="2026-01-21 11:27:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:27:53.946057164 +0000 UTC m=+1016.239086455" watchObservedRunningTime="2026-01-21 11:27:53.947537316 +0000 UTC m=+1016.240566607" Jan 21 11:27:53 crc kubenswrapper[4824]: I0121 11:27:53.981137 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Jan 21 11:27:54 crc kubenswrapper[4824]: I0121 11:27:54.057143 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="305dccfd-067c-4062-b9d1-0200e105e8fc" path="/var/lib/kubelet/pods/305dccfd-067c-4062-b9d1-0200e105e8fc/volumes" Jan 21 11:27:54 crc kubenswrapper[4824]: I0121 11:27:54.057907 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="964436a3-7963-4a72-bcb6-b7bab7e96574" path="/var/lib/kubelet/pods/964436a3-7963-4a72-bcb6-b7bab7e96574/volumes" Jan 21 11:27:54 crc kubenswrapper[4824]: I0121 11:27:54.058619 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f5046792-741c-487c-adb8-43502e9fdba1" path="/var/lib/kubelet/pods/f5046792-741c-487c-adb8-43502e9fdba1/volumes" Jan 21 11:27:54 crc kubenswrapper[4824]: I0121 11:27:54.939852 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4229a37f-5aa8-46c7-8ec4-74dde8937047","Type":"ContainerStarted","Data":"dd565547c1040520ce1112c6d6a660fde99c6c04a0693bd3a1b6b9ee3760d887"} Jan 21 11:27:54 crc kubenswrapper[4824]: I0121 11:27:54.940083 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4229a37f-5aa8-46c7-8ec4-74dde8937047","Type":"ContainerStarted","Data":"092b4d89a4764210f21850893c608f45eeb6897569b66ec19869b2ec6f56a849"} Jan 21 11:27:54 crc kubenswrapper[4824]: I0121 11:27:54.982626 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Jan 21 11:27:55 crc kubenswrapper[4824]: I0121 11:27:55.947253 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4229a37f-5aa8-46c7-8ec4-74dde8937047","Type":"ContainerStarted","Data":"9e5a9a376f5ca8e9a715e7d63a4d682d3e22d112810ff9c8bcac3ca694b1c63f"} Jan 21 11:27:56 crc kubenswrapper[4824]: I0121 11:27:56.012784 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Jan 21 11:27:56 crc kubenswrapper[4824]: I0121 11:27:56.955338 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4229a37f-5aa8-46c7-8ec4-74dde8937047","Type":"ContainerStarted","Data":"c385e10009efadf59111b9ba2f835f53a57502df4ba216b1657e806da68fc0c6"} Jan 21 11:27:56 crc kubenswrapper[4824]: I0121 11:27:56.955434 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell0-conductor-0" podUID="5d989da6-218c-4d56-b121-ce2002e90616" containerName="nova-cell0-conductor-conductor" containerID="cri-o://30df82afcb16527a31ea3e3ed18927eb1b094d788d010b009a7cc8f18ac7450c" gracePeriod=30 Jan 21 11:27:57 crc kubenswrapper[4824]: I0121 11:27:57.646596 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Jan 21 11:27:57 crc kubenswrapper[4824]: I0121 11:27:57.753457 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ddqrh\" (UniqueName: \"kubernetes.io/projected/5d989da6-218c-4d56-b121-ce2002e90616-kube-api-access-ddqrh\") pod \"5d989da6-218c-4d56-b121-ce2002e90616\" (UID: \"5d989da6-218c-4d56-b121-ce2002e90616\") " Jan 21 11:27:57 crc kubenswrapper[4824]: I0121 11:27:57.753718 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d989da6-218c-4d56-b121-ce2002e90616-combined-ca-bundle\") pod \"5d989da6-218c-4d56-b121-ce2002e90616\" (UID: \"5d989da6-218c-4d56-b121-ce2002e90616\") " Jan 21 11:27:57 crc kubenswrapper[4824]: I0121 11:27:57.753827 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5d989da6-218c-4d56-b121-ce2002e90616-config-data\") pod \"5d989da6-218c-4d56-b121-ce2002e90616\" (UID: \"5d989da6-218c-4d56-b121-ce2002e90616\") " Jan 21 11:27:57 crc kubenswrapper[4824]: I0121 11:27:57.757218 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5d989da6-218c-4d56-b121-ce2002e90616-kube-api-access-ddqrh" (OuterVolumeSpecName: "kube-api-access-ddqrh") pod "5d989da6-218c-4d56-b121-ce2002e90616" (UID: "5d989da6-218c-4d56-b121-ce2002e90616"). InnerVolumeSpecName "kube-api-access-ddqrh". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:27:57 crc kubenswrapper[4824]: I0121 11:27:57.773081 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d989da6-218c-4d56-b121-ce2002e90616-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5d989da6-218c-4d56-b121-ce2002e90616" (UID: "5d989da6-218c-4d56-b121-ce2002e90616"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:27:57 crc kubenswrapper[4824]: I0121 11:27:57.774316 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d989da6-218c-4d56-b121-ce2002e90616-config-data" (OuterVolumeSpecName: "config-data") pod "5d989da6-218c-4d56-b121-ce2002e90616" (UID: "5d989da6-218c-4d56-b121-ce2002e90616"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:27:57 crc kubenswrapper[4824]: I0121 11:27:57.856086 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ddqrh\" (UniqueName: \"kubernetes.io/projected/5d989da6-218c-4d56-b121-ce2002e90616-kube-api-access-ddqrh\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:57 crc kubenswrapper[4824]: I0121 11:27:57.856111 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d989da6-218c-4d56-b121-ce2002e90616-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:57 crc kubenswrapper[4824]: I0121 11:27:57.856120 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5d989da6-218c-4d56-b121-ce2002e90616-config-data\") on node \"crc\" DevicePath \"\"" Jan 21 11:27:57 crc kubenswrapper[4824]: I0121 11:27:57.964372 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4229a37f-5aa8-46c7-8ec4-74dde8937047","Type":"ContainerStarted","Data":"a7189e58853aa7540d18485396cf7983417d9550e4fd62b2794a48453697a520"} Jan 21 11:27:57 crc kubenswrapper[4824]: I0121 11:27:57.964452 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4229a37f-5aa8-46c7-8ec4-74dde8937047" containerName="ceilometer-central-agent" containerID="cri-o://dd565547c1040520ce1112c6d6a660fde99c6c04a0693bd3a1b6b9ee3760d887" gracePeriod=30 Jan 21 11:27:57 crc kubenswrapper[4824]: I0121 11:27:57.964513 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Jan 21 11:27:57 crc kubenswrapper[4824]: I0121 11:27:57.964536 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4229a37f-5aa8-46c7-8ec4-74dde8937047" containerName="proxy-httpd" containerID="cri-o://a7189e58853aa7540d18485396cf7983417d9550e4fd62b2794a48453697a520" gracePeriod=30 Jan 21 11:27:57 crc kubenswrapper[4824]: I0121 11:27:57.964572 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4229a37f-5aa8-46c7-8ec4-74dde8937047" containerName="sg-core" containerID="cri-o://c385e10009efadf59111b9ba2f835f53a57502df4ba216b1657e806da68fc0c6" gracePeriod=30 Jan 21 11:27:57 crc kubenswrapper[4824]: I0121 11:27:57.964606 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4229a37f-5aa8-46c7-8ec4-74dde8937047" containerName="ceilometer-notification-agent" containerID="cri-o://9e5a9a376f5ca8e9a715e7d63a4d682d3e22d112810ff9c8bcac3ca694b1c63f" gracePeriod=30 Jan 21 11:27:57 crc kubenswrapper[4824]: I0121 11:27:57.971074 4824 generic.go:334] "Generic (PLEG): container finished" podID="5d989da6-218c-4d56-b121-ce2002e90616" containerID="30df82afcb16527a31ea3e3ed18927eb1b094d788d010b009a7cc8f18ac7450c" exitCode=0 Jan 21 11:27:57 crc kubenswrapper[4824]: I0121 11:27:57.971103 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"5d989da6-218c-4d56-b121-ce2002e90616","Type":"ContainerDied","Data":"30df82afcb16527a31ea3e3ed18927eb1b094d788d010b009a7cc8f18ac7450c"} Jan 21 11:27:57 crc kubenswrapper[4824]: I0121 11:27:57.971128 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"5d989da6-218c-4d56-b121-ce2002e90616","Type":"ContainerDied","Data":"41bf3501b387352adac25f51221ca7a56a480e13956b916db174150a5458d50b"} Jan 21 11:27:57 crc kubenswrapper[4824]: I0121 11:27:57.971140 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Jan 21 11:27:57 crc kubenswrapper[4824]: I0121 11:27:57.971153 4824 scope.go:117] "RemoveContainer" containerID="30df82afcb16527a31ea3e3ed18927eb1b094d788d010b009a7cc8f18ac7450c" Jan 21 11:27:57 crc kubenswrapper[4824]: I0121 11:27:57.989736 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.524160698 podStartE2EDuration="4.989723109s" podCreationTimestamp="2026-01-21 11:27:53 +0000 UTC" firstStartedPulling="2026-01-21 11:27:53.983759258 +0000 UTC m=+1016.276788551" lastFinishedPulling="2026-01-21 11:27:57.44932168 +0000 UTC m=+1019.742350962" observedRunningTime="2026-01-21 11:27:57.986120375 +0000 UTC m=+1020.279149667" watchObservedRunningTime="2026-01-21 11:27:57.989723109 +0000 UTC m=+1020.282752402" Jan 21 11:27:58 crc kubenswrapper[4824]: I0121 11:27:58.000094 4824 scope.go:117] "RemoveContainer" containerID="30df82afcb16527a31ea3e3ed18927eb1b094d788d010b009a7cc8f18ac7450c" Jan 21 11:27:58 crc kubenswrapper[4824]: E0121 11:27:58.000810 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"30df82afcb16527a31ea3e3ed18927eb1b094d788d010b009a7cc8f18ac7450c\": container with ID starting with 30df82afcb16527a31ea3e3ed18927eb1b094d788d010b009a7cc8f18ac7450c not found: ID does not exist" containerID="30df82afcb16527a31ea3e3ed18927eb1b094d788d010b009a7cc8f18ac7450c" Jan 21 11:27:58 crc kubenswrapper[4824]: I0121 11:27:58.000849 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"30df82afcb16527a31ea3e3ed18927eb1b094d788d010b009a7cc8f18ac7450c"} err="failed to get container status \"30df82afcb16527a31ea3e3ed18927eb1b094d788d010b009a7cc8f18ac7450c\": rpc error: code = NotFound desc = could not find container \"30df82afcb16527a31ea3e3ed18927eb1b094d788d010b009a7cc8f18ac7450c\": container with ID starting with 30df82afcb16527a31ea3e3ed18927eb1b094d788d010b009a7cc8f18ac7450c not found: ID does not exist" Jan 21 11:27:58 crc kubenswrapper[4824]: I0121 11:27:58.007163 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Jan 21 11:27:58 crc kubenswrapper[4824]: I0121 11:27:58.021547 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-0"] Jan 21 11:27:58 crc kubenswrapper[4824]: I0121 11:27:58.047591 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Jan 21 11:27:58 crc kubenswrapper[4824]: E0121 11:27:58.050677 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5d989da6-218c-4d56-b121-ce2002e90616" containerName="nova-cell0-conductor-conductor" Jan 21 11:27:58 crc kubenswrapper[4824]: I0121 11:27:58.050706 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="5d989da6-218c-4d56-b121-ce2002e90616" containerName="nova-cell0-conductor-conductor" Jan 21 11:27:58 crc kubenswrapper[4824]: I0121 11:27:58.051326 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="5d989da6-218c-4d56-b121-ce2002e90616" containerName="nova-cell0-conductor-conductor" Jan 21 11:27:58 crc kubenswrapper[4824]: I0121 11:27:58.052377 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Jan 21 11:27:58 crc kubenswrapper[4824]: I0121 11:27:58.060663 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-gbh62" Jan 21 11:27:58 crc kubenswrapper[4824]: I0121 11:27:58.063318 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Jan 21 11:27:58 crc kubenswrapper[4824]: I0121 11:27:58.064186 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5d989da6-218c-4d56-b121-ce2002e90616" path="/var/lib/kubelet/pods/5d989da6-218c-4d56-b121-ce2002e90616/volumes" Jan 21 11:27:58 crc kubenswrapper[4824]: I0121 11:27:58.064641 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Jan 21 11:27:58 crc kubenswrapper[4824]: I0121 11:27:58.160567 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x989p\" (UniqueName: \"kubernetes.io/projected/8effdacf-8fec-40b4-8f61-8856ac0f232c-kube-api-access-x989p\") pod \"nova-cell0-conductor-0\" (UID: \"8effdacf-8fec-40b4-8f61-8856ac0f232c\") " pod="openstack/nova-cell0-conductor-0" Jan 21 11:27:58 crc kubenswrapper[4824]: I0121 11:27:58.160882 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8effdacf-8fec-40b4-8f61-8856ac0f232c-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"8effdacf-8fec-40b4-8f61-8856ac0f232c\") " pod="openstack/nova-cell0-conductor-0" Jan 21 11:27:58 crc kubenswrapper[4824]: I0121 11:27:58.161073 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8effdacf-8fec-40b4-8f61-8856ac0f232c-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"8effdacf-8fec-40b4-8f61-8856ac0f232c\") " pod="openstack/nova-cell0-conductor-0" Jan 21 11:27:58 crc kubenswrapper[4824]: I0121 11:27:58.262636 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8effdacf-8fec-40b4-8f61-8856ac0f232c-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"8effdacf-8fec-40b4-8f61-8856ac0f232c\") " pod="openstack/nova-cell0-conductor-0" Jan 21 11:27:58 crc kubenswrapper[4824]: I0121 11:27:58.262758 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8effdacf-8fec-40b4-8f61-8856ac0f232c-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"8effdacf-8fec-40b4-8f61-8856ac0f232c\") " pod="openstack/nova-cell0-conductor-0" Jan 21 11:27:58 crc kubenswrapper[4824]: I0121 11:27:58.262927 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x989p\" (UniqueName: \"kubernetes.io/projected/8effdacf-8fec-40b4-8f61-8856ac0f232c-kube-api-access-x989p\") pod \"nova-cell0-conductor-0\" (UID: \"8effdacf-8fec-40b4-8f61-8856ac0f232c\") " pod="openstack/nova-cell0-conductor-0" Jan 21 11:27:58 crc kubenswrapper[4824]: I0121 11:27:58.266374 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8effdacf-8fec-40b4-8f61-8856ac0f232c-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"8effdacf-8fec-40b4-8f61-8856ac0f232c\") " pod="openstack/nova-cell0-conductor-0" Jan 21 11:27:58 crc kubenswrapper[4824]: I0121 11:27:58.273813 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8effdacf-8fec-40b4-8f61-8856ac0f232c-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"8effdacf-8fec-40b4-8f61-8856ac0f232c\") " pod="openstack/nova-cell0-conductor-0" Jan 21 11:27:58 crc kubenswrapper[4824]: I0121 11:27:58.275765 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x989p\" (UniqueName: \"kubernetes.io/projected/8effdacf-8fec-40b4-8f61-8856ac0f232c-kube-api-access-x989p\") pod \"nova-cell0-conductor-0\" (UID: \"8effdacf-8fec-40b4-8f61-8856ac0f232c\") " pod="openstack/nova-cell0-conductor-0" Jan 21 11:27:58 crc kubenswrapper[4824]: I0121 11:27:58.374319 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Jan 21 11:27:58 crc kubenswrapper[4824]: I0121 11:27:58.748440 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Jan 21 11:27:58 crc kubenswrapper[4824]: W0121 11:27:58.754877 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8effdacf_8fec_40b4_8f61_8856ac0f232c.slice/crio-25998cf7bd255205c22a0312ba7cdf9b107c0a619888da32f9b0e42fd7815107 WatchSource:0}: Error finding container 25998cf7bd255205c22a0312ba7cdf9b107c0a619888da32f9b0e42fd7815107: Status 404 returned error can't find the container with id 25998cf7bd255205c22a0312ba7cdf9b107c0a619888da32f9b0e42fd7815107 Jan 21 11:27:58 crc kubenswrapper[4824]: I0121 11:27:58.980190 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"8effdacf-8fec-40b4-8f61-8856ac0f232c","Type":"ContainerStarted","Data":"cb4175471f7b293199c781fd7c6f75c7749c4e9253027bc4a10e5ee9ff6b8076"} Jan 21 11:27:58 crc kubenswrapper[4824]: I0121 11:27:58.980230 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"8effdacf-8fec-40b4-8f61-8856ac0f232c","Type":"ContainerStarted","Data":"25998cf7bd255205c22a0312ba7cdf9b107c0a619888da32f9b0e42fd7815107"} Jan 21 11:27:58 crc kubenswrapper[4824]: I0121 11:27:58.980415 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Jan 21 11:27:58 crc kubenswrapper[4824]: I0121 11:27:58.982690 4824 generic.go:334] "Generic (PLEG): container finished" podID="4229a37f-5aa8-46c7-8ec4-74dde8937047" containerID="a7189e58853aa7540d18485396cf7983417d9550e4fd62b2794a48453697a520" exitCode=0 Jan 21 11:27:58 crc kubenswrapper[4824]: I0121 11:27:58.982716 4824 generic.go:334] "Generic (PLEG): container finished" podID="4229a37f-5aa8-46c7-8ec4-74dde8937047" containerID="c385e10009efadf59111b9ba2f835f53a57502df4ba216b1657e806da68fc0c6" exitCode=2 Jan 21 11:27:58 crc kubenswrapper[4824]: I0121 11:27:58.982726 4824 generic.go:334] "Generic (PLEG): container finished" podID="4229a37f-5aa8-46c7-8ec4-74dde8937047" containerID="9e5a9a376f5ca8e9a715e7d63a4d682d3e22d112810ff9c8bcac3ca694b1c63f" exitCode=0 Jan 21 11:27:58 crc kubenswrapper[4824]: I0121 11:27:58.982753 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4229a37f-5aa8-46c7-8ec4-74dde8937047","Type":"ContainerDied","Data":"a7189e58853aa7540d18485396cf7983417d9550e4fd62b2794a48453697a520"} Jan 21 11:27:58 crc kubenswrapper[4824]: I0121 11:27:58.982784 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4229a37f-5aa8-46c7-8ec4-74dde8937047","Type":"ContainerDied","Data":"c385e10009efadf59111b9ba2f835f53a57502df4ba216b1657e806da68fc0c6"} Jan 21 11:27:58 crc kubenswrapper[4824]: I0121 11:27:58.982795 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4229a37f-5aa8-46c7-8ec4-74dde8937047","Type":"ContainerDied","Data":"9e5a9a376f5ca8e9a715e7d63a4d682d3e22d112810ff9c8bcac3ca694b1c63f"} Jan 21 11:27:58 crc kubenswrapper[4824]: I0121 11:27:58.998545 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=0.998532424 podStartE2EDuration="998.532424ms" podCreationTimestamp="2026-01-21 11:27:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:27:58.99225283 +0000 UTC m=+1021.285282122" watchObservedRunningTime="2026-01-21 11:27:58.998532424 +0000 UTC m=+1021.291561717" Jan 21 11:27:59 crc kubenswrapper[4824]: I0121 11:27:59.718680 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-engine-5bc6fc99dd-nv45v" Jan 21 11:27:59 crc kubenswrapper[4824]: I0121 11:27:59.756717 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-engine-656b845fb4-qst2s"] Jan 21 11:27:59 crc kubenswrapper[4824]: I0121 11:27:59.756907 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/heat-engine-656b845fb4-qst2s" podUID="d8d66537-20e1-4f87-9077-484e264bae81" containerName="heat-engine" containerID="cri-o://05efb240856120694f40b5baf0ccbaf59da69510aa66da3d1ac87094a30bb9ea" gracePeriod=60 Jan 21 11:28:01 crc kubenswrapper[4824]: I0121 11:28:01.720754 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 21 11:28:01 crc kubenswrapper[4824]: I0121 11:28:01.823226 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4229a37f-5aa8-46c7-8ec4-74dde8937047-run-httpd\") pod \"4229a37f-5aa8-46c7-8ec4-74dde8937047\" (UID: \"4229a37f-5aa8-46c7-8ec4-74dde8937047\") " Jan 21 11:28:01 crc kubenswrapper[4824]: I0121 11:28:01.823275 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4229a37f-5aa8-46c7-8ec4-74dde8937047-sg-core-conf-yaml\") pod \"4229a37f-5aa8-46c7-8ec4-74dde8937047\" (UID: \"4229a37f-5aa8-46c7-8ec4-74dde8937047\") " Jan 21 11:28:01 crc kubenswrapper[4824]: I0121 11:28:01.823332 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4229a37f-5aa8-46c7-8ec4-74dde8937047-scripts\") pod \"4229a37f-5aa8-46c7-8ec4-74dde8937047\" (UID: \"4229a37f-5aa8-46c7-8ec4-74dde8937047\") " Jan 21 11:28:01 crc kubenswrapper[4824]: I0121 11:28:01.823386 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4229a37f-5aa8-46c7-8ec4-74dde8937047-config-data\") pod \"4229a37f-5aa8-46c7-8ec4-74dde8937047\" (UID: \"4229a37f-5aa8-46c7-8ec4-74dde8937047\") " Jan 21 11:28:01 crc kubenswrapper[4824]: I0121 11:28:01.823450 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4229a37f-5aa8-46c7-8ec4-74dde8937047-log-httpd\") pod \"4229a37f-5aa8-46c7-8ec4-74dde8937047\" (UID: \"4229a37f-5aa8-46c7-8ec4-74dde8937047\") " Jan 21 11:28:01 crc kubenswrapper[4824]: I0121 11:28:01.823478 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6zwmx\" (UniqueName: \"kubernetes.io/projected/4229a37f-5aa8-46c7-8ec4-74dde8937047-kube-api-access-6zwmx\") pod \"4229a37f-5aa8-46c7-8ec4-74dde8937047\" (UID: \"4229a37f-5aa8-46c7-8ec4-74dde8937047\") " Jan 21 11:28:01 crc kubenswrapper[4824]: I0121 11:28:01.823497 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4229a37f-5aa8-46c7-8ec4-74dde8937047-combined-ca-bundle\") pod \"4229a37f-5aa8-46c7-8ec4-74dde8937047\" (UID: \"4229a37f-5aa8-46c7-8ec4-74dde8937047\") " Jan 21 11:28:01 crc kubenswrapper[4824]: I0121 11:28:01.823685 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4229a37f-5aa8-46c7-8ec4-74dde8937047-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "4229a37f-5aa8-46c7-8ec4-74dde8937047" (UID: "4229a37f-5aa8-46c7-8ec4-74dde8937047"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:28:01 crc kubenswrapper[4824]: I0121 11:28:01.823835 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4229a37f-5aa8-46c7-8ec4-74dde8937047-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "4229a37f-5aa8-46c7-8ec4-74dde8937047" (UID: "4229a37f-5aa8-46c7-8ec4-74dde8937047"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:28:01 crc kubenswrapper[4824]: I0121 11:28:01.824165 4824 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4229a37f-5aa8-46c7-8ec4-74dde8937047-run-httpd\") on node \"crc\" DevicePath \"\"" Jan 21 11:28:01 crc kubenswrapper[4824]: I0121 11:28:01.824181 4824 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4229a37f-5aa8-46c7-8ec4-74dde8937047-log-httpd\") on node \"crc\" DevicePath \"\"" Jan 21 11:28:01 crc kubenswrapper[4824]: I0121 11:28:01.832884 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4229a37f-5aa8-46c7-8ec4-74dde8937047-kube-api-access-6zwmx" (OuterVolumeSpecName: "kube-api-access-6zwmx") pod "4229a37f-5aa8-46c7-8ec4-74dde8937047" (UID: "4229a37f-5aa8-46c7-8ec4-74dde8937047"). InnerVolumeSpecName "kube-api-access-6zwmx". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:28:01 crc kubenswrapper[4824]: I0121 11:28:01.838102 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4229a37f-5aa8-46c7-8ec4-74dde8937047-scripts" (OuterVolumeSpecName: "scripts") pod "4229a37f-5aa8-46c7-8ec4-74dde8937047" (UID: "4229a37f-5aa8-46c7-8ec4-74dde8937047"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:28:01 crc kubenswrapper[4824]: I0121 11:28:01.855102 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4229a37f-5aa8-46c7-8ec4-74dde8937047-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "4229a37f-5aa8-46c7-8ec4-74dde8937047" (UID: "4229a37f-5aa8-46c7-8ec4-74dde8937047"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:28:01 crc kubenswrapper[4824]: I0121 11:28:01.891639 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4229a37f-5aa8-46c7-8ec4-74dde8937047-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4229a37f-5aa8-46c7-8ec4-74dde8937047" (UID: "4229a37f-5aa8-46c7-8ec4-74dde8937047"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:28:01 crc kubenswrapper[4824]: I0121 11:28:01.905301 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4229a37f-5aa8-46c7-8ec4-74dde8937047-config-data" (OuterVolumeSpecName: "config-data") pod "4229a37f-5aa8-46c7-8ec4-74dde8937047" (UID: "4229a37f-5aa8-46c7-8ec4-74dde8937047"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:28:01 crc kubenswrapper[4824]: I0121 11:28:01.925987 4824 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4229a37f-5aa8-46c7-8ec4-74dde8937047-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Jan 21 11:28:01 crc kubenswrapper[4824]: I0121 11:28:01.926011 4824 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4229a37f-5aa8-46c7-8ec4-74dde8937047-scripts\") on node \"crc\" DevicePath \"\"" Jan 21 11:28:01 crc kubenswrapper[4824]: I0121 11:28:01.926021 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4229a37f-5aa8-46c7-8ec4-74dde8937047-config-data\") on node \"crc\" DevicePath \"\"" Jan 21 11:28:01 crc kubenswrapper[4824]: I0121 11:28:01.926029 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6zwmx\" (UniqueName: \"kubernetes.io/projected/4229a37f-5aa8-46c7-8ec4-74dde8937047-kube-api-access-6zwmx\") on node \"crc\" DevicePath \"\"" Jan 21 11:28:01 crc kubenswrapper[4824]: I0121 11:28:01.926040 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4229a37f-5aa8-46c7-8ec4-74dde8937047-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 21 11:28:02 crc kubenswrapper[4824]: I0121 11:28:02.004632 4824 generic.go:334] "Generic (PLEG): container finished" podID="4229a37f-5aa8-46c7-8ec4-74dde8937047" containerID="dd565547c1040520ce1112c6d6a660fde99c6c04a0693bd3a1b6b9ee3760d887" exitCode=0 Jan 21 11:28:02 crc kubenswrapper[4824]: I0121 11:28:02.004669 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4229a37f-5aa8-46c7-8ec4-74dde8937047","Type":"ContainerDied","Data":"dd565547c1040520ce1112c6d6a660fde99c6c04a0693bd3a1b6b9ee3760d887"} Jan 21 11:28:02 crc kubenswrapper[4824]: I0121 11:28:02.004693 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4229a37f-5aa8-46c7-8ec4-74dde8937047","Type":"ContainerDied","Data":"092b4d89a4764210f21850893c608f45eeb6897569b66ec19869b2ec6f56a849"} Jan 21 11:28:02 crc kubenswrapper[4824]: I0121 11:28:02.004707 4824 scope.go:117] "RemoveContainer" containerID="a7189e58853aa7540d18485396cf7983417d9550e4fd62b2794a48453697a520" Jan 21 11:28:02 crc kubenswrapper[4824]: I0121 11:28:02.004818 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 21 11:28:02 crc kubenswrapper[4824]: I0121 11:28:02.034678 4824 scope.go:117] "RemoveContainer" containerID="c385e10009efadf59111b9ba2f835f53a57502df4ba216b1657e806da68fc0c6" Jan 21 11:28:02 crc kubenswrapper[4824]: I0121 11:28:02.041842 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Jan 21 11:28:02 crc kubenswrapper[4824]: I0121 11:28:02.052704 4824 scope.go:117] "RemoveContainer" containerID="9e5a9a376f5ca8e9a715e7d63a4d682d3e22d112810ff9c8bcac3ca694b1c63f" Jan 21 11:28:02 crc kubenswrapper[4824]: I0121 11:28:02.065678 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Jan 21 11:28:02 crc kubenswrapper[4824]: I0121 11:28:02.066739 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Jan 21 11:28:02 crc kubenswrapper[4824]: E0121 11:28:02.067062 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4229a37f-5aa8-46c7-8ec4-74dde8937047" containerName="ceilometer-notification-agent" Jan 21 11:28:02 crc kubenswrapper[4824]: I0121 11:28:02.067080 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="4229a37f-5aa8-46c7-8ec4-74dde8937047" containerName="ceilometer-notification-agent" Jan 21 11:28:02 crc kubenswrapper[4824]: E0121 11:28:02.067094 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4229a37f-5aa8-46c7-8ec4-74dde8937047" containerName="ceilometer-central-agent" Jan 21 11:28:02 crc kubenswrapper[4824]: I0121 11:28:02.067100 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="4229a37f-5aa8-46c7-8ec4-74dde8937047" containerName="ceilometer-central-agent" Jan 21 11:28:02 crc kubenswrapper[4824]: E0121 11:28:02.067108 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4229a37f-5aa8-46c7-8ec4-74dde8937047" containerName="sg-core" Jan 21 11:28:02 crc kubenswrapper[4824]: I0121 11:28:02.067113 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="4229a37f-5aa8-46c7-8ec4-74dde8937047" containerName="sg-core" Jan 21 11:28:02 crc kubenswrapper[4824]: E0121 11:28:02.067132 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4229a37f-5aa8-46c7-8ec4-74dde8937047" containerName="proxy-httpd" Jan 21 11:28:02 crc kubenswrapper[4824]: I0121 11:28:02.067138 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="4229a37f-5aa8-46c7-8ec4-74dde8937047" containerName="proxy-httpd" Jan 21 11:28:02 crc kubenswrapper[4824]: I0121 11:28:02.067309 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="4229a37f-5aa8-46c7-8ec4-74dde8937047" containerName="sg-core" Jan 21 11:28:02 crc kubenswrapper[4824]: I0121 11:28:02.067319 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="4229a37f-5aa8-46c7-8ec4-74dde8937047" containerName="ceilometer-central-agent" Jan 21 11:28:02 crc kubenswrapper[4824]: I0121 11:28:02.067332 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="4229a37f-5aa8-46c7-8ec4-74dde8937047" containerName="ceilometer-notification-agent" Jan 21 11:28:02 crc kubenswrapper[4824]: I0121 11:28:02.067347 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="4229a37f-5aa8-46c7-8ec4-74dde8937047" containerName="proxy-httpd" Jan 21 11:28:02 crc kubenswrapper[4824]: I0121 11:28:02.068715 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 21 11:28:02 crc kubenswrapper[4824]: I0121 11:28:02.073260 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Jan 21 11:28:02 crc kubenswrapper[4824]: I0121 11:28:02.073503 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Jan 21 11:28:02 crc kubenswrapper[4824]: I0121 11:28:02.074204 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Jan 21 11:28:02 crc kubenswrapper[4824]: I0121 11:28:02.081030 4824 scope.go:117] "RemoveContainer" containerID="dd565547c1040520ce1112c6d6a660fde99c6c04a0693bd3a1b6b9ee3760d887" Jan 21 11:28:02 crc kubenswrapper[4824]: I0121 11:28:02.101460 4824 scope.go:117] "RemoveContainer" containerID="a7189e58853aa7540d18485396cf7983417d9550e4fd62b2794a48453697a520" Jan 21 11:28:02 crc kubenswrapper[4824]: E0121 11:28:02.101748 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a7189e58853aa7540d18485396cf7983417d9550e4fd62b2794a48453697a520\": container with ID starting with a7189e58853aa7540d18485396cf7983417d9550e4fd62b2794a48453697a520 not found: ID does not exist" containerID="a7189e58853aa7540d18485396cf7983417d9550e4fd62b2794a48453697a520" Jan 21 11:28:02 crc kubenswrapper[4824]: I0121 11:28:02.101812 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a7189e58853aa7540d18485396cf7983417d9550e4fd62b2794a48453697a520"} err="failed to get container status \"a7189e58853aa7540d18485396cf7983417d9550e4fd62b2794a48453697a520\": rpc error: code = NotFound desc = could not find container \"a7189e58853aa7540d18485396cf7983417d9550e4fd62b2794a48453697a520\": container with ID starting with a7189e58853aa7540d18485396cf7983417d9550e4fd62b2794a48453697a520 not found: ID does not exist" Jan 21 11:28:02 crc kubenswrapper[4824]: I0121 11:28:02.101832 4824 scope.go:117] "RemoveContainer" containerID="c385e10009efadf59111b9ba2f835f53a57502df4ba216b1657e806da68fc0c6" Jan 21 11:28:02 crc kubenswrapper[4824]: E0121 11:28:02.102066 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c385e10009efadf59111b9ba2f835f53a57502df4ba216b1657e806da68fc0c6\": container with ID starting with c385e10009efadf59111b9ba2f835f53a57502df4ba216b1657e806da68fc0c6 not found: ID does not exist" containerID="c385e10009efadf59111b9ba2f835f53a57502df4ba216b1657e806da68fc0c6" Jan 21 11:28:02 crc kubenswrapper[4824]: I0121 11:28:02.102091 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c385e10009efadf59111b9ba2f835f53a57502df4ba216b1657e806da68fc0c6"} err="failed to get container status \"c385e10009efadf59111b9ba2f835f53a57502df4ba216b1657e806da68fc0c6\": rpc error: code = NotFound desc = could not find container \"c385e10009efadf59111b9ba2f835f53a57502df4ba216b1657e806da68fc0c6\": container with ID starting with c385e10009efadf59111b9ba2f835f53a57502df4ba216b1657e806da68fc0c6 not found: ID does not exist" Jan 21 11:28:02 crc kubenswrapper[4824]: I0121 11:28:02.102105 4824 scope.go:117] "RemoveContainer" containerID="9e5a9a376f5ca8e9a715e7d63a4d682d3e22d112810ff9c8bcac3ca694b1c63f" Jan 21 11:28:02 crc kubenswrapper[4824]: E0121 11:28:02.102319 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9e5a9a376f5ca8e9a715e7d63a4d682d3e22d112810ff9c8bcac3ca694b1c63f\": container with ID starting with 9e5a9a376f5ca8e9a715e7d63a4d682d3e22d112810ff9c8bcac3ca694b1c63f not found: ID does not exist" containerID="9e5a9a376f5ca8e9a715e7d63a4d682d3e22d112810ff9c8bcac3ca694b1c63f" Jan 21 11:28:02 crc kubenswrapper[4824]: I0121 11:28:02.102339 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9e5a9a376f5ca8e9a715e7d63a4d682d3e22d112810ff9c8bcac3ca694b1c63f"} err="failed to get container status \"9e5a9a376f5ca8e9a715e7d63a4d682d3e22d112810ff9c8bcac3ca694b1c63f\": rpc error: code = NotFound desc = could not find container \"9e5a9a376f5ca8e9a715e7d63a4d682d3e22d112810ff9c8bcac3ca694b1c63f\": container with ID starting with 9e5a9a376f5ca8e9a715e7d63a4d682d3e22d112810ff9c8bcac3ca694b1c63f not found: ID does not exist" Jan 21 11:28:02 crc kubenswrapper[4824]: I0121 11:28:02.102353 4824 scope.go:117] "RemoveContainer" containerID="dd565547c1040520ce1112c6d6a660fde99c6c04a0693bd3a1b6b9ee3760d887" Jan 21 11:28:02 crc kubenswrapper[4824]: E0121 11:28:02.102595 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dd565547c1040520ce1112c6d6a660fde99c6c04a0693bd3a1b6b9ee3760d887\": container with ID starting with dd565547c1040520ce1112c6d6a660fde99c6c04a0693bd3a1b6b9ee3760d887 not found: ID does not exist" containerID="dd565547c1040520ce1112c6d6a660fde99c6c04a0693bd3a1b6b9ee3760d887" Jan 21 11:28:02 crc kubenswrapper[4824]: I0121 11:28:02.102613 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dd565547c1040520ce1112c6d6a660fde99c6c04a0693bd3a1b6b9ee3760d887"} err="failed to get container status \"dd565547c1040520ce1112c6d6a660fde99c6c04a0693bd3a1b6b9ee3760d887\": rpc error: code = NotFound desc = could not find container \"dd565547c1040520ce1112c6d6a660fde99c6c04a0693bd3a1b6b9ee3760d887\": container with ID starting with dd565547c1040520ce1112c6d6a660fde99c6c04a0693bd3a1b6b9ee3760d887 not found: ID does not exist" Jan 21 11:28:02 crc kubenswrapper[4824]: I0121 11:28:02.230463 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/deb46ee7-a757-433b-8595-ac21f9d3a03f-scripts\") pod \"ceilometer-0\" (UID: \"deb46ee7-a757-433b-8595-ac21f9d3a03f\") " pod="openstack/ceilometer-0" Jan 21 11:28:02 crc kubenswrapper[4824]: I0121 11:28:02.230813 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/deb46ee7-a757-433b-8595-ac21f9d3a03f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"deb46ee7-a757-433b-8595-ac21f9d3a03f\") " pod="openstack/ceilometer-0" Jan 21 11:28:02 crc kubenswrapper[4824]: I0121 11:28:02.230874 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/deb46ee7-a757-433b-8595-ac21f9d3a03f-config-data\") pod \"ceilometer-0\" (UID: \"deb46ee7-a757-433b-8595-ac21f9d3a03f\") " pod="openstack/ceilometer-0" Jan 21 11:28:02 crc kubenswrapper[4824]: I0121 11:28:02.230921 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/deb46ee7-a757-433b-8595-ac21f9d3a03f-log-httpd\") pod \"ceilometer-0\" (UID: \"deb46ee7-a757-433b-8595-ac21f9d3a03f\") " pod="openstack/ceilometer-0" Jan 21 11:28:02 crc kubenswrapper[4824]: I0121 11:28:02.231054 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/deb46ee7-a757-433b-8595-ac21f9d3a03f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"deb46ee7-a757-433b-8595-ac21f9d3a03f\") " pod="openstack/ceilometer-0" Jan 21 11:28:02 crc kubenswrapper[4824]: I0121 11:28:02.231092 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d25xm\" (UniqueName: \"kubernetes.io/projected/deb46ee7-a757-433b-8595-ac21f9d3a03f-kube-api-access-d25xm\") pod \"ceilometer-0\" (UID: \"deb46ee7-a757-433b-8595-ac21f9d3a03f\") " pod="openstack/ceilometer-0" Jan 21 11:28:02 crc kubenswrapper[4824]: I0121 11:28:02.231207 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/deb46ee7-a757-433b-8595-ac21f9d3a03f-run-httpd\") pod \"ceilometer-0\" (UID: \"deb46ee7-a757-433b-8595-ac21f9d3a03f\") " pod="openstack/ceilometer-0" Jan 21 11:28:02 crc kubenswrapper[4824]: I0121 11:28:02.332984 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/deb46ee7-a757-433b-8595-ac21f9d3a03f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"deb46ee7-a757-433b-8595-ac21f9d3a03f\") " pod="openstack/ceilometer-0" Jan 21 11:28:02 crc kubenswrapper[4824]: I0121 11:28:02.333184 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/deb46ee7-a757-433b-8595-ac21f9d3a03f-config-data\") pod \"ceilometer-0\" (UID: \"deb46ee7-a757-433b-8595-ac21f9d3a03f\") " pod="openstack/ceilometer-0" Jan 21 11:28:02 crc kubenswrapper[4824]: I0121 11:28:02.333293 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/deb46ee7-a757-433b-8595-ac21f9d3a03f-log-httpd\") pod \"ceilometer-0\" (UID: \"deb46ee7-a757-433b-8595-ac21f9d3a03f\") " pod="openstack/ceilometer-0" Jan 21 11:28:02 crc kubenswrapper[4824]: I0121 11:28:02.333389 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/deb46ee7-a757-433b-8595-ac21f9d3a03f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"deb46ee7-a757-433b-8595-ac21f9d3a03f\") " pod="openstack/ceilometer-0" Jan 21 11:28:02 crc kubenswrapper[4824]: I0121 11:28:02.333466 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d25xm\" (UniqueName: \"kubernetes.io/projected/deb46ee7-a757-433b-8595-ac21f9d3a03f-kube-api-access-d25xm\") pod \"ceilometer-0\" (UID: \"deb46ee7-a757-433b-8595-ac21f9d3a03f\") " pod="openstack/ceilometer-0" Jan 21 11:28:02 crc kubenswrapper[4824]: I0121 11:28:02.333577 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/deb46ee7-a757-433b-8595-ac21f9d3a03f-run-httpd\") pod \"ceilometer-0\" (UID: \"deb46ee7-a757-433b-8595-ac21f9d3a03f\") " pod="openstack/ceilometer-0" Jan 21 11:28:02 crc kubenswrapper[4824]: I0121 11:28:02.333669 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/deb46ee7-a757-433b-8595-ac21f9d3a03f-scripts\") pod \"ceilometer-0\" (UID: \"deb46ee7-a757-433b-8595-ac21f9d3a03f\") " pod="openstack/ceilometer-0" Jan 21 11:28:02 crc kubenswrapper[4824]: I0121 11:28:02.334128 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/deb46ee7-a757-433b-8595-ac21f9d3a03f-log-httpd\") pod \"ceilometer-0\" (UID: \"deb46ee7-a757-433b-8595-ac21f9d3a03f\") " pod="openstack/ceilometer-0" Jan 21 11:28:02 crc kubenswrapper[4824]: I0121 11:28:02.334163 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/deb46ee7-a757-433b-8595-ac21f9d3a03f-run-httpd\") pod \"ceilometer-0\" (UID: \"deb46ee7-a757-433b-8595-ac21f9d3a03f\") " pod="openstack/ceilometer-0" Jan 21 11:28:02 crc kubenswrapper[4824]: I0121 11:28:02.336591 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/deb46ee7-a757-433b-8595-ac21f9d3a03f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"deb46ee7-a757-433b-8595-ac21f9d3a03f\") " pod="openstack/ceilometer-0" Jan 21 11:28:02 crc kubenswrapper[4824]: I0121 11:28:02.336635 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/deb46ee7-a757-433b-8595-ac21f9d3a03f-scripts\") pod \"ceilometer-0\" (UID: \"deb46ee7-a757-433b-8595-ac21f9d3a03f\") " pod="openstack/ceilometer-0" Jan 21 11:28:02 crc kubenswrapper[4824]: I0121 11:28:02.337081 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/deb46ee7-a757-433b-8595-ac21f9d3a03f-config-data\") pod \"ceilometer-0\" (UID: \"deb46ee7-a757-433b-8595-ac21f9d3a03f\") " pod="openstack/ceilometer-0" Jan 21 11:28:02 crc kubenswrapper[4824]: I0121 11:28:02.337185 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/deb46ee7-a757-433b-8595-ac21f9d3a03f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"deb46ee7-a757-433b-8595-ac21f9d3a03f\") " pod="openstack/ceilometer-0" Jan 21 11:28:02 crc kubenswrapper[4824]: I0121 11:28:02.347383 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d25xm\" (UniqueName: \"kubernetes.io/projected/deb46ee7-a757-433b-8595-ac21f9d3a03f-kube-api-access-d25xm\") pod \"ceilometer-0\" (UID: \"deb46ee7-a757-433b-8595-ac21f9d3a03f\") " pod="openstack/ceilometer-0" Jan 21 11:28:02 crc kubenswrapper[4824]: I0121 11:28:02.395569 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 21 11:28:02 crc kubenswrapper[4824]: E0121 11:28:02.687293 4824 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="05efb240856120694f40b5baf0ccbaf59da69510aa66da3d1ac87094a30bb9ea" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Jan 21 11:28:02 crc kubenswrapper[4824]: E0121 11:28:02.688419 4824 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="05efb240856120694f40b5baf0ccbaf59da69510aa66da3d1ac87094a30bb9ea" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Jan 21 11:28:02 crc kubenswrapper[4824]: E0121 11:28:02.689717 4824 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="05efb240856120694f40b5baf0ccbaf59da69510aa66da3d1ac87094a30bb9ea" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Jan 21 11:28:02 crc kubenswrapper[4824]: E0121 11:28:02.689750 4824 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/heat-engine-656b845fb4-qst2s" podUID="d8d66537-20e1-4f87-9077-484e264bae81" containerName="heat-engine" Jan 21 11:28:02 crc kubenswrapper[4824]: I0121 11:28:02.787486 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Jan 21 11:28:03 crc kubenswrapper[4824]: I0121 11:28:03.014702 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"deb46ee7-a757-433b-8595-ac21f9d3a03f","Type":"ContainerStarted","Data":"bac5f30d5f4848de2bccab1913e20b18c8655df8b33e352d91fc9c00699884ca"} Jan 21 11:28:03 crc kubenswrapper[4824]: I0121 11:28:03.401835 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Jan 21 11:28:03 crc kubenswrapper[4824]: I0121 11:28:03.795841 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Jan 21 11:28:03 crc kubenswrapper[4824]: I0121 11:28:03.967995 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-g9vq5"] Jan 21 11:28:03 crc kubenswrapper[4824]: I0121 11:28:03.969615 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-g9vq5" Jan 21 11:28:03 crc kubenswrapper[4824]: I0121 11:28:03.977550 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Jan 21 11:28:03 crc kubenswrapper[4824]: I0121 11:28:03.977796 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Jan 21 11:28:03 crc kubenswrapper[4824]: I0121 11:28:03.990424 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-g9vq5"] Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.030860 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"deb46ee7-a757-433b-8595-ac21f9d3a03f","Type":"ContainerStarted","Data":"b2dbd538cfc7628790de49627fe1faff1d0f54a393e99dc66963d7845a74077e"} Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.068754 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4229a37f-5aa8-46c7-8ec4-74dde8937047" path="/var/lib/kubelet/pods/4229a37f-5aa8-46c7-8ec4-74dde8937047/volumes" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.077295 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/baa34da9-3711-4cd4-a5eb-ee5814ac51a7-scripts\") pod \"nova-cell0-cell-mapping-g9vq5\" (UID: \"baa34da9-3711-4cd4-a5eb-ee5814ac51a7\") " pod="openstack/nova-cell0-cell-mapping-g9vq5" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.077375 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/baa34da9-3711-4cd4-a5eb-ee5814ac51a7-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-g9vq5\" (UID: \"baa34da9-3711-4cd4-a5eb-ee5814ac51a7\") " pod="openstack/nova-cell0-cell-mapping-g9vq5" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.077580 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/baa34da9-3711-4cd4-a5eb-ee5814ac51a7-config-data\") pod \"nova-cell0-cell-mapping-g9vq5\" (UID: \"baa34da9-3711-4cd4-a5eb-ee5814ac51a7\") " pod="openstack/nova-cell0-cell-mapping-g9vq5" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.077620 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2xnkr\" (UniqueName: \"kubernetes.io/projected/baa34da9-3711-4cd4-a5eb-ee5814ac51a7-kube-api-access-2xnkr\") pod \"nova-cell0-cell-mapping-g9vq5\" (UID: \"baa34da9-3711-4cd4-a5eb-ee5814ac51a7\") " pod="openstack/nova-cell0-cell-mapping-g9vq5" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.081890 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.103004 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.103088 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.104315 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.105407 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.109597 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.109832 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.159433 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.178124 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.179262 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.185004 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.191697 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/baa34da9-3711-4cd4-a5eb-ee5814ac51a7-config-data\") pod \"nova-cell0-cell-mapping-g9vq5\" (UID: \"baa34da9-3711-4cd4-a5eb-ee5814ac51a7\") " pod="openstack/nova-cell0-cell-mapping-g9vq5" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.192384 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2xnkr\" (UniqueName: \"kubernetes.io/projected/baa34da9-3711-4cd4-a5eb-ee5814ac51a7-kube-api-access-2xnkr\") pod \"nova-cell0-cell-mapping-g9vq5\" (UID: \"baa34da9-3711-4cd4-a5eb-ee5814ac51a7\") " pod="openstack/nova-cell0-cell-mapping-g9vq5" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.192549 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/baa34da9-3711-4cd4-a5eb-ee5814ac51a7-scripts\") pod \"nova-cell0-cell-mapping-g9vq5\" (UID: \"baa34da9-3711-4cd4-a5eb-ee5814ac51a7\") " pod="openstack/nova-cell0-cell-mapping-g9vq5" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.192677 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/baa34da9-3711-4cd4-a5eb-ee5814ac51a7-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-g9vq5\" (UID: \"baa34da9-3711-4cd4-a5eb-ee5814ac51a7\") " pod="openstack/nova-cell0-cell-mapping-g9vq5" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.196745 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/baa34da9-3711-4cd4-a5eb-ee5814ac51a7-config-data\") pod \"nova-cell0-cell-mapping-g9vq5\" (UID: \"baa34da9-3711-4cd4-a5eb-ee5814ac51a7\") " pod="openstack/nova-cell0-cell-mapping-g9vq5" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.197935 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/baa34da9-3711-4cd4-a5eb-ee5814ac51a7-scripts\") pod \"nova-cell0-cell-mapping-g9vq5\" (UID: \"baa34da9-3711-4cd4-a5eb-ee5814ac51a7\") " pod="openstack/nova-cell0-cell-mapping-g9vq5" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.216427 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/baa34da9-3711-4cd4-a5eb-ee5814ac51a7-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-g9vq5\" (UID: \"baa34da9-3711-4cd4-a5eb-ee5814ac51a7\") " pod="openstack/nova-cell0-cell-mapping-g9vq5" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.231327 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2xnkr\" (UniqueName: \"kubernetes.io/projected/baa34da9-3711-4cd4-a5eb-ee5814ac51a7-kube-api-access-2xnkr\") pod \"nova-cell0-cell-mapping-g9vq5\" (UID: \"baa34da9-3711-4cd4-a5eb-ee5814ac51a7\") " pod="openstack/nova-cell0-cell-mapping-g9vq5" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.289920 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.294130 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-g9vq5" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.295237 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4a414d51-aaa8-43a6-88f5-f3ca59eb87a6-config-data\") pod \"nova-scheduler-0\" (UID: \"4a414d51-aaa8-43a6-88f5-f3ca59eb87a6\") " pod="openstack/nova-scheduler-0" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.295271 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f3a7f750-590d-43f3-a243-ec2548d1d4d8-logs\") pod \"nova-metadata-0\" (UID: \"f3a7f750-590d-43f3-a243-ec2548d1d4d8\") " pod="openstack/nova-metadata-0" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.295288 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-znk42\" (UniqueName: \"kubernetes.io/projected/4a414d51-aaa8-43a6-88f5-f3ca59eb87a6-kube-api-access-znk42\") pod \"nova-scheduler-0\" (UID: \"4a414d51-aaa8-43a6-88f5-f3ca59eb87a6\") " pod="openstack/nova-scheduler-0" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.295309 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a414d51-aaa8-43a6-88f5-f3ca59eb87a6-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"4a414d51-aaa8-43a6-88f5-f3ca59eb87a6\") " pod="openstack/nova-scheduler-0" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.295325 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9934f06c-6f15-4ca5-95f1-4c5bb27e72e8-logs\") pod \"nova-api-0\" (UID: \"9934f06c-6f15-4ca5-95f1-4c5bb27e72e8\") " pod="openstack/nova-api-0" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.295348 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f3a7f750-590d-43f3-a243-ec2548d1d4d8-config-data\") pod \"nova-metadata-0\" (UID: \"f3a7f750-590d-43f3-a243-ec2548d1d4d8\") " pod="openstack/nova-metadata-0" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.295381 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pjvlf\" (UniqueName: \"kubernetes.io/projected/9934f06c-6f15-4ca5-95f1-4c5bb27e72e8-kube-api-access-pjvlf\") pod \"nova-api-0\" (UID: \"9934f06c-6f15-4ca5-95f1-4c5bb27e72e8\") " pod="openstack/nova-api-0" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.295438 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3a7f750-590d-43f3-a243-ec2548d1d4d8-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"f3a7f750-590d-43f3-a243-ec2548d1d4d8\") " pod="openstack/nova-metadata-0" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.295462 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9934f06c-6f15-4ca5-95f1-4c5bb27e72e8-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"9934f06c-6f15-4ca5-95f1-4c5bb27e72e8\") " pod="openstack/nova-api-0" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.295526 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l9drq\" (UniqueName: \"kubernetes.io/projected/f3a7f750-590d-43f3-a243-ec2548d1d4d8-kube-api-access-l9drq\") pod \"nova-metadata-0\" (UID: \"f3a7f750-590d-43f3-a243-ec2548d1d4d8\") " pod="openstack/nova-metadata-0" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.295550 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9934f06c-6f15-4ca5-95f1-4c5bb27e72e8-config-data\") pod \"nova-api-0\" (UID: \"9934f06c-6f15-4ca5-95f1-4c5bb27e72e8\") " pod="openstack/nova-api-0" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.337576 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.338696 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.341809 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.357885 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.373754 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-9b86998b5-lwrg9"] Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.375233 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-9b86998b5-lwrg9" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.392531 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-9b86998b5-lwrg9"] Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.397097 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4a414d51-aaa8-43a6-88f5-f3ca59eb87a6-config-data\") pod \"nova-scheduler-0\" (UID: \"4a414d51-aaa8-43a6-88f5-f3ca59eb87a6\") " pod="openstack/nova-scheduler-0" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.397143 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f3a7f750-590d-43f3-a243-ec2548d1d4d8-logs\") pod \"nova-metadata-0\" (UID: \"f3a7f750-590d-43f3-a243-ec2548d1d4d8\") " pod="openstack/nova-metadata-0" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.397164 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-znk42\" (UniqueName: \"kubernetes.io/projected/4a414d51-aaa8-43a6-88f5-f3ca59eb87a6-kube-api-access-znk42\") pod \"nova-scheduler-0\" (UID: \"4a414d51-aaa8-43a6-88f5-f3ca59eb87a6\") " pod="openstack/nova-scheduler-0" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.397207 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a414d51-aaa8-43a6-88f5-f3ca59eb87a6-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"4a414d51-aaa8-43a6-88f5-f3ca59eb87a6\") " pod="openstack/nova-scheduler-0" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.397229 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9934f06c-6f15-4ca5-95f1-4c5bb27e72e8-logs\") pod \"nova-api-0\" (UID: \"9934f06c-6f15-4ca5-95f1-4c5bb27e72e8\") " pod="openstack/nova-api-0" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.397270 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f3a7f750-590d-43f3-a243-ec2548d1d4d8-config-data\") pod \"nova-metadata-0\" (UID: \"f3a7f750-590d-43f3-a243-ec2548d1d4d8\") " pod="openstack/nova-metadata-0" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.397344 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pjvlf\" (UniqueName: \"kubernetes.io/projected/9934f06c-6f15-4ca5-95f1-4c5bb27e72e8-kube-api-access-pjvlf\") pod \"nova-api-0\" (UID: \"9934f06c-6f15-4ca5-95f1-4c5bb27e72e8\") " pod="openstack/nova-api-0" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.397458 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3a7f750-590d-43f3-a243-ec2548d1d4d8-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"f3a7f750-590d-43f3-a243-ec2548d1d4d8\") " pod="openstack/nova-metadata-0" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.397500 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9934f06c-6f15-4ca5-95f1-4c5bb27e72e8-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"9934f06c-6f15-4ca5-95f1-4c5bb27e72e8\") " pod="openstack/nova-api-0" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.397643 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l9drq\" (UniqueName: \"kubernetes.io/projected/f3a7f750-590d-43f3-a243-ec2548d1d4d8-kube-api-access-l9drq\") pod \"nova-metadata-0\" (UID: \"f3a7f750-590d-43f3-a243-ec2548d1d4d8\") " pod="openstack/nova-metadata-0" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.397688 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9934f06c-6f15-4ca5-95f1-4c5bb27e72e8-config-data\") pod \"nova-api-0\" (UID: \"9934f06c-6f15-4ca5-95f1-4c5bb27e72e8\") " pod="openstack/nova-api-0" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.399601 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f3a7f750-590d-43f3-a243-ec2548d1d4d8-logs\") pod \"nova-metadata-0\" (UID: \"f3a7f750-590d-43f3-a243-ec2548d1d4d8\") " pod="openstack/nova-metadata-0" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.400137 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9934f06c-6f15-4ca5-95f1-4c5bb27e72e8-logs\") pod \"nova-api-0\" (UID: \"9934f06c-6f15-4ca5-95f1-4c5bb27e72e8\") " pod="openstack/nova-api-0" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.405771 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9934f06c-6f15-4ca5-95f1-4c5bb27e72e8-config-data\") pod \"nova-api-0\" (UID: \"9934f06c-6f15-4ca5-95f1-4c5bb27e72e8\") " pod="openstack/nova-api-0" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.411268 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a414d51-aaa8-43a6-88f5-f3ca59eb87a6-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"4a414d51-aaa8-43a6-88f5-f3ca59eb87a6\") " pod="openstack/nova-scheduler-0" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.415224 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3a7f750-590d-43f3-a243-ec2548d1d4d8-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"f3a7f750-590d-43f3-a243-ec2548d1d4d8\") " pod="openstack/nova-metadata-0" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.415486 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-znk42\" (UniqueName: \"kubernetes.io/projected/4a414d51-aaa8-43a6-88f5-f3ca59eb87a6-kube-api-access-znk42\") pod \"nova-scheduler-0\" (UID: \"4a414d51-aaa8-43a6-88f5-f3ca59eb87a6\") " pod="openstack/nova-scheduler-0" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.416261 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f3a7f750-590d-43f3-a243-ec2548d1d4d8-config-data\") pod \"nova-metadata-0\" (UID: \"f3a7f750-590d-43f3-a243-ec2548d1d4d8\") " pod="openstack/nova-metadata-0" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.416507 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4a414d51-aaa8-43a6-88f5-f3ca59eb87a6-config-data\") pod \"nova-scheduler-0\" (UID: \"4a414d51-aaa8-43a6-88f5-f3ca59eb87a6\") " pod="openstack/nova-scheduler-0" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.416646 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pjvlf\" (UniqueName: \"kubernetes.io/projected/9934f06c-6f15-4ca5-95f1-4c5bb27e72e8-kube-api-access-pjvlf\") pod \"nova-api-0\" (UID: \"9934f06c-6f15-4ca5-95f1-4c5bb27e72e8\") " pod="openstack/nova-api-0" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.421566 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9934f06c-6f15-4ca5-95f1-4c5bb27e72e8-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"9934f06c-6f15-4ca5-95f1-4c5bb27e72e8\") " pod="openstack/nova-api-0" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.433396 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l9drq\" (UniqueName: \"kubernetes.io/projected/f3a7f750-590d-43f3-a243-ec2548d1d4d8-kube-api-access-l9drq\") pod \"nova-metadata-0\" (UID: \"f3a7f750-590d-43f3-a243-ec2548d1d4d8\") " pod="openstack/nova-metadata-0" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.501873 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8247185f-987d-4f6e-b633-042907078e39-dns-swift-storage-0\") pod \"dnsmasq-dns-9b86998b5-lwrg9\" (UID: \"8247185f-987d-4f6e-b633-042907078e39\") " pod="openstack/dnsmasq-dns-9b86998b5-lwrg9" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.502234 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8247185f-987d-4f6e-b633-042907078e39-ovsdbserver-sb\") pod \"dnsmasq-dns-9b86998b5-lwrg9\" (UID: \"8247185f-987d-4f6e-b633-042907078e39\") " pod="openstack/dnsmasq-dns-9b86998b5-lwrg9" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.502257 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8247185f-987d-4f6e-b633-042907078e39-config\") pod \"dnsmasq-dns-9b86998b5-lwrg9\" (UID: \"8247185f-987d-4f6e-b633-042907078e39\") " pod="openstack/dnsmasq-dns-9b86998b5-lwrg9" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.502304 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d373f159-2b64-4dd6-9169-a96a01afcfce-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"d373f159-2b64-4dd6-9169-a96a01afcfce\") " pod="openstack/nova-cell1-novncproxy-0" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.502331 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8247185f-987d-4f6e-b633-042907078e39-dns-svc\") pod \"dnsmasq-dns-9b86998b5-lwrg9\" (UID: \"8247185f-987d-4f6e-b633-042907078e39\") " pod="openstack/dnsmasq-dns-9b86998b5-lwrg9" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.502386 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9kz5l\" (UniqueName: \"kubernetes.io/projected/8247185f-987d-4f6e-b633-042907078e39-kube-api-access-9kz5l\") pod \"dnsmasq-dns-9b86998b5-lwrg9\" (UID: \"8247185f-987d-4f6e-b633-042907078e39\") " pod="openstack/dnsmasq-dns-9b86998b5-lwrg9" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.502426 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d373f159-2b64-4dd6-9169-a96a01afcfce-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"d373f159-2b64-4dd6-9169-a96a01afcfce\") " pod="openstack/nova-cell1-novncproxy-0" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.502488 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z6l4s\" (UniqueName: \"kubernetes.io/projected/d373f159-2b64-4dd6-9169-a96a01afcfce-kube-api-access-z6l4s\") pod \"nova-cell1-novncproxy-0\" (UID: \"d373f159-2b64-4dd6-9169-a96a01afcfce\") " pod="openstack/nova-cell1-novncproxy-0" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.502561 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8247185f-987d-4f6e-b633-042907078e39-ovsdbserver-nb\") pod \"dnsmasq-dns-9b86998b5-lwrg9\" (UID: \"8247185f-987d-4f6e-b633-042907078e39\") " pod="openstack/dnsmasq-dns-9b86998b5-lwrg9" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.526332 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.537143 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.575281 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.605941 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8247185f-987d-4f6e-b633-042907078e39-ovsdbserver-sb\") pod \"dnsmasq-dns-9b86998b5-lwrg9\" (UID: \"8247185f-987d-4f6e-b633-042907078e39\") " pod="openstack/dnsmasq-dns-9b86998b5-lwrg9" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.613176 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8247185f-987d-4f6e-b633-042907078e39-config\") pod \"dnsmasq-dns-9b86998b5-lwrg9\" (UID: \"8247185f-987d-4f6e-b633-042907078e39\") " pod="openstack/dnsmasq-dns-9b86998b5-lwrg9" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.606688 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8247185f-987d-4f6e-b633-042907078e39-ovsdbserver-sb\") pod \"dnsmasq-dns-9b86998b5-lwrg9\" (UID: \"8247185f-987d-4f6e-b633-042907078e39\") " pod="openstack/dnsmasq-dns-9b86998b5-lwrg9" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.613232 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d373f159-2b64-4dd6-9169-a96a01afcfce-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"d373f159-2b64-4dd6-9169-a96a01afcfce\") " pod="openstack/nova-cell1-novncproxy-0" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.613798 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8247185f-987d-4f6e-b633-042907078e39-dns-svc\") pod \"dnsmasq-dns-9b86998b5-lwrg9\" (UID: \"8247185f-987d-4f6e-b633-042907078e39\") " pod="openstack/dnsmasq-dns-9b86998b5-lwrg9" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.613832 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8247185f-987d-4f6e-b633-042907078e39-config\") pod \"dnsmasq-dns-9b86998b5-lwrg9\" (UID: \"8247185f-987d-4f6e-b633-042907078e39\") " pod="openstack/dnsmasq-dns-9b86998b5-lwrg9" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.613899 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9kz5l\" (UniqueName: \"kubernetes.io/projected/8247185f-987d-4f6e-b633-042907078e39-kube-api-access-9kz5l\") pod \"dnsmasq-dns-9b86998b5-lwrg9\" (UID: \"8247185f-987d-4f6e-b633-042907078e39\") " pod="openstack/dnsmasq-dns-9b86998b5-lwrg9" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.613949 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d373f159-2b64-4dd6-9169-a96a01afcfce-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"d373f159-2b64-4dd6-9169-a96a01afcfce\") " pod="openstack/nova-cell1-novncproxy-0" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.614037 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z6l4s\" (UniqueName: \"kubernetes.io/projected/d373f159-2b64-4dd6-9169-a96a01afcfce-kube-api-access-z6l4s\") pod \"nova-cell1-novncproxy-0\" (UID: \"d373f159-2b64-4dd6-9169-a96a01afcfce\") " pod="openstack/nova-cell1-novncproxy-0" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.614121 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8247185f-987d-4f6e-b633-042907078e39-ovsdbserver-nb\") pod \"dnsmasq-dns-9b86998b5-lwrg9\" (UID: \"8247185f-987d-4f6e-b633-042907078e39\") " pod="openstack/dnsmasq-dns-9b86998b5-lwrg9" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.614166 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8247185f-987d-4f6e-b633-042907078e39-dns-swift-storage-0\") pod \"dnsmasq-dns-9b86998b5-lwrg9\" (UID: \"8247185f-987d-4f6e-b633-042907078e39\") " pod="openstack/dnsmasq-dns-9b86998b5-lwrg9" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.614480 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8247185f-987d-4f6e-b633-042907078e39-dns-svc\") pod \"dnsmasq-dns-9b86998b5-lwrg9\" (UID: \"8247185f-987d-4f6e-b633-042907078e39\") " pod="openstack/dnsmasq-dns-9b86998b5-lwrg9" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.614720 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8247185f-987d-4f6e-b633-042907078e39-dns-swift-storage-0\") pod \"dnsmasq-dns-9b86998b5-lwrg9\" (UID: \"8247185f-987d-4f6e-b633-042907078e39\") " pod="openstack/dnsmasq-dns-9b86998b5-lwrg9" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.615319 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8247185f-987d-4f6e-b633-042907078e39-ovsdbserver-nb\") pod \"dnsmasq-dns-9b86998b5-lwrg9\" (UID: \"8247185f-987d-4f6e-b633-042907078e39\") " pod="openstack/dnsmasq-dns-9b86998b5-lwrg9" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.618517 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d373f159-2b64-4dd6-9169-a96a01afcfce-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"d373f159-2b64-4dd6-9169-a96a01afcfce\") " pod="openstack/nova-cell1-novncproxy-0" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.620297 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d373f159-2b64-4dd6-9169-a96a01afcfce-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"d373f159-2b64-4dd6-9169-a96a01afcfce\") " pod="openstack/nova-cell1-novncproxy-0" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.632372 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9kz5l\" (UniqueName: \"kubernetes.io/projected/8247185f-987d-4f6e-b633-042907078e39-kube-api-access-9kz5l\") pod \"dnsmasq-dns-9b86998b5-lwrg9\" (UID: \"8247185f-987d-4f6e-b633-042907078e39\") " pod="openstack/dnsmasq-dns-9b86998b5-lwrg9" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.635261 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z6l4s\" (UniqueName: \"kubernetes.io/projected/d373f159-2b64-4dd6-9169-a96a01afcfce-kube-api-access-z6l4s\") pod \"nova-cell1-novncproxy-0\" (UID: \"d373f159-2b64-4dd6-9169-a96a01afcfce\") " pod="openstack/nova-cell1-novncproxy-0" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.674479 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.825723 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-9b86998b5-lwrg9" Jan 21 11:28:04 crc kubenswrapper[4824]: I0121 11:28:04.851443 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-g9vq5"] Jan 21 11:28:05 crc kubenswrapper[4824]: I0121 11:28:05.067527 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-g9vq5" event={"ID":"baa34da9-3711-4cd4-a5eb-ee5814ac51a7","Type":"ContainerStarted","Data":"b818724198e29357d179e92d306ffc141219f83ff62bb7dd6d3881b003b27078"} Jan 21 11:28:05 crc kubenswrapper[4824]: I0121 11:28:05.072830 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"deb46ee7-a757-433b-8595-ac21f9d3a03f","Type":"ContainerStarted","Data":"03674de70b88a6dae6f9a3d4e9d1399fcaa1ecb7056d38626af5e46c5a1612f6"} Jan 21 11:28:05 crc kubenswrapper[4824]: I0121 11:28:05.122350 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Jan 21 11:28:05 crc kubenswrapper[4824]: I0121 11:28:05.150942 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Jan 21 11:28:05 crc kubenswrapper[4824]: I0121 11:28:05.271809 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Jan 21 11:28:05 crc kubenswrapper[4824]: I0121 11:28:05.436914 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Jan 21 11:28:05 crc kubenswrapper[4824]: I0121 11:28:05.603035 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-9b86998b5-lwrg9"] Jan 21 11:28:06 crc kubenswrapper[4824]: I0121 11:28:06.114465 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"9934f06c-6f15-4ca5-95f1-4c5bb27e72e8","Type":"ContainerStarted","Data":"2c15246588a81cb1f14c8e607d836b282630b2c662037ff7639bacf65e5fa2ae"} Jan 21 11:28:06 crc kubenswrapper[4824]: I0121 11:28:06.116807 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"deb46ee7-a757-433b-8595-ac21f9d3a03f","Type":"ContainerStarted","Data":"64748c9826d4f022251fc061ac66e44525e85e4e0e8ee1dc2d4a0b3f328468c3"} Jan 21 11:28:06 crc kubenswrapper[4824]: I0121 11:28:06.120651 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"4a414d51-aaa8-43a6-88f5-f3ca59eb87a6","Type":"ContainerStarted","Data":"720379547d053b3767f55293ad70e28fa03d8b901bf9feb0905f084febf1fbbe"} Jan 21 11:28:06 crc kubenswrapper[4824]: I0121 11:28:06.126220 4824 generic.go:334] "Generic (PLEG): container finished" podID="8247185f-987d-4f6e-b633-042907078e39" containerID="3f015c1fd2be4ed1835afc4169552ab51d97958ba2fb966dc794c0c63861420b" exitCode=0 Jan 21 11:28:06 crc kubenswrapper[4824]: I0121 11:28:06.126267 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-9b86998b5-lwrg9" event={"ID":"8247185f-987d-4f6e-b633-042907078e39","Type":"ContainerDied","Data":"3f015c1fd2be4ed1835afc4169552ab51d97958ba2fb966dc794c0c63861420b"} Jan 21 11:28:06 crc kubenswrapper[4824]: I0121 11:28:06.126286 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-9b86998b5-lwrg9" event={"ID":"8247185f-987d-4f6e-b633-042907078e39","Type":"ContainerStarted","Data":"cee4d42d6e8f113c3a920c0a14aa2caa944d9803b04e59596b4d760c6b1d3e91"} Jan 21 11:28:06 crc kubenswrapper[4824]: I0121 11:28:06.136173 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f3a7f750-590d-43f3-a243-ec2548d1d4d8","Type":"ContainerStarted","Data":"d793c49926c6cf8bd7a478c6e30627a0b61508eda4904423a7dd0b8012e4b62b"} Jan 21 11:28:06 crc kubenswrapper[4824]: I0121 11:28:06.138190 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"d373f159-2b64-4dd6-9169-a96a01afcfce","Type":"ContainerStarted","Data":"7669fbc7cdadc946c7f940b38d316891febbd77d056792c26182131fc73631b1"} Jan 21 11:28:06 crc kubenswrapper[4824]: I0121 11:28:06.139866 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-g9vq5" event={"ID":"baa34da9-3711-4cd4-a5eb-ee5814ac51a7","Type":"ContainerStarted","Data":"dd76c88e84c244299a1fbb85cad5296bd9ab946000d0d268303768f60e22c9c4"} Jan 21 11:28:06 crc kubenswrapper[4824]: I0121 11:28:06.525664 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-g9vq5" podStartSLOduration=3.5256496950000002 podStartE2EDuration="3.525649695s" podCreationTimestamp="2026-01-21 11:28:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:28:06.170845113 +0000 UTC m=+1028.463874406" watchObservedRunningTime="2026-01-21 11:28:06.525649695 +0000 UTC m=+1028.818678988" Jan 21 11:28:06 crc kubenswrapper[4824]: I0121 11:28:06.576378 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-7gtcg"] Jan 21 11:28:06 crc kubenswrapper[4824]: I0121 11:28:06.577487 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-7gtcg" Jan 21 11:28:06 crc kubenswrapper[4824]: I0121 11:28:06.581551 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Jan 21 11:28:06 crc kubenswrapper[4824]: I0121 11:28:06.588372 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Jan 21 11:28:06 crc kubenswrapper[4824]: I0121 11:28:06.597850 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-7gtcg"] Jan 21 11:28:06 crc kubenswrapper[4824]: I0121 11:28:06.683123 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vv6bv\" (UniqueName: \"kubernetes.io/projected/135a8573-b55a-4c5f-9cb2-a7c3adea9720-kube-api-access-vv6bv\") pod \"nova-cell1-conductor-db-sync-7gtcg\" (UID: \"135a8573-b55a-4c5f-9cb2-a7c3adea9720\") " pod="openstack/nova-cell1-conductor-db-sync-7gtcg" Jan 21 11:28:06 crc kubenswrapper[4824]: I0121 11:28:06.683222 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/135a8573-b55a-4c5f-9cb2-a7c3adea9720-config-data\") pod \"nova-cell1-conductor-db-sync-7gtcg\" (UID: \"135a8573-b55a-4c5f-9cb2-a7c3adea9720\") " pod="openstack/nova-cell1-conductor-db-sync-7gtcg" Jan 21 11:28:06 crc kubenswrapper[4824]: I0121 11:28:06.683246 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/135a8573-b55a-4c5f-9cb2-a7c3adea9720-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-7gtcg\" (UID: \"135a8573-b55a-4c5f-9cb2-a7c3adea9720\") " pod="openstack/nova-cell1-conductor-db-sync-7gtcg" Jan 21 11:28:06 crc kubenswrapper[4824]: I0121 11:28:06.683309 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/135a8573-b55a-4c5f-9cb2-a7c3adea9720-scripts\") pod \"nova-cell1-conductor-db-sync-7gtcg\" (UID: \"135a8573-b55a-4c5f-9cb2-a7c3adea9720\") " pod="openstack/nova-cell1-conductor-db-sync-7gtcg" Jan 21 11:28:06 crc kubenswrapper[4824]: I0121 11:28:06.785888 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/135a8573-b55a-4c5f-9cb2-a7c3adea9720-config-data\") pod \"nova-cell1-conductor-db-sync-7gtcg\" (UID: \"135a8573-b55a-4c5f-9cb2-a7c3adea9720\") " pod="openstack/nova-cell1-conductor-db-sync-7gtcg" Jan 21 11:28:06 crc kubenswrapper[4824]: I0121 11:28:06.785939 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/135a8573-b55a-4c5f-9cb2-a7c3adea9720-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-7gtcg\" (UID: \"135a8573-b55a-4c5f-9cb2-a7c3adea9720\") " pod="openstack/nova-cell1-conductor-db-sync-7gtcg" Jan 21 11:28:06 crc kubenswrapper[4824]: I0121 11:28:06.786020 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/135a8573-b55a-4c5f-9cb2-a7c3adea9720-scripts\") pod \"nova-cell1-conductor-db-sync-7gtcg\" (UID: \"135a8573-b55a-4c5f-9cb2-a7c3adea9720\") " pod="openstack/nova-cell1-conductor-db-sync-7gtcg" Jan 21 11:28:06 crc kubenswrapper[4824]: I0121 11:28:06.786147 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vv6bv\" (UniqueName: \"kubernetes.io/projected/135a8573-b55a-4c5f-9cb2-a7c3adea9720-kube-api-access-vv6bv\") pod \"nova-cell1-conductor-db-sync-7gtcg\" (UID: \"135a8573-b55a-4c5f-9cb2-a7c3adea9720\") " pod="openstack/nova-cell1-conductor-db-sync-7gtcg" Jan 21 11:28:06 crc kubenswrapper[4824]: I0121 11:28:06.802880 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/135a8573-b55a-4c5f-9cb2-a7c3adea9720-config-data\") pod \"nova-cell1-conductor-db-sync-7gtcg\" (UID: \"135a8573-b55a-4c5f-9cb2-a7c3adea9720\") " pod="openstack/nova-cell1-conductor-db-sync-7gtcg" Jan 21 11:28:06 crc kubenswrapper[4824]: I0121 11:28:06.808550 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/135a8573-b55a-4c5f-9cb2-a7c3adea9720-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-7gtcg\" (UID: \"135a8573-b55a-4c5f-9cb2-a7c3adea9720\") " pod="openstack/nova-cell1-conductor-db-sync-7gtcg" Jan 21 11:28:06 crc kubenswrapper[4824]: I0121 11:28:06.810403 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vv6bv\" (UniqueName: \"kubernetes.io/projected/135a8573-b55a-4c5f-9cb2-a7c3adea9720-kube-api-access-vv6bv\") pod \"nova-cell1-conductor-db-sync-7gtcg\" (UID: \"135a8573-b55a-4c5f-9cb2-a7c3adea9720\") " pod="openstack/nova-cell1-conductor-db-sync-7gtcg" Jan 21 11:28:06 crc kubenswrapper[4824]: I0121 11:28:06.811485 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/135a8573-b55a-4c5f-9cb2-a7c3adea9720-scripts\") pod \"nova-cell1-conductor-db-sync-7gtcg\" (UID: \"135a8573-b55a-4c5f-9cb2-a7c3adea9720\") " pod="openstack/nova-cell1-conductor-db-sync-7gtcg" Jan 21 11:28:06 crc kubenswrapper[4824]: I0121 11:28:06.948039 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-7gtcg" Jan 21 11:28:07 crc kubenswrapper[4824]: I0121 11:28:07.157341 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"deb46ee7-a757-433b-8595-ac21f9d3a03f","Type":"ContainerStarted","Data":"70559b4f16a38253804eec43f7ada16a4c4b14e98a6505d8ed5b38a8a1ac6f82"} Jan 21 11:28:07 crc kubenswrapper[4824]: I0121 11:28:07.157513 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="deb46ee7-a757-433b-8595-ac21f9d3a03f" containerName="ceilometer-central-agent" containerID="cri-o://b2dbd538cfc7628790de49627fe1faff1d0f54a393e99dc66963d7845a74077e" gracePeriod=30 Jan 21 11:28:07 crc kubenswrapper[4824]: I0121 11:28:07.157776 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="deb46ee7-a757-433b-8595-ac21f9d3a03f" containerName="proxy-httpd" containerID="cri-o://70559b4f16a38253804eec43f7ada16a4c4b14e98a6505d8ed5b38a8a1ac6f82" gracePeriod=30 Jan 21 11:28:07 crc kubenswrapper[4824]: I0121 11:28:07.157811 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Jan 21 11:28:07 crc kubenswrapper[4824]: I0121 11:28:07.157836 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="deb46ee7-a757-433b-8595-ac21f9d3a03f" containerName="sg-core" containerID="cri-o://64748c9826d4f022251fc061ac66e44525e85e4e0e8ee1dc2d4a0b3f328468c3" gracePeriod=30 Jan 21 11:28:07 crc kubenswrapper[4824]: I0121 11:28:07.157890 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="deb46ee7-a757-433b-8595-ac21f9d3a03f" containerName="ceilometer-notification-agent" containerID="cri-o://03674de70b88a6dae6f9a3d4e9d1399fcaa1ecb7056d38626af5e46c5a1612f6" gracePeriod=30 Jan 21 11:28:07 crc kubenswrapper[4824]: I0121 11:28:07.170565 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-9b86998b5-lwrg9" event={"ID":"8247185f-987d-4f6e-b633-042907078e39","Type":"ContainerStarted","Data":"32fb68e8168ce229b5eee2857a8b6f19fa0565f1c232d572bde25ae83e70fb37"} Jan 21 11:28:07 crc kubenswrapper[4824]: I0121 11:28:07.179905 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.638577278 podStartE2EDuration="5.17989422s" podCreationTimestamp="2026-01-21 11:28:02 +0000 UTC" firstStartedPulling="2026-01-21 11:28:02.790558451 +0000 UTC m=+1025.083587743" lastFinishedPulling="2026-01-21 11:28:06.331875393 +0000 UTC m=+1028.624904685" observedRunningTime="2026-01-21 11:28:07.17479897 +0000 UTC m=+1029.467828262" watchObservedRunningTime="2026-01-21 11:28:07.17989422 +0000 UTC m=+1029.472923512" Jan 21 11:28:07 crc kubenswrapper[4824]: I0121 11:28:07.182624 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-9b86998b5-lwrg9" Jan 21 11:28:07 crc kubenswrapper[4824]: I0121 11:28:07.517771 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-9b86998b5-lwrg9" podStartSLOduration=3.517756134 podStartE2EDuration="3.517756134s" podCreationTimestamp="2026-01-21 11:28:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:28:07.206919767 +0000 UTC m=+1029.499949059" watchObservedRunningTime="2026-01-21 11:28:07.517756134 +0000 UTC m=+1029.810785427" Jan 21 11:28:07 crc kubenswrapper[4824]: I0121 11:28:07.523437 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Jan 21 11:28:07 crc kubenswrapper[4824]: I0121 11:28:07.530024 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-7gtcg"] Jan 21 11:28:07 crc kubenswrapper[4824]: I0121 11:28:07.536333 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Jan 21 11:28:08 crc kubenswrapper[4824]: I0121 11:28:08.182209 4824 generic.go:334] "Generic (PLEG): container finished" podID="deb46ee7-a757-433b-8595-ac21f9d3a03f" containerID="70559b4f16a38253804eec43f7ada16a4c4b14e98a6505d8ed5b38a8a1ac6f82" exitCode=0 Jan 21 11:28:08 crc kubenswrapper[4824]: I0121 11:28:08.182401 4824 generic.go:334] "Generic (PLEG): container finished" podID="deb46ee7-a757-433b-8595-ac21f9d3a03f" containerID="64748c9826d4f022251fc061ac66e44525e85e4e0e8ee1dc2d4a0b3f328468c3" exitCode=2 Jan 21 11:28:08 crc kubenswrapper[4824]: I0121 11:28:08.182410 4824 generic.go:334] "Generic (PLEG): container finished" podID="deb46ee7-a757-433b-8595-ac21f9d3a03f" containerID="03674de70b88a6dae6f9a3d4e9d1399fcaa1ecb7056d38626af5e46c5a1612f6" exitCode=0 Jan 21 11:28:08 crc kubenswrapper[4824]: I0121 11:28:08.183255 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"deb46ee7-a757-433b-8595-ac21f9d3a03f","Type":"ContainerDied","Data":"70559b4f16a38253804eec43f7ada16a4c4b14e98a6505d8ed5b38a8a1ac6f82"} Jan 21 11:28:08 crc kubenswrapper[4824]: I0121 11:28:08.183279 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"deb46ee7-a757-433b-8595-ac21f9d3a03f","Type":"ContainerDied","Data":"64748c9826d4f022251fc061ac66e44525e85e4e0e8ee1dc2d4a0b3f328468c3"} Jan 21 11:28:08 crc kubenswrapper[4824]: I0121 11:28:08.183290 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"deb46ee7-a757-433b-8595-ac21f9d3a03f","Type":"ContainerDied","Data":"03674de70b88a6dae6f9a3d4e9d1399fcaa1ecb7056d38626af5e46c5a1612f6"} Jan 21 11:28:08 crc kubenswrapper[4824]: W0121 11:28:08.708314 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod135a8573_b55a_4c5f_9cb2_a7c3adea9720.slice/crio-74afad19ccfca1492a879f61489b7e643f3472e59cfcfff81bc9271163b2f5df WatchSource:0}: Error finding container 74afad19ccfca1492a879f61489b7e643f3472e59cfcfff81bc9271163b2f5df: Status 404 returned error can't find the container with id 74afad19ccfca1492a879f61489b7e643f3472e59cfcfff81bc9271163b2f5df Jan 21 11:28:09 crc kubenswrapper[4824]: I0121 11:28:09.193759 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-7gtcg" event={"ID":"135a8573-b55a-4c5f-9cb2-a7c3adea9720","Type":"ContainerStarted","Data":"74afad19ccfca1492a879f61489b7e643f3472e59cfcfff81bc9271163b2f5df"} Jan 21 11:28:10 crc kubenswrapper[4824]: I0121 11:28:10.202652 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f3a7f750-590d-43f3-a243-ec2548d1d4d8","Type":"ContainerStarted","Data":"3c697465f3dc9cbfbb17099585b5fcea916a9a851ee6407d34d2bb4f6f77dbb0"} Jan 21 11:28:10 crc kubenswrapper[4824]: I0121 11:28:10.202883 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f3a7f750-590d-43f3-a243-ec2548d1d4d8","Type":"ContainerStarted","Data":"1f864312bcd5d9294e3207dbe32578ac30ec3fffebcb657307393b540937095e"} Jan 21 11:28:10 crc kubenswrapper[4824]: I0121 11:28:10.202770 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="f3a7f750-590d-43f3-a243-ec2548d1d4d8" containerName="nova-metadata-metadata" containerID="cri-o://3c697465f3dc9cbfbb17099585b5fcea916a9a851ee6407d34d2bb4f6f77dbb0" gracePeriod=30 Jan 21 11:28:10 crc kubenswrapper[4824]: I0121 11:28:10.202711 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="f3a7f750-590d-43f3-a243-ec2548d1d4d8" containerName="nova-metadata-log" containerID="cri-o://1f864312bcd5d9294e3207dbe32578ac30ec3fffebcb657307393b540937095e" gracePeriod=30 Jan 21 11:28:10 crc kubenswrapper[4824]: I0121 11:28:10.204840 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"d373f159-2b64-4dd6-9169-a96a01afcfce","Type":"ContainerStarted","Data":"82cb1335346c23ca78526b5fd5db5f17dcd0b8161c12a767fa8321e0f3ab845f"} Jan 21 11:28:10 crc kubenswrapper[4824]: I0121 11:28:10.204918 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="d373f159-2b64-4dd6-9169-a96a01afcfce" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://82cb1335346c23ca78526b5fd5db5f17dcd0b8161c12a767fa8321e0f3ab845f" gracePeriod=30 Jan 21 11:28:10 crc kubenswrapper[4824]: I0121 11:28:10.209630 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"9934f06c-6f15-4ca5-95f1-4c5bb27e72e8","Type":"ContainerStarted","Data":"2dc23f89a2ecb53e2543111bc5e25eb8fa216de246339731c0b2bddd341062d7"} Jan 21 11:28:10 crc kubenswrapper[4824]: I0121 11:28:10.209666 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"9934f06c-6f15-4ca5-95f1-4c5bb27e72e8","Type":"ContainerStarted","Data":"79798a638b185d61504b6917c46892e6c77377d55b27c826aa0f1424cea017bf"} Jan 21 11:28:10 crc kubenswrapper[4824]: I0121 11:28:10.211204 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-7gtcg" event={"ID":"135a8573-b55a-4c5f-9cb2-a7c3adea9720","Type":"ContainerStarted","Data":"e6d7bd7f99afa7be3d7eee7ba3147eedee78587b76d733c8cb546689d1b785fe"} Jan 21 11:28:10 crc kubenswrapper[4824]: I0121 11:28:10.215292 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"4a414d51-aaa8-43a6-88f5-f3ca59eb87a6","Type":"ContainerStarted","Data":"e871bc06d477443b88ec205254bfc60975ceca19ed6da8e295be1de5bf3e986e"} Jan 21 11:28:10 crc kubenswrapper[4824]: I0121 11:28:10.224628 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.233256646 podStartE2EDuration="6.224613476s" podCreationTimestamp="2026-01-21 11:28:04 +0000 UTC" firstStartedPulling="2026-01-21 11:28:05.161974974 +0000 UTC m=+1027.455004266" lastFinishedPulling="2026-01-21 11:28:09.153331803 +0000 UTC m=+1031.446361096" observedRunningTime="2026-01-21 11:28:10.219770541 +0000 UTC m=+1032.512799832" watchObservedRunningTime="2026-01-21 11:28:10.224613476 +0000 UTC m=+1032.517642767" Jan 21 11:28:10 crc kubenswrapper[4824]: I0121 11:28:10.234780 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.544110698 podStartE2EDuration="6.234763501s" podCreationTimestamp="2026-01-21 11:28:04 +0000 UTC" firstStartedPulling="2026-01-21 11:28:05.45758397 +0000 UTC m=+1027.750613262" lastFinishedPulling="2026-01-21 11:28:09.148236773 +0000 UTC m=+1031.441266065" observedRunningTime="2026-01-21 11:28:10.233325138 +0000 UTC m=+1032.526354430" watchObservedRunningTime="2026-01-21 11:28:10.234763501 +0000 UTC m=+1032.527792792" Jan 21 11:28:10 crc kubenswrapper[4824]: I0121 11:28:10.251802 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.264828487 podStartE2EDuration="6.25177563s" podCreationTimestamp="2026-01-21 11:28:04 +0000 UTC" firstStartedPulling="2026-01-21 11:28:05.163597966 +0000 UTC m=+1027.456627257" lastFinishedPulling="2026-01-21 11:28:09.150545108 +0000 UTC m=+1031.443574400" observedRunningTime="2026-01-21 11:28:10.247140707 +0000 UTC m=+1032.540169999" watchObservedRunningTime="2026-01-21 11:28:10.25177563 +0000 UTC m=+1032.544804921" Jan 21 11:28:10 crc kubenswrapper[4824]: I0121 11:28:10.260262 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-7gtcg" podStartSLOduration=4.260245456 podStartE2EDuration="4.260245456s" podCreationTimestamp="2026-01-21 11:28:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:28:10.257749247 +0000 UTC m=+1032.550778539" watchObservedRunningTime="2026-01-21 11:28:10.260245456 +0000 UTC m=+1032.553274748" Jan 21 11:28:10 crc kubenswrapper[4824]: I0121 11:28:10.286980 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.421375541 podStartE2EDuration="6.28695164s" podCreationTimestamp="2026-01-21 11:28:04 +0000 UTC" firstStartedPulling="2026-01-21 11:28:05.284995309 +0000 UTC m=+1027.578024600" lastFinishedPulling="2026-01-21 11:28:09.150571407 +0000 UTC m=+1031.443600699" observedRunningTime="2026-01-21 11:28:10.269593748 +0000 UTC m=+1032.562623040" watchObservedRunningTime="2026-01-21 11:28:10.28695164 +0000 UTC m=+1032.579980932" Jan 21 11:28:10 crc kubenswrapper[4824]: I0121 11:28:10.710151 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Jan 21 11:28:10 crc kubenswrapper[4824]: I0121 11:28:10.773689 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3a7f750-590d-43f3-a243-ec2548d1d4d8-combined-ca-bundle\") pod \"f3a7f750-590d-43f3-a243-ec2548d1d4d8\" (UID: \"f3a7f750-590d-43f3-a243-ec2548d1d4d8\") " Jan 21 11:28:10 crc kubenswrapper[4824]: I0121 11:28:10.773767 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f3a7f750-590d-43f3-a243-ec2548d1d4d8-logs\") pod \"f3a7f750-590d-43f3-a243-ec2548d1d4d8\" (UID: \"f3a7f750-590d-43f3-a243-ec2548d1d4d8\") " Jan 21 11:28:10 crc kubenswrapper[4824]: I0121 11:28:10.773833 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f3a7f750-590d-43f3-a243-ec2548d1d4d8-config-data\") pod \"f3a7f750-590d-43f3-a243-ec2548d1d4d8\" (UID: \"f3a7f750-590d-43f3-a243-ec2548d1d4d8\") " Jan 21 11:28:10 crc kubenswrapper[4824]: I0121 11:28:10.773909 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l9drq\" (UniqueName: \"kubernetes.io/projected/f3a7f750-590d-43f3-a243-ec2548d1d4d8-kube-api-access-l9drq\") pod \"f3a7f750-590d-43f3-a243-ec2548d1d4d8\" (UID: \"f3a7f750-590d-43f3-a243-ec2548d1d4d8\") " Jan 21 11:28:10 crc kubenswrapper[4824]: I0121 11:28:10.774381 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f3a7f750-590d-43f3-a243-ec2548d1d4d8-logs" (OuterVolumeSpecName: "logs") pod "f3a7f750-590d-43f3-a243-ec2548d1d4d8" (UID: "f3a7f750-590d-43f3-a243-ec2548d1d4d8"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:28:10 crc kubenswrapper[4824]: I0121 11:28:10.790121 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f3a7f750-590d-43f3-a243-ec2548d1d4d8-kube-api-access-l9drq" (OuterVolumeSpecName: "kube-api-access-l9drq") pod "f3a7f750-590d-43f3-a243-ec2548d1d4d8" (UID: "f3a7f750-590d-43f3-a243-ec2548d1d4d8"). InnerVolumeSpecName "kube-api-access-l9drq". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:28:10 crc kubenswrapper[4824]: I0121 11:28:10.799327 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f3a7f750-590d-43f3-a243-ec2548d1d4d8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f3a7f750-590d-43f3-a243-ec2548d1d4d8" (UID: "f3a7f750-590d-43f3-a243-ec2548d1d4d8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:28:10 crc kubenswrapper[4824]: I0121 11:28:10.802161 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f3a7f750-590d-43f3-a243-ec2548d1d4d8-config-data" (OuterVolumeSpecName: "config-data") pod "f3a7f750-590d-43f3-a243-ec2548d1d4d8" (UID: "f3a7f750-590d-43f3-a243-ec2548d1d4d8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:28:10 crc kubenswrapper[4824]: I0121 11:28:10.876236 4824 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f3a7f750-590d-43f3-a243-ec2548d1d4d8-logs\") on node \"crc\" DevicePath \"\"" Jan 21 11:28:10 crc kubenswrapper[4824]: I0121 11:28:10.876282 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f3a7f750-590d-43f3-a243-ec2548d1d4d8-config-data\") on node \"crc\" DevicePath \"\"" Jan 21 11:28:10 crc kubenswrapper[4824]: I0121 11:28:10.876292 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l9drq\" (UniqueName: \"kubernetes.io/projected/f3a7f750-590d-43f3-a243-ec2548d1d4d8-kube-api-access-l9drq\") on node \"crc\" DevicePath \"\"" Jan 21 11:28:10 crc kubenswrapper[4824]: I0121 11:28:10.876302 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3a7f750-590d-43f3-a243-ec2548d1d4d8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 21 11:28:11 crc kubenswrapper[4824]: I0121 11:28:11.223465 4824 generic.go:334] "Generic (PLEG): container finished" podID="f3a7f750-590d-43f3-a243-ec2548d1d4d8" containerID="3c697465f3dc9cbfbb17099585b5fcea916a9a851ee6407d34d2bb4f6f77dbb0" exitCode=0 Jan 21 11:28:11 crc kubenswrapper[4824]: I0121 11:28:11.223668 4824 generic.go:334] "Generic (PLEG): container finished" podID="f3a7f750-590d-43f3-a243-ec2548d1d4d8" containerID="1f864312bcd5d9294e3207dbe32578ac30ec3fffebcb657307393b540937095e" exitCode=143 Jan 21 11:28:11 crc kubenswrapper[4824]: I0121 11:28:11.223536 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f3a7f750-590d-43f3-a243-ec2548d1d4d8","Type":"ContainerDied","Data":"3c697465f3dc9cbfbb17099585b5fcea916a9a851ee6407d34d2bb4f6f77dbb0"} Jan 21 11:28:11 crc kubenswrapper[4824]: I0121 11:28:11.223891 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f3a7f750-590d-43f3-a243-ec2548d1d4d8","Type":"ContainerDied","Data":"1f864312bcd5d9294e3207dbe32578ac30ec3fffebcb657307393b540937095e"} Jan 21 11:28:11 crc kubenswrapper[4824]: I0121 11:28:11.223915 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f3a7f750-590d-43f3-a243-ec2548d1d4d8","Type":"ContainerDied","Data":"d793c49926c6cf8bd7a478c6e30627a0b61508eda4904423a7dd0b8012e4b62b"} Jan 21 11:28:11 crc kubenswrapper[4824]: I0121 11:28:11.223930 4824 scope.go:117] "RemoveContainer" containerID="3c697465f3dc9cbfbb17099585b5fcea916a9a851ee6407d34d2bb4f6f77dbb0" Jan 21 11:28:11 crc kubenswrapper[4824]: I0121 11:28:11.223517 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Jan 21 11:28:11 crc kubenswrapper[4824]: I0121 11:28:11.250516 4824 scope.go:117] "RemoveContainer" containerID="1f864312bcd5d9294e3207dbe32578ac30ec3fffebcb657307393b540937095e" Jan 21 11:28:11 crc kubenswrapper[4824]: I0121 11:28:11.256997 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Jan 21 11:28:11 crc kubenswrapper[4824]: I0121 11:28:11.269043 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Jan 21 11:28:11 crc kubenswrapper[4824]: I0121 11:28:11.274339 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Jan 21 11:28:11 crc kubenswrapper[4824]: E0121 11:28:11.274917 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f3a7f750-590d-43f3-a243-ec2548d1d4d8" containerName="nova-metadata-log" Jan 21 11:28:11 crc kubenswrapper[4824]: I0121 11:28:11.274938 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="f3a7f750-590d-43f3-a243-ec2548d1d4d8" containerName="nova-metadata-log" Jan 21 11:28:11 crc kubenswrapper[4824]: E0121 11:28:11.274973 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f3a7f750-590d-43f3-a243-ec2548d1d4d8" containerName="nova-metadata-metadata" Jan 21 11:28:11 crc kubenswrapper[4824]: I0121 11:28:11.274979 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="f3a7f750-590d-43f3-a243-ec2548d1d4d8" containerName="nova-metadata-metadata" Jan 21 11:28:11 crc kubenswrapper[4824]: I0121 11:28:11.275345 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="f3a7f750-590d-43f3-a243-ec2548d1d4d8" containerName="nova-metadata-metadata" Jan 21 11:28:11 crc kubenswrapper[4824]: I0121 11:28:11.275383 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="f3a7f750-590d-43f3-a243-ec2548d1d4d8" containerName="nova-metadata-log" Jan 21 11:28:11 crc kubenswrapper[4824]: I0121 11:28:11.277324 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Jan 21 11:28:11 crc kubenswrapper[4824]: I0121 11:28:11.282411 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Jan 21 11:28:11 crc kubenswrapper[4824]: I0121 11:28:11.282588 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Jan 21 11:28:11 crc kubenswrapper[4824]: I0121 11:28:11.286450 4824 scope.go:117] "RemoveContainer" containerID="3c697465f3dc9cbfbb17099585b5fcea916a9a851ee6407d34d2bb4f6f77dbb0" Jan 21 11:28:11 crc kubenswrapper[4824]: I0121 11:28:11.286851 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Jan 21 11:28:11 crc kubenswrapper[4824]: E0121 11:28:11.287023 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3c697465f3dc9cbfbb17099585b5fcea916a9a851ee6407d34d2bb4f6f77dbb0\": container with ID starting with 3c697465f3dc9cbfbb17099585b5fcea916a9a851ee6407d34d2bb4f6f77dbb0 not found: ID does not exist" containerID="3c697465f3dc9cbfbb17099585b5fcea916a9a851ee6407d34d2bb4f6f77dbb0" Jan 21 11:28:11 crc kubenswrapper[4824]: I0121 11:28:11.287057 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3c697465f3dc9cbfbb17099585b5fcea916a9a851ee6407d34d2bb4f6f77dbb0"} err="failed to get container status \"3c697465f3dc9cbfbb17099585b5fcea916a9a851ee6407d34d2bb4f6f77dbb0\": rpc error: code = NotFound desc = could not find container \"3c697465f3dc9cbfbb17099585b5fcea916a9a851ee6407d34d2bb4f6f77dbb0\": container with ID starting with 3c697465f3dc9cbfbb17099585b5fcea916a9a851ee6407d34d2bb4f6f77dbb0 not found: ID does not exist" Jan 21 11:28:11 crc kubenswrapper[4824]: I0121 11:28:11.287080 4824 scope.go:117] "RemoveContainer" containerID="1f864312bcd5d9294e3207dbe32578ac30ec3fffebcb657307393b540937095e" Jan 21 11:28:11 crc kubenswrapper[4824]: E0121 11:28:11.290274 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1f864312bcd5d9294e3207dbe32578ac30ec3fffebcb657307393b540937095e\": container with ID starting with 1f864312bcd5d9294e3207dbe32578ac30ec3fffebcb657307393b540937095e not found: ID does not exist" containerID="1f864312bcd5d9294e3207dbe32578ac30ec3fffebcb657307393b540937095e" Jan 21 11:28:11 crc kubenswrapper[4824]: I0121 11:28:11.290299 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1f864312bcd5d9294e3207dbe32578ac30ec3fffebcb657307393b540937095e"} err="failed to get container status \"1f864312bcd5d9294e3207dbe32578ac30ec3fffebcb657307393b540937095e\": rpc error: code = NotFound desc = could not find container \"1f864312bcd5d9294e3207dbe32578ac30ec3fffebcb657307393b540937095e\": container with ID starting with 1f864312bcd5d9294e3207dbe32578ac30ec3fffebcb657307393b540937095e not found: ID does not exist" Jan 21 11:28:11 crc kubenswrapper[4824]: I0121 11:28:11.290315 4824 scope.go:117] "RemoveContainer" containerID="3c697465f3dc9cbfbb17099585b5fcea916a9a851ee6407d34d2bb4f6f77dbb0" Jan 21 11:28:11 crc kubenswrapper[4824]: I0121 11:28:11.291531 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3c697465f3dc9cbfbb17099585b5fcea916a9a851ee6407d34d2bb4f6f77dbb0"} err="failed to get container status \"3c697465f3dc9cbfbb17099585b5fcea916a9a851ee6407d34d2bb4f6f77dbb0\": rpc error: code = NotFound desc = could not find container \"3c697465f3dc9cbfbb17099585b5fcea916a9a851ee6407d34d2bb4f6f77dbb0\": container with ID starting with 3c697465f3dc9cbfbb17099585b5fcea916a9a851ee6407d34d2bb4f6f77dbb0 not found: ID does not exist" Jan 21 11:28:11 crc kubenswrapper[4824]: I0121 11:28:11.291552 4824 scope.go:117] "RemoveContainer" containerID="1f864312bcd5d9294e3207dbe32578ac30ec3fffebcb657307393b540937095e" Jan 21 11:28:11 crc kubenswrapper[4824]: I0121 11:28:11.293336 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1f864312bcd5d9294e3207dbe32578ac30ec3fffebcb657307393b540937095e"} err="failed to get container status \"1f864312bcd5d9294e3207dbe32578ac30ec3fffebcb657307393b540937095e\": rpc error: code = NotFound desc = could not find container \"1f864312bcd5d9294e3207dbe32578ac30ec3fffebcb657307393b540937095e\": container with ID starting with 1f864312bcd5d9294e3207dbe32578ac30ec3fffebcb657307393b540937095e not found: ID does not exist" Jan 21 11:28:11 crc kubenswrapper[4824]: I0121 11:28:11.385983 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9372ff5e-019e-47b0-b2b1-fd3f5cbd16f3-config-data\") pod \"nova-metadata-0\" (UID: \"9372ff5e-019e-47b0-b2b1-fd3f5cbd16f3\") " pod="openstack/nova-metadata-0" Jan 21 11:28:11 crc kubenswrapper[4824]: I0121 11:28:11.386069 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/9372ff5e-019e-47b0-b2b1-fd3f5cbd16f3-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"9372ff5e-019e-47b0-b2b1-fd3f5cbd16f3\") " pod="openstack/nova-metadata-0" Jan 21 11:28:11 crc kubenswrapper[4824]: I0121 11:28:11.386092 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9372ff5e-019e-47b0-b2b1-fd3f5cbd16f3-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"9372ff5e-019e-47b0-b2b1-fd3f5cbd16f3\") " pod="openstack/nova-metadata-0" Jan 21 11:28:11 crc kubenswrapper[4824]: I0121 11:28:11.386117 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9372ff5e-019e-47b0-b2b1-fd3f5cbd16f3-logs\") pod \"nova-metadata-0\" (UID: \"9372ff5e-019e-47b0-b2b1-fd3f5cbd16f3\") " pod="openstack/nova-metadata-0" Jan 21 11:28:11 crc kubenswrapper[4824]: I0121 11:28:11.386444 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m24v8\" (UniqueName: \"kubernetes.io/projected/9372ff5e-019e-47b0-b2b1-fd3f5cbd16f3-kube-api-access-m24v8\") pod \"nova-metadata-0\" (UID: \"9372ff5e-019e-47b0-b2b1-fd3f5cbd16f3\") " pod="openstack/nova-metadata-0" Jan 21 11:28:11 crc kubenswrapper[4824]: I0121 11:28:11.489519 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9372ff5e-019e-47b0-b2b1-fd3f5cbd16f3-logs\") pod \"nova-metadata-0\" (UID: \"9372ff5e-019e-47b0-b2b1-fd3f5cbd16f3\") " pod="openstack/nova-metadata-0" Jan 21 11:28:11 crc kubenswrapper[4824]: I0121 11:28:11.489705 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m24v8\" (UniqueName: \"kubernetes.io/projected/9372ff5e-019e-47b0-b2b1-fd3f5cbd16f3-kube-api-access-m24v8\") pod \"nova-metadata-0\" (UID: \"9372ff5e-019e-47b0-b2b1-fd3f5cbd16f3\") " pod="openstack/nova-metadata-0" Jan 21 11:28:11 crc kubenswrapper[4824]: I0121 11:28:11.489775 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9372ff5e-019e-47b0-b2b1-fd3f5cbd16f3-config-data\") pod \"nova-metadata-0\" (UID: \"9372ff5e-019e-47b0-b2b1-fd3f5cbd16f3\") " pod="openstack/nova-metadata-0" Jan 21 11:28:11 crc kubenswrapper[4824]: I0121 11:28:11.489854 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/9372ff5e-019e-47b0-b2b1-fd3f5cbd16f3-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"9372ff5e-019e-47b0-b2b1-fd3f5cbd16f3\") " pod="openstack/nova-metadata-0" Jan 21 11:28:11 crc kubenswrapper[4824]: I0121 11:28:11.489875 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9372ff5e-019e-47b0-b2b1-fd3f5cbd16f3-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"9372ff5e-019e-47b0-b2b1-fd3f5cbd16f3\") " pod="openstack/nova-metadata-0" Jan 21 11:28:11 crc kubenswrapper[4824]: I0121 11:28:11.489991 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9372ff5e-019e-47b0-b2b1-fd3f5cbd16f3-logs\") pod \"nova-metadata-0\" (UID: \"9372ff5e-019e-47b0-b2b1-fd3f5cbd16f3\") " pod="openstack/nova-metadata-0" Jan 21 11:28:11 crc kubenswrapper[4824]: I0121 11:28:11.493273 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9372ff5e-019e-47b0-b2b1-fd3f5cbd16f3-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"9372ff5e-019e-47b0-b2b1-fd3f5cbd16f3\") " pod="openstack/nova-metadata-0" Jan 21 11:28:11 crc kubenswrapper[4824]: I0121 11:28:11.493734 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/9372ff5e-019e-47b0-b2b1-fd3f5cbd16f3-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"9372ff5e-019e-47b0-b2b1-fd3f5cbd16f3\") " pod="openstack/nova-metadata-0" Jan 21 11:28:11 crc kubenswrapper[4824]: I0121 11:28:11.495251 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9372ff5e-019e-47b0-b2b1-fd3f5cbd16f3-config-data\") pod \"nova-metadata-0\" (UID: \"9372ff5e-019e-47b0-b2b1-fd3f5cbd16f3\") " pod="openstack/nova-metadata-0" Jan 21 11:28:11 crc kubenswrapper[4824]: I0121 11:28:11.504746 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m24v8\" (UniqueName: \"kubernetes.io/projected/9372ff5e-019e-47b0-b2b1-fd3f5cbd16f3-kube-api-access-m24v8\") pod \"nova-metadata-0\" (UID: \"9372ff5e-019e-47b0-b2b1-fd3f5cbd16f3\") " pod="openstack/nova-metadata-0" Jan 21 11:28:11 crc kubenswrapper[4824]: I0121 11:28:11.662110 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Jan 21 11:28:11 crc kubenswrapper[4824]: I0121 11:28:11.667544 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-656b845fb4-qst2s" Jan 21 11:28:11 crc kubenswrapper[4824]: I0121 11:28:11.767021 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 21 11:28:11 crc kubenswrapper[4824]: I0121 11:28:11.794368 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d8d66537-20e1-4f87-9077-484e264bae81-config-data\") pod \"d8d66537-20e1-4f87-9077-484e264bae81\" (UID: \"d8d66537-20e1-4f87-9077-484e264bae81\") " Jan 21 11:28:11 crc kubenswrapper[4824]: I0121 11:28:11.794537 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8d66537-20e1-4f87-9077-484e264bae81-combined-ca-bundle\") pod \"d8d66537-20e1-4f87-9077-484e264bae81\" (UID: \"d8d66537-20e1-4f87-9077-484e264bae81\") " Jan 21 11:28:11 crc kubenswrapper[4824]: I0121 11:28:11.794594 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d8d66537-20e1-4f87-9077-484e264bae81-config-data-custom\") pod \"d8d66537-20e1-4f87-9077-484e264bae81\" (UID: \"d8d66537-20e1-4f87-9077-484e264bae81\") " Jan 21 11:28:11 crc kubenswrapper[4824]: I0121 11:28:11.794623 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rhf6r\" (UniqueName: \"kubernetes.io/projected/d8d66537-20e1-4f87-9077-484e264bae81-kube-api-access-rhf6r\") pod \"d8d66537-20e1-4f87-9077-484e264bae81\" (UID: \"d8d66537-20e1-4f87-9077-484e264bae81\") " Jan 21 11:28:11 crc kubenswrapper[4824]: I0121 11:28:11.798401 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d8d66537-20e1-4f87-9077-484e264bae81-kube-api-access-rhf6r" (OuterVolumeSpecName: "kube-api-access-rhf6r") pod "d8d66537-20e1-4f87-9077-484e264bae81" (UID: "d8d66537-20e1-4f87-9077-484e264bae81"). InnerVolumeSpecName "kube-api-access-rhf6r". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:28:11 crc kubenswrapper[4824]: I0121 11:28:11.799393 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rhf6r\" (UniqueName: \"kubernetes.io/projected/d8d66537-20e1-4f87-9077-484e264bae81-kube-api-access-rhf6r\") on node \"crc\" DevicePath \"\"" Jan 21 11:28:11 crc kubenswrapper[4824]: I0121 11:28:11.807942 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d8d66537-20e1-4f87-9077-484e264bae81-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "d8d66537-20e1-4f87-9077-484e264bae81" (UID: "d8d66537-20e1-4f87-9077-484e264bae81"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:28:11 crc kubenswrapper[4824]: I0121 11:28:11.826596 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d8d66537-20e1-4f87-9077-484e264bae81-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d8d66537-20e1-4f87-9077-484e264bae81" (UID: "d8d66537-20e1-4f87-9077-484e264bae81"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:28:11 crc kubenswrapper[4824]: I0121 11:28:11.840596 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d8d66537-20e1-4f87-9077-484e264bae81-config-data" (OuterVolumeSpecName: "config-data") pod "d8d66537-20e1-4f87-9077-484e264bae81" (UID: "d8d66537-20e1-4f87-9077-484e264bae81"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:28:11 crc kubenswrapper[4824]: I0121 11:28:11.900974 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/deb46ee7-a757-433b-8595-ac21f9d3a03f-run-httpd\") pod \"deb46ee7-a757-433b-8595-ac21f9d3a03f\" (UID: \"deb46ee7-a757-433b-8595-ac21f9d3a03f\") " Jan 21 11:28:11 crc kubenswrapper[4824]: I0121 11:28:11.901231 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/deb46ee7-a757-433b-8595-ac21f9d3a03f-sg-core-conf-yaml\") pod \"deb46ee7-a757-433b-8595-ac21f9d3a03f\" (UID: \"deb46ee7-a757-433b-8595-ac21f9d3a03f\") " Jan 21 11:28:11 crc kubenswrapper[4824]: I0121 11:28:11.901361 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/deb46ee7-a757-433b-8595-ac21f9d3a03f-combined-ca-bundle\") pod \"deb46ee7-a757-433b-8595-ac21f9d3a03f\" (UID: \"deb46ee7-a757-433b-8595-ac21f9d3a03f\") " Jan 21 11:28:11 crc kubenswrapper[4824]: I0121 11:28:11.901421 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/deb46ee7-a757-433b-8595-ac21f9d3a03f-scripts\") pod \"deb46ee7-a757-433b-8595-ac21f9d3a03f\" (UID: \"deb46ee7-a757-433b-8595-ac21f9d3a03f\") " Jan 21 11:28:11 crc kubenswrapper[4824]: I0121 11:28:11.901445 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/deb46ee7-a757-433b-8595-ac21f9d3a03f-log-httpd\") pod \"deb46ee7-a757-433b-8595-ac21f9d3a03f\" (UID: \"deb46ee7-a757-433b-8595-ac21f9d3a03f\") " Jan 21 11:28:11 crc kubenswrapper[4824]: I0121 11:28:11.901357 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/deb46ee7-a757-433b-8595-ac21f9d3a03f-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "deb46ee7-a757-433b-8595-ac21f9d3a03f" (UID: "deb46ee7-a757-433b-8595-ac21f9d3a03f"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:28:11 crc kubenswrapper[4824]: I0121 11:28:11.901597 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/deb46ee7-a757-433b-8595-ac21f9d3a03f-config-data\") pod \"deb46ee7-a757-433b-8595-ac21f9d3a03f\" (UID: \"deb46ee7-a757-433b-8595-ac21f9d3a03f\") " Jan 21 11:28:11 crc kubenswrapper[4824]: I0121 11:28:11.901643 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d25xm\" (UniqueName: \"kubernetes.io/projected/deb46ee7-a757-433b-8595-ac21f9d3a03f-kube-api-access-d25xm\") pod \"deb46ee7-a757-433b-8595-ac21f9d3a03f\" (UID: \"deb46ee7-a757-433b-8595-ac21f9d3a03f\") " Jan 21 11:28:11 crc kubenswrapper[4824]: I0121 11:28:11.901884 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/deb46ee7-a757-433b-8595-ac21f9d3a03f-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "deb46ee7-a757-433b-8595-ac21f9d3a03f" (UID: "deb46ee7-a757-433b-8595-ac21f9d3a03f"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:28:11 crc kubenswrapper[4824]: I0121 11:28:11.902726 4824 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/deb46ee7-a757-433b-8595-ac21f9d3a03f-log-httpd\") on node \"crc\" DevicePath \"\"" Jan 21 11:28:11 crc kubenswrapper[4824]: I0121 11:28:11.902747 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8d66537-20e1-4f87-9077-484e264bae81-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 21 11:28:11 crc kubenswrapper[4824]: I0121 11:28:11.902758 4824 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d8d66537-20e1-4f87-9077-484e264bae81-config-data-custom\") on node \"crc\" DevicePath \"\"" Jan 21 11:28:11 crc kubenswrapper[4824]: I0121 11:28:11.902766 4824 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/deb46ee7-a757-433b-8595-ac21f9d3a03f-run-httpd\") on node \"crc\" DevicePath \"\"" Jan 21 11:28:11 crc kubenswrapper[4824]: I0121 11:28:11.902774 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d8d66537-20e1-4f87-9077-484e264bae81-config-data\") on node \"crc\" DevicePath \"\"" Jan 21 11:28:11 crc kubenswrapper[4824]: I0121 11:28:11.904566 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/deb46ee7-a757-433b-8595-ac21f9d3a03f-scripts" (OuterVolumeSpecName: "scripts") pod "deb46ee7-a757-433b-8595-ac21f9d3a03f" (UID: "deb46ee7-a757-433b-8595-ac21f9d3a03f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:28:11 crc kubenswrapper[4824]: I0121 11:28:11.909472 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/deb46ee7-a757-433b-8595-ac21f9d3a03f-kube-api-access-d25xm" (OuterVolumeSpecName: "kube-api-access-d25xm") pod "deb46ee7-a757-433b-8595-ac21f9d3a03f" (UID: "deb46ee7-a757-433b-8595-ac21f9d3a03f"). InnerVolumeSpecName "kube-api-access-d25xm". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:28:11 crc kubenswrapper[4824]: I0121 11:28:11.927227 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/deb46ee7-a757-433b-8595-ac21f9d3a03f-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "deb46ee7-a757-433b-8595-ac21f9d3a03f" (UID: "deb46ee7-a757-433b-8595-ac21f9d3a03f"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:28:11 crc kubenswrapper[4824]: I0121 11:28:11.961370 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/deb46ee7-a757-433b-8595-ac21f9d3a03f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "deb46ee7-a757-433b-8595-ac21f9d3a03f" (UID: "deb46ee7-a757-433b-8595-ac21f9d3a03f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:28:11 crc kubenswrapper[4824]: I0121 11:28:11.987144 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/deb46ee7-a757-433b-8595-ac21f9d3a03f-config-data" (OuterVolumeSpecName: "config-data") pod "deb46ee7-a757-433b-8595-ac21f9d3a03f" (UID: "deb46ee7-a757-433b-8595-ac21f9d3a03f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:28:12 crc kubenswrapper[4824]: I0121 11:28:12.004238 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/deb46ee7-a757-433b-8595-ac21f9d3a03f-config-data\") on node \"crc\" DevicePath \"\"" Jan 21 11:28:12 crc kubenswrapper[4824]: I0121 11:28:12.004272 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d25xm\" (UniqueName: \"kubernetes.io/projected/deb46ee7-a757-433b-8595-ac21f9d3a03f-kube-api-access-d25xm\") on node \"crc\" DevicePath \"\"" Jan 21 11:28:12 crc kubenswrapper[4824]: I0121 11:28:12.004286 4824 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/deb46ee7-a757-433b-8595-ac21f9d3a03f-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Jan 21 11:28:12 crc kubenswrapper[4824]: I0121 11:28:12.004295 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/deb46ee7-a757-433b-8595-ac21f9d3a03f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 21 11:28:12 crc kubenswrapper[4824]: I0121 11:28:12.004303 4824 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/deb46ee7-a757-433b-8595-ac21f9d3a03f-scripts\") on node \"crc\" DevicePath \"\"" Jan 21 11:28:12 crc kubenswrapper[4824]: I0121 11:28:12.057873 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f3a7f750-590d-43f3-a243-ec2548d1d4d8" path="/var/lib/kubelet/pods/f3a7f750-590d-43f3-a243-ec2548d1d4d8/volumes" Jan 21 11:28:12 crc kubenswrapper[4824]: I0121 11:28:12.086428 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Jan 21 11:28:12 crc kubenswrapper[4824]: W0121 11:28:12.092658 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9372ff5e_019e_47b0_b2b1_fd3f5cbd16f3.slice/crio-3b22de5e8cf949114c3b9b5d7ecfe3c1b6ca7ce93636a1e49cddaaa4c2711814 WatchSource:0}: Error finding container 3b22de5e8cf949114c3b9b5d7ecfe3c1b6ca7ce93636a1e49cddaaa4c2711814: Status 404 returned error can't find the container with id 3b22de5e8cf949114c3b9b5d7ecfe3c1b6ca7ce93636a1e49cddaaa4c2711814 Jan 21 11:28:12 crc kubenswrapper[4824]: I0121 11:28:12.232531 4824 generic.go:334] "Generic (PLEG): container finished" podID="baa34da9-3711-4cd4-a5eb-ee5814ac51a7" containerID="dd76c88e84c244299a1fbb85cad5296bd9ab946000d0d268303768f60e22c9c4" exitCode=0 Jan 21 11:28:12 crc kubenswrapper[4824]: I0121 11:28:12.232595 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-g9vq5" event={"ID":"baa34da9-3711-4cd4-a5eb-ee5814ac51a7","Type":"ContainerDied","Data":"dd76c88e84c244299a1fbb85cad5296bd9ab946000d0d268303768f60e22c9c4"} Jan 21 11:28:12 crc kubenswrapper[4824]: I0121 11:28:12.235107 4824 generic.go:334] "Generic (PLEG): container finished" podID="deb46ee7-a757-433b-8595-ac21f9d3a03f" containerID="b2dbd538cfc7628790de49627fe1faff1d0f54a393e99dc66963d7845a74077e" exitCode=0 Jan 21 11:28:12 crc kubenswrapper[4824]: I0121 11:28:12.235143 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"deb46ee7-a757-433b-8595-ac21f9d3a03f","Type":"ContainerDied","Data":"b2dbd538cfc7628790de49627fe1faff1d0f54a393e99dc66963d7845a74077e"} Jan 21 11:28:12 crc kubenswrapper[4824]: I0121 11:28:12.235179 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"deb46ee7-a757-433b-8595-ac21f9d3a03f","Type":"ContainerDied","Data":"bac5f30d5f4848de2bccab1913e20b18c8655df8b33e352d91fc9c00699884ca"} Jan 21 11:28:12 crc kubenswrapper[4824]: I0121 11:28:12.235197 4824 scope.go:117] "RemoveContainer" containerID="70559b4f16a38253804eec43f7ada16a4c4b14e98a6505d8ed5b38a8a1ac6f82" Jan 21 11:28:12 crc kubenswrapper[4824]: I0121 11:28:12.235203 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 21 11:28:12 crc kubenswrapper[4824]: I0121 11:28:12.239561 4824 generic.go:334] "Generic (PLEG): container finished" podID="d8d66537-20e1-4f87-9077-484e264bae81" containerID="05efb240856120694f40b5baf0ccbaf59da69510aa66da3d1ac87094a30bb9ea" exitCode=0 Jan 21 11:28:12 crc kubenswrapper[4824]: I0121 11:28:12.239742 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-656b845fb4-qst2s" Jan 21 11:28:12 crc kubenswrapper[4824]: I0121 11:28:12.239762 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-656b845fb4-qst2s" event={"ID":"d8d66537-20e1-4f87-9077-484e264bae81","Type":"ContainerDied","Data":"05efb240856120694f40b5baf0ccbaf59da69510aa66da3d1ac87094a30bb9ea"} Jan 21 11:28:12 crc kubenswrapper[4824]: I0121 11:28:12.239833 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-656b845fb4-qst2s" event={"ID":"d8d66537-20e1-4f87-9077-484e264bae81","Type":"ContainerDied","Data":"6515b7c77086397376de7620cef207d703b735c2b9d2c10ad1793aedbc3eb3b1"} Jan 21 11:28:12 crc kubenswrapper[4824]: I0121 11:28:12.243077 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"9372ff5e-019e-47b0-b2b1-fd3f5cbd16f3","Type":"ContainerStarted","Data":"3b32312f5047bb2f4067be047f24a0224f9cb4437dee6fe43810999d5a0e1540"} Jan 21 11:28:12 crc kubenswrapper[4824]: I0121 11:28:12.243103 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"9372ff5e-019e-47b0-b2b1-fd3f5cbd16f3","Type":"ContainerStarted","Data":"3b22de5e8cf949114c3b9b5d7ecfe3c1b6ca7ce93636a1e49cddaaa4c2711814"} Jan 21 11:28:12 crc kubenswrapper[4824]: I0121 11:28:12.259688 4824 scope.go:117] "RemoveContainer" containerID="64748c9826d4f022251fc061ac66e44525e85e4e0e8ee1dc2d4a0b3f328468c3" Jan 21 11:28:12 crc kubenswrapper[4824]: I0121 11:28:12.264511 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-engine-656b845fb4-qst2s"] Jan 21 11:28:12 crc kubenswrapper[4824]: I0121 11:28:12.271270 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-engine-656b845fb4-qst2s"] Jan 21 11:28:12 crc kubenswrapper[4824]: I0121 11:28:12.281580 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Jan 21 11:28:12 crc kubenswrapper[4824]: I0121 11:28:12.289133 4824 scope.go:117] "RemoveContainer" containerID="03674de70b88a6dae6f9a3d4e9d1399fcaa1ecb7056d38626af5e46c5a1612f6" Jan 21 11:28:12 crc kubenswrapper[4824]: I0121 11:28:12.289506 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Jan 21 11:28:12 crc kubenswrapper[4824]: I0121 11:28:12.297919 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Jan 21 11:28:12 crc kubenswrapper[4824]: E0121 11:28:12.298254 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="deb46ee7-a757-433b-8595-ac21f9d3a03f" containerName="ceilometer-notification-agent" Jan 21 11:28:12 crc kubenswrapper[4824]: I0121 11:28:12.298271 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="deb46ee7-a757-433b-8595-ac21f9d3a03f" containerName="ceilometer-notification-agent" Jan 21 11:28:12 crc kubenswrapper[4824]: E0121 11:28:12.298288 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="deb46ee7-a757-433b-8595-ac21f9d3a03f" containerName="proxy-httpd" Jan 21 11:28:12 crc kubenswrapper[4824]: I0121 11:28:12.298294 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="deb46ee7-a757-433b-8595-ac21f9d3a03f" containerName="proxy-httpd" Jan 21 11:28:12 crc kubenswrapper[4824]: E0121 11:28:12.298312 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="deb46ee7-a757-433b-8595-ac21f9d3a03f" containerName="ceilometer-central-agent" Jan 21 11:28:12 crc kubenswrapper[4824]: I0121 11:28:12.298318 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="deb46ee7-a757-433b-8595-ac21f9d3a03f" containerName="ceilometer-central-agent" Jan 21 11:28:12 crc kubenswrapper[4824]: E0121 11:28:12.298325 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="deb46ee7-a757-433b-8595-ac21f9d3a03f" containerName="sg-core" Jan 21 11:28:12 crc kubenswrapper[4824]: I0121 11:28:12.298330 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="deb46ee7-a757-433b-8595-ac21f9d3a03f" containerName="sg-core" Jan 21 11:28:12 crc kubenswrapper[4824]: E0121 11:28:12.298343 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d8d66537-20e1-4f87-9077-484e264bae81" containerName="heat-engine" Jan 21 11:28:12 crc kubenswrapper[4824]: I0121 11:28:12.298349 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="d8d66537-20e1-4f87-9077-484e264bae81" containerName="heat-engine" Jan 21 11:28:12 crc kubenswrapper[4824]: I0121 11:28:12.298535 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="d8d66537-20e1-4f87-9077-484e264bae81" containerName="heat-engine" Jan 21 11:28:12 crc kubenswrapper[4824]: I0121 11:28:12.298550 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="deb46ee7-a757-433b-8595-ac21f9d3a03f" containerName="ceilometer-notification-agent" Jan 21 11:28:12 crc kubenswrapper[4824]: I0121 11:28:12.298557 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="deb46ee7-a757-433b-8595-ac21f9d3a03f" containerName="proxy-httpd" Jan 21 11:28:12 crc kubenswrapper[4824]: I0121 11:28:12.298570 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="deb46ee7-a757-433b-8595-ac21f9d3a03f" containerName="sg-core" Jan 21 11:28:12 crc kubenswrapper[4824]: I0121 11:28:12.298579 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="deb46ee7-a757-433b-8595-ac21f9d3a03f" containerName="ceilometer-central-agent" Jan 21 11:28:12 crc kubenswrapper[4824]: I0121 11:28:12.299993 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 21 11:28:12 crc kubenswrapper[4824]: I0121 11:28:12.304941 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Jan 21 11:28:12 crc kubenswrapper[4824]: I0121 11:28:12.305032 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Jan 21 11:28:12 crc kubenswrapper[4824]: I0121 11:28:12.336113 4824 scope.go:117] "RemoveContainer" containerID="b2dbd538cfc7628790de49627fe1faff1d0f54a393e99dc66963d7845a74077e" Jan 21 11:28:12 crc kubenswrapper[4824]: I0121 11:28:12.345350 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Jan 21 11:28:12 crc kubenswrapper[4824]: I0121 11:28:12.362700 4824 scope.go:117] "RemoveContainer" containerID="70559b4f16a38253804eec43f7ada16a4c4b14e98a6505d8ed5b38a8a1ac6f82" Jan 21 11:28:12 crc kubenswrapper[4824]: E0121 11:28:12.363117 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"70559b4f16a38253804eec43f7ada16a4c4b14e98a6505d8ed5b38a8a1ac6f82\": container with ID starting with 70559b4f16a38253804eec43f7ada16a4c4b14e98a6505d8ed5b38a8a1ac6f82 not found: ID does not exist" containerID="70559b4f16a38253804eec43f7ada16a4c4b14e98a6505d8ed5b38a8a1ac6f82" Jan 21 11:28:12 crc kubenswrapper[4824]: I0121 11:28:12.363154 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"70559b4f16a38253804eec43f7ada16a4c4b14e98a6505d8ed5b38a8a1ac6f82"} err="failed to get container status \"70559b4f16a38253804eec43f7ada16a4c4b14e98a6505d8ed5b38a8a1ac6f82\": rpc error: code = NotFound desc = could not find container \"70559b4f16a38253804eec43f7ada16a4c4b14e98a6505d8ed5b38a8a1ac6f82\": container with ID starting with 70559b4f16a38253804eec43f7ada16a4c4b14e98a6505d8ed5b38a8a1ac6f82 not found: ID does not exist" Jan 21 11:28:12 crc kubenswrapper[4824]: I0121 11:28:12.363180 4824 scope.go:117] "RemoveContainer" containerID="64748c9826d4f022251fc061ac66e44525e85e4e0e8ee1dc2d4a0b3f328468c3" Jan 21 11:28:12 crc kubenswrapper[4824]: E0121 11:28:12.363481 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"64748c9826d4f022251fc061ac66e44525e85e4e0e8ee1dc2d4a0b3f328468c3\": container with ID starting with 64748c9826d4f022251fc061ac66e44525e85e4e0e8ee1dc2d4a0b3f328468c3 not found: ID does not exist" containerID="64748c9826d4f022251fc061ac66e44525e85e4e0e8ee1dc2d4a0b3f328468c3" Jan 21 11:28:12 crc kubenswrapper[4824]: I0121 11:28:12.363505 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"64748c9826d4f022251fc061ac66e44525e85e4e0e8ee1dc2d4a0b3f328468c3"} err="failed to get container status \"64748c9826d4f022251fc061ac66e44525e85e4e0e8ee1dc2d4a0b3f328468c3\": rpc error: code = NotFound desc = could not find container \"64748c9826d4f022251fc061ac66e44525e85e4e0e8ee1dc2d4a0b3f328468c3\": container with ID starting with 64748c9826d4f022251fc061ac66e44525e85e4e0e8ee1dc2d4a0b3f328468c3 not found: ID does not exist" Jan 21 11:28:12 crc kubenswrapper[4824]: I0121 11:28:12.363521 4824 scope.go:117] "RemoveContainer" containerID="03674de70b88a6dae6f9a3d4e9d1399fcaa1ecb7056d38626af5e46c5a1612f6" Jan 21 11:28:12 crc kubenswrapper[4824]: E0121 11:28:12.364852 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"03674de70b88a6dae6f9a3d4e9d1399fcaa1ecb7056d38626af5e46c5a1612f6\": container with ID starting with 03674de70b88a6dae6f9a3d4e9d1399fcaa1ecb7056d38626af5e46c5a1612f6 not found: ID does not exist" containerID="03674de70b88a6dae6f9a3d4e9d1399fcaa1ecb7056d38626af5e46c5a1612f6" Jan 21 11:28:12 crc kubenswrapper[4824]: I0121 11:28:12.364888 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"03674de70b88a6dae6f9a3d4e9d1399fcaa1ecb7056d38626af5e46c5a1612f6"} err="failed to get container status \"03674de70b88a6dae6f9a3d4e9d1399fcaa1ecb7056d38626af5e46c5a1612f6\": rpc error: code = NotFound desc = could not find container \"03674de70b88a6dae6f9a3d4e9d1399fcaa1ecb7056d38626af5e46c5a1612f6\": container with ID starting with 03674de70b88a6dae6f9a3d4e9d1399fcaa1ecb7056d38626af5e46c5a1612f6 not found: ID does not exist" Jan 21 11:28:12 crc kubenswrapper[4824]: I0121 11:28:12.364909 4824 scope.go:117] "RemoveContainer" containerID="b2dbd538cfc7628790de49627fe1faff1d0f54a393e99dc66963d7845a74077e" Jan 21 11:28:12 crc kubenswrapper[4824]: E0121 11:28:12.365175 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b2dbd538cfc7628790de49627fe1faff1d0f54a393e99dc66963d7845a74077e\": container with ID starting with b2dbd538cfc7628790de49627fe1faff1d0f54a393e99dc66963d7845a74077e not found: ID does not exist" containerID="b2dbd538cfc7628790de49627fe1faff1d0f54a393e99dc66963d7845a74077e" Jan 21 11:28:12 crc kubenswrapper[4824]: I0121 11:28:12.365202 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b2dbd538cfc7628790de49627fe1faff1d0f54a393e99dc66963d7845a74077e"} err="failed to get container status \"b2dbd538cfc7628790de49627fe1faff1d0f54a393e99dc66963d7845a74077e\": rpc error: code = NotFound desc = could not find container \"b2dbd538cfc7628790de49627fe1faff1d0f54a393e99dc66963d7845a74077e\": container with ID starting with b2dbd538cfc7628790de49627fe1faff1d0f54a393e99dc66963d7845a74077e not found: ID does not exist" Jan 21 11:28:12 crc kubenswrapper[4824]: I0121 11:28:12.365217 4824 scope.go:117] "RemoveContainer" containerID="05efb240856120694f40b5baf0ccbaf59da69510aa66da3d1ac87094a30bb9ea" Jan 21 11:28:12 crc kubenswrapper[4824]: I0121 11:28:12.382246 4824 scope.go:117] "RemoveContainer" containerID="05efb240856120694f40b5baf0ccbaf59da69510aa66da3d1ac87094a30bb9ea" Jan 21 11:28:12 crc kubenswrapper[4824]: E0121 11:28:12.382534 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"05efb240856120694f40b5baf0ccbaf59da69510aa66da3d1ac87094a30bb9ea\": container with ID starting with 05efb240856120694f40b5baf0ccbaf59da69510aa66da3d1ac87094a30bb9ea not found: ID does not exist" containerID="05efb240856120694f40b5baf0ccbaf59da69510aa66da3d1ac87094a30bb9ea" Jan 21 11:28:12 crc kubenswrapper[4824]: I0121 11:28:12.382564 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"05efb240856120694f40b5baf0ccbaf59da69510aa66da3d1ac87094a30bb9ea"} err="failed to get container status \"05efb240856120694f40b5baf0ccbaf59da69510aa66da3d1ac87094a30bb9ea\": rpc error: code = NotFound desc = could not find container \"05efb240856120694f40b5baf0ccbaf59da69510aa66da3d1ac87094a30bb9ea\": container with ID starting with 05efb240856120694f40b5baf0ccbaf59da69510aa66da3d1ac87094a30bb9ea not found: ID does not exist" Jan 21 11:28:12 crc kubenswrapper[4824]: I0121 11:28:12.419861 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/42d7c353-4ca7-4fa0-b4da-4793ce3e36a1-config-data\") pod \"ceilometer-0\" (UID: \"42d7c353-4ca7-4fa0-b4da-4793ce3e36a1\") " pod="openstack/ceilometer-0" Jan 21 11:28:12 crc kubenswrapper[4824]: I0121 11:28:12.419902 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/42d7c353-4ca7-4fa0-b4da-4793ce3e36a1-run-httpd\") pod \"ceilometer-0\" (UID: \"42d7c353-4ca7-4fa0-b4da-4793ce3e36a1\") " pod="openstack/ceilometer-0" Jan 21 11:28:12 crc kubenswrapper[4824]: I0121 11:28:12.419988 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42d7c353-4ca7-4fa0-b4da-4793ce3e36a1-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"42d7c353-4ca7-4fa0-b4da-4793ce3e36a1\") " pod="openstack/ceilometer-0" Jan 21 11:28:12 crc kubenswrapper[4824]: I0121 11:28:12.420026 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/42d7c353-4ca7-4fa0-b4da-4793ce3e36a1-log-httpd\") pod \"ceilometer-0\" (UID: \"42d7c353-4ca7-4fa0-b4da-4793ce3e36a1\") " pod="openstack/ceilometer-0" Jan 21 11:28:12 crc kubenswrapper[4824]: I0121 11:28:12.420063 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dmqz4\" (UniqueName: \"kubernetes.io/projected/42d7c353-4ca7-4fa0-b4da-4793ce3e36a1-kube-api-access-dmqz4\") pod \"ceilometer-0\" (UID: \"42d7c353-4ca7-4fa0-b4da-4793ce3e36a1\") " pod="openstack/ceilometer-0" Jan 21 11:28:12 crc kubenswrapper[4824]: I0121 11:28:12.420261 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/42d7c353-4ca7-4fa0-b4da-4793ce3e36a1-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"42d7c353-4ca7-4fa0-b4da-4793ce3e36a1\") " pod="openstack/ceilometer-0" Jan 21 11:28:12 crc kubenswrapper[4824]: I0121 11:28:12.420310 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/42d7c353-4ca7-4fa0-b4da-4793ce3e36a1-scripts\") pod \"ceilometer-0\" (UID: \"42d7c353-4ca7-4fa0-b4da-4793ce3e36a1\") " pod="openstack/ceilometer-0" Jan 21 11:28:12 crc kubenswrapper[4824]: I0121 11:28:12.522338 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42d7c353-4ca7-4fa0-b4da-4793ce3e36a1-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"42d7c353-4ca7-4fa0-b4da-4793ce3e36a1\") " pod="openstack/ceilometer-0" Jan 21 11:28:12 crc kubenswrapper[4824]: I0121 11:28:12.522599 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/42d7c353-4ca7-4fa0-b4da-4793ce3e36a1-log-httpd\") pod \"ceilometer-0\" (UID: \"42d7c353-4ca7-4fa0-b4da-4793ce3e36a1\") " pod="openstack/ceilometer-0" Jan 21 11:28:12 crc kubenswrapper[4824]: I0121 11:28:12.522725 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dmqz4\" (UniqueName: \"kubernetes.io/projected/42d7c353-4ca7-4fa0-b4da-4793ce3e36a1-kube-api-access-dmqz4\") pod \"ceilometer-0\" (UID: \"42d7c353-4ca7-4fa0-b4da-4793ce3e36a1\") " pod="openstack/ceilometer-0" Jan 21 11:28:12 crc kubenswrapper[4824]: I0121 11:28:12.522937 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/42d7c353-4ca7-4fa0-b4da-4793ce3e36a1-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"42d7c353-4ca7-4fa0-b4da-4793ce3e36a1\") " pod="openstack/ceilometer-0" Jan 21 11:28:12 crc kubenswrapper[4824]: I0121 11:28:12.523032 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/42d7c353-4ca7-4fa0-b4da-4793ce3e36a1-scripts\") pod \"ceilometer-0\" (UID: \"42d7c353-4ca7-4fa0-b4da-4793ce3e36a1\") " pod="openstack/ceilometer-0" Jan 21 11:28:12 crc kubenswrapper[4824]: I0121 11:28:12.523035 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/42d7c353-4ca7-4fa0-b4da-4793ce3e36a1-log-httpd\") pod \"ceilometer-0\" (UID: \"42d7c353-4ca7-4fa0-b4da-4793ce3e36a1\") " pod="openstack/ceilometer-0" Jan 21 11:28:12 crc kubenswrapper[4824]: I0121 11:28:12.523314 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/42d7c353-4ca7-4fa0-b4da-4793ce3e36a1-config-data\") pod \"ceilometer-0\" (UID: \"42d7c353-4ca7-4fa0-b4da-4793ce3e36a1\") " pod="openstack/ceilometer-0" Jan 21 11:28:12 crc kubenswrapper[4824]: I0121 11:28:12.523403 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/42d7c353-4ca7-4fa0-b4da-4793ce3e36a1-run-httpd\") pod \"ceilometer-0\" (UID: \"42d7c353-4ca7-4fa0-b4da-4793ce3e36a1\") " pod="openstack/ceilometer-0" Jan 21 11:28:12 crc kubenswrapper[4824]: I0121 11:28:12.523692 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/42d7c353-4ca7-4fa0-b4da-4793ce3e36a1-run-httpd\") pod \"ceilometer-0\" (UID: \"42d7c353-4ca7-4fa0-b4da-4793ce3e36a1\") " pod="openstack/ceilometer-0" Jan 21 11:28:12 crc kubenswrapper[4824]: I0121 11:28:12.526469 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/42d7c353-4ca7-4fa0-b4da-4793ce3e36a1-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"42d7c353-4ca7-4fa0-b4da-4793ce3e36a1\") " pod="openstack/ceilometer-0" Jan 21 11:28:12 crc kubenswrapper[4824]: I0121 11:28:12.526580 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/42d7c353-4ca7-4fa0-b4da-4793ce3e36a1-scripts\") pod \"ceilometer-0\" (UID: \"42d7c353-4ca7-4fa0-b4da-4793ce3e36a1\") " pod="openstack/ceilometer-0" Jan 21 11:28:12 crc kubenswrapper[4824]: I0121 11:28:12.526834 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/42d7c353-4ca7-4fa0-b4da-4793ce3e36a1-config-data\") pod \"ceilometer-0\" (UID: \"42d7c353-4ca7-4fa0-b4da-4793ce3e36a1\") " pod="openstack/ceilometer-0" Jan 21 11:28:12 crc kubenswrapper[4824]: I0121 11:28:12.526993 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42d7c353-4ca7-4fa0-b4da-4793ce3e36a1-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"42d7c353-4ca7-4fa0-b4da-4793ce3e36a1\") " pod="openstack/ceilometer-0" Jan 21 11:28:12 crc kubenswrapper[4824]: I0121 11:28:12.538617 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dmqz4\" (UniqueName: \"kubernetes.io/projected/42d7c353-4ca7-4fa0-b4da-4793ce3e36a1-kube-api-access-dmqz4\") pod \"ceilometer-0\" (UID: \"42d7c353-4ca7-4fa0-b4da-4793ce3e36a1\") " pod="openstack/ceilometer-0" Jan 21 11:28:12 crc kubenswrapper[4824]: I0121 11:28:12.628777 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 21 11:28:12 crc kubenswrapper[4824]: I0121 11:28:12.995172 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Jan 21 11:28:13 crc kubenswrapper[4824]: I0121 11:28:13.253906 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"9372ff5e-019e-47b0-b2b1-fd3f5cbd16f3","Type":"ContainerStarted","Data":"7db077deeddf26df005d9b8e08dcd67b955383aa1afae0547bd5d1b88562a3ff"} Jan 21 11:28:13 crc kubenswrapper[4824]: I0121 11:28:13.256585 4824 generic.go:334] "Generic (PLEG): container finished" podID="135a8573-b55a-4c5f-9cb2-a7c3adea9720" containerID="e6d7bd7f99afa7be3d7eee7ba3147eedee78587b76d733c8cb546689d1b785fe" exitCode=0 Jan 21 11:28:13 crc kubenswrapper[4824]: I0121 11:28:13.256658 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-7gtcg" event={"ID":"135a8573-b55a-4c5f-9cb2-a7c3adea9720","Type":"ContainerDied","Data":"e6d7bd7f99afa7be3d7eee7ba3147eedee78587b76d733c8cb546689d1b785fe"} Jan 21 11:28:13 crc kubenswrapper[4824]: I0121 11:28:13.259264 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"42d7c353-4ca7-4fa0-b4da-4793ce3e36a1","Type":"ContainerStarted","Data":"3debb3f0581abba1b7346533099c933cc5a8bb94ead0c540a6932f4be37b2b39"} Jan 21 11:28:13 crc kubenswrapper[4824]: I0121 11:28:13.292827 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.292813951 podStartE2EDuration="2.292813951s" podCreationTimestamp="2026-01-21 11:28:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:28:13.288741209 +0000 UTC m=+1035.581770502" watchObservedRunningTime="2026-01-21 11:28:13.292813951 +0000 UTC m=+1035.585843244" Jan 21 11:28:13 crc kubenswrapper[4824]: I0121 11:28:13.698331 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-g9vq5" Jan 21 11:28:13 crc kubenswrapper[4824]: I0121 11:28:13.744249 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/baa34da9-3711-4cd4-a5eb-ee5814ac51a7-config-data\") pod \"baa34da9-3711-4cd4-a5eb-ee5814ac51a7\" (UID: \"baa34da9-3711-4cd4-a5eb-ee5814ac51a7\") " Jan 21 11:28:13 crc kubenswrapper[4824]: I0121 11:28:13.744286 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2xnkr\" (UniqueName: \"kubernetes.io/projected/baa34da9-3711-4cd4-a5eb-ee5814ac51a7-kube-api-access-2xnkr\") pod \"baa34da9-3711-4cd4-a5eb-ee5814ac51a7\" (UID: \"baa34da9-3711-4cd4-a5eb-ee5814ac51a7\") " Jan 21 11:28:13 crc kubenswrapper[4824]: I0121 11:28:13.744340 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/baa34da9-3711-4cd4-a5eb-ee5814ac51a7-scripts\") pod \"baa34da9-3711-4cd4-a5eb-ee5814ac51a7\" (UID: \"baa34da9-3711-4cd4-a5eb-ee5814ac51a7\") " Jan 21 11:28:13 crc kubenswrapper[4824]: I0121 11:28:13.744444 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/baa34da9-3711-4cd4-a5eb-ee5814ac51a7-combined-ca-bundle\") pod \"baa34da9-3711-4cd4-a5eb-ee5814ac51a7\" (UID: \"baa34da9-3711-4cd4-a5eb-ee5814ac51a7\") " Jan 21 11:28:13 crc kubenswrapper[4824]: I0121 11:28:13.747373 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/baa34da9-3711-4cd4-a5eb-ee5814ac51a7-scripts" (OuterVolumeSpecName: "scripts") pod "baa34da9-3711-4cd4-a5eb-ee5814ac51a7" (UID: "baa34da9-3711-4cd4-a5eb-ee5814ac51a7"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:28:13 crc kubenswrapper[4824]: I0121 11:28:13.747639 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/baa34da9-3711-4cd4-a5eb-ee5814ac51a7-kube-api-access-2xnkr" (OuterVolumeSpecName: "kube-api-access-2xnkr") pod "baa34da9-3711-4cd4-a5eb-ee5814ac51a7" (UID: "baa34da9-3711-4cd4-a5eb-ee5814ac51a7"). InnerVolumeSpecName "kube-api-access-2xnkr". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:28:13 crc kubenswrapper[4824]: I0121 11:28:13.763997 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/baa34da9-3711-4cd4-a5eb-ee5814ac51a7-config-data" (OuterVolumeSpecName: "config-data") pod "baa34da9-3711-4cd4-a5eb-ee5814ac51a7" (UID: "baa34da9-3711-4cd4-a5eb-ee5814ac51a7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:28:13 crc kubenswrapper[4824]: I0121 11:28:13.765193 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/baa34da9-3711-4cd4-a5eb-ee5814ac51a7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "baa34da9-3711-4cd4-a5eb-ee5814ac51a7" (UID: "baa34da9-3711-4cd4-a5eb-ee5814ac51a7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:28:13 crc kubenswrapper[4824]: I0121 11:28:13.846611 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/baa34da9-3711-4cd4-a5eb-ee5814ac51a7-config-data\") on node \"crc\" DevicePath \"\"" Jan 21 11:28:13 crc kubenswrapper[4824]: I0121 11:28:13.846849 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2xnkr\" (UniqueName: \"kubernetes.io/projected/baa34da9-3711-4cd4-a5eb-ee5814ac51a7-kube-api-access-2xnkr\") on node \"crc\" DevicePath \"\"" Jan 21 11:28:13 crc kubenswrapper[4824]: I0121 11:28:13.846861 4824 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/baa34da9-3711-4cd4-a5eb-ee5814ac51a7-scripts\") on node \"crc\" DevicePath \"\"" Jan 21 11:28:13 crc kubenswrapper[4824]: I0121 11:28:13.846871 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/baa34da9-3711-4cd4-a5eb-ee5814ac51a7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 21 11:28:14 crc kubenswrapper[4824]: I0121 11:28:14.057426 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d8d66537-20e1-4f87-9077-484e264bae81" path="/var/lib/kubelet/pods/d8d66537-20e1-4f87-9077-484e264bae81/volumes" Jan 21 11:28:14 crc kubenswrapper[4824]: I0121 11:28:14.057992 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="deb46ee7-a757-433b-8595-ac21f9d3a03f" path="/var/lib/kubelet/pods/deb46ee7-a757-433b-8595-ac21f9d3a03f/volumes" Jan 21 11:28:14 crc kubenswrapper[4824]: I0121 11:28:14.273436 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"42d7c353-4ca7-4fa0-b4da-4793ce3e36a1","Type":"ContainerStarted","Data":"82b46e3d25bb50cb39acd7bbf34eb02ad735f71fcba98f606eba2559159f21c7"} Jan 21 11:28:14 crc kubenswrapper[4824]: I0121 11:28:14.274745 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-g9vq5" event={"ID":"baa34da9-3711-4cd4-a5eb-ee5814ac51a7","Type":"ContainerDied","Data":"b818724198e29357d179e92d306ffc141219f83ff62bb7dd6d3881b003b27078"} Jan 21 11:28:14 crc kubenswrapper[4824]: I0121 11:28:14.274779 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b818724198e29357d179e92d306ffc141219f83ff62bb7dd6d3881b003b27078" Jan 21 11:28:14 crc kubenswrapper[4824]: I0121 11:28:14.274780 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-g9vq5" Jan 21 11:28:14 crc kubenswrapper[4824]: I0121 11:28:14.414579 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Jan 21 11:28:14 crc kubenswrapper[4824]: I0121 11:28:14.414826 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="9934f06c-6f15-4ca5-95f1-4c5bb27e72e8" containerName="nova-api-api" containerID="cri-o://2dc23f89a2ecb53e2543111bc5e25eb8fa216de246339731c0b2bddd341062d7" gracePeriod=30 Jan 21 11:28:14 crc kubenswrapper[4824]: I0121 11:28:14.414778 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="9934f06c-6f15-4ca5-95f1-4c5bb27e72e8" containerName="nova-api-log" containerID="cri-o://79798a638b185d61504b6917c46892e6c77377d55b27c826aa0f1424cea017bf" gracePeriod=30 Jan 21 11:28:14 crc kubenswrapper[4824]: I0121 11:28:14.427045 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Jan 21 11:28:14 crc kubenswrapper[4824]: I0121 11:28:14.433239 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="4a414d51-aaa8-43a6-88f5-f3ca59eb87a6" containerName="nova-scheduler-scheduler" containerID="cri-o://e871bc06d477443b88ec205254bfc60975ceca19ed6da8e295be1de5bf3e986e" gracePeriod=30 Jan 21 11:28:14 crc kubenswrapper[4824]: I0121 11:28:14.437952 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Jan 21 11:28:14 crc kubenswrapper[4824]: I0121 11:28:14.575913 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Jan 21 11:28:14 crc kubenswrapper[4824]: I0121 11:28:14.576434 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-7gtcg" Jan 21 11:28:14 crc kubenswrapper[4824]: I0121 11:28:14.662971 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/135a8573-b55a-4c5f-9cb2-a7c3adea9720-config-data\") pod \"135a8573-b55a-4c5f-9cb2-a7c3adea9720\" (UID: \"135a8573-b55a-4c5f-9cb2-a7c3adea9720\") " Jan 21 11:28:14 crc kubenswrapper[4824]: I0121 11:28:14.663059 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/135a8573-b55a-4c5f-9cb2-a7c3adea9720-combined-ca-bundle\") pod \"135a8573-b55a-4c5f-9cb2-a7c3adea9720\" (UID: \"135a8573-b55a-4c5f-9cb2-a7c3adea9720\") " Jan 21 11:28:14 crc kubenswrapper[4824]: I0121 11:28:14.663083 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vv6bv\" (UniqueName: \"kubernetes.io/projected/135a8573-b55a-4c5f-9cb2-a7c3adea9720-kube-api-access-vv6bv\") pod \"135a8573-b55a-4c5f-9cb2-a7c3adea9720\" (UID: \"135a8573-b55a-4c5f-9cb2-a7c3adea9720\") " Jan 21 11:28:14 crc kubenswrapper[4824]: I0121 11:28:14.663220 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/135a8573-b55a-4c5f-9cb2-a7c3adea9720-scripts\") pod \"135a8573-b55a-4c5f-9cb2-a7c3adea9720\" (UID: \"135a8573-b55a-4c5f-9cb2-a7c3adea9720\") " Jan 21 11:28:14 crc kubenswrapper[4824]: I0121 11:28:14.668172 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/135a8573-b55a-4c5f-9cb2-a7c3adea9720-scripts" (OuterVolumeSpecName: "scripts") pod "135a8573-b55a-4c5f-9cb2-a7c3adea9720" (UID: "135a8573-b55a-4c5f-9cb2-a7c3adea9720"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:28:14 crc kubenswrapper[4824]: I0121 11:28:14.669124 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/135a8573-b55a-4c5f-9cb2-a7c3adea9720-kube-api-access-vv6bv" (OuterVolumeSpecName: "kube-api-access-vv6bv") pod "135a8573-b55a-4c5f-9cb2-a7c3adea9720" (UID: "135a8573-b55a-4c5f-9cb2-a7c3adea9720"). InnerVolumeSpecName "kube-api-access-vv6bv". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:28:14 crc kubenswrapper[4824]: I0121 11:28:14.675117 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Jan 21 11:28:14 crc kubenswrapper[4824]: I0121 11:28:14.687724 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/135a8573-b55a-4c5f-9cb2-a7c3adea9720-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "135a8573-b55a-4c5f-9cb2-a7c3adea9720" (UID: "135a8573-b55a-4c5f-9cb2-a7c3adea9720"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:28:14 crc kubenswrapper[4824]: I0121 11:28:14.693026 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/135a8573-b55a-4c5f-9cb2-a7c3adea9720-config-data" (OuterVolumeSpecName: "config-data") pod "135a8573-b55a-4c5f-9cb2-a7c3adea9720" (UID: "135a8573-b55a-4c5f-9cb2-a7c3adea9720"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:28:14 crc kubenswrapper[4824]: I0121 11:28:14.765623 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/135a8573-b55a-4c5f-9cb2-a7c3adea9720-config-data\") on node \"crc\" DevicePath \"\"" Jan 21 11:28:14 crc kubenswrapper[4824]: I0121 11:28:14.765651 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/135a8573-b55a-4c5f-9cb2-a7c3adea9720-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 21 11:28:14 crc kubenswrapper[4824]: I0121 11:28:14.765661 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vv6bv\" (UniqueName: \"kubernetes.io/projected/135a8573-b55a-4c5f-9cb2-a7c3adea9720-kube-api-access-vv6bv\") on node \"crc\" DevicePath \"\"" Jan 21 11:28:14 crc kubenswrapper[4824]: I0121 11:28:14.765671 4824 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/135a8573-b55a-4c5f-9cb2-a7c3adea9720-scripts\") on node \"crc\" DevicePath \"\"" Jan 21 11:28:14 crc kubenswrapper[4824]: I0121 11:28:14.829118 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-9b86998b5-lwrg9" Jan 21 11:28:14 crc kubenswrapper[4824]: I0121 11:28:14.880321 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7756b9d78c-ftf8g"] Jan 21 11:28:14 crc kubenswrapper[4824]: I0121 11:28:14.880501 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7756b9d78c-ftf8g" podUID="f2ec0128-5881-4c67-8d78-d53d8be3a2d3" containerName="dnsmasq-dns" containerID="cri-o://18927e4fff1f2c120dbfd5a267daf57c4c84dbc311e5fd297d4153df2a6f21f4" gracePeriod=10 Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:14.997581 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.070523 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9934f06c-6f15-4ca5-95f1-4c5bb27e72e8-combined-ca-bundle\") pod \"9934f06c-6f15-4ca5-95f1-4c5bb27e72e8\" (UID: \"9934f06c-6f15-4ca5-95f1-4c5bb27e72e8\") " Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.070637 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9934f06c-6f15-4ca5-95f1-4c5bb27e72e8-logs\") pod \"9934f06c-6f15-4ca5-95f1-4c5bb27e72e8\" (UID: \"9934f06c-6f15-4ca5-95f1-4c5bb27e72e8\") " Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.070659 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9934f06c-6f15-4ca5-95f1-4c5bb27e72e8-config-data\") pod \"9934f06c-6f15-4ca5-95f1-4c5bb27e72e8\" (UID: \"9934f06c-6f15-4ca5-95f1-4c5bb27e72e8\") " Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.070767 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjvlf\" (UniqueName: \"kubernetes.io/projected/9934f06c-6f15-4ca5-95f1-4c5bb27e72e8-kube-api-access-pjvlf\") pod \"9934f06c-6f15-4ca5-95f1-4c5bb27e72e8\" (UID: \"9934f06c-6f15-4ca5-95f1-4c5bb27e72e8\") " Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.071319 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9934f06c-6f15-4ca5-95f1-4c5bb27e72e8-logs" (OuterVolumeSpecName: "logs") pod "9934f06c-6f15-4ca5-95f1-4c5bb27e72e8" (UID: "9934f06c-6f15-4ca5-95f1-4c5bb27e72e8"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.078097 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9934f06c-6f15-4ca5-95f1-4c5bb27e72e8-kube-api-access-pjvlf" (OuterVolumeSpecName: "kube-api-access-pjvlf") pod "9934f06c-6f15-4ca5-95f1-4c5bb27e72e8" (UID: "9934f06c-6f15-4ca5-95f1-4c5bb27e72e8"). InnerVolumeSpecName "kube-api-access-pjvlf". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.111075 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9934f06c-6f15-4ca5-95f1-4c5bb27e72e8-config-data" (OuterVolumeSpecName: "config-data") pod "9934f06c-6f15-4ca5-95f1-4c5bb27e72e8" (UID: "9934f06c-6f15-4ca5-95f1-4c5bb27e72e8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.127012 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9934f06c-6f15-4ca5-95f1-4c5bb27e72e8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9934f06c-6f15-4ca5-95f1-4c5bb27e72e8" (UID: "9934f06c-6f15-4ca5-95f1-4c5bb27e72e8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.174300 4824 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9934f06c-6f15-4ca5-95f1-4c5bb27e72e8-logs\") on node \"crc\" DevicePath \"\"" Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.174324 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9934f06c-6f15-4ca5-95f1-4c5bb27e72e8-config-data\") on node \"crc\" DevicePath \"\"" Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.174335 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjvlf\" (UniqueName: \"kubernetes.io/projected/9934f06c-6f15-4ca5-95f1-4c5bb27e72e8-kube-api-access-pjvlf\") on node \"crc\" DevicePath \"\"" Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.174352 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9934f06c-6f15-4ca5-95f1-4c5bb27e72e8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.291223 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-7gtcg" event={"ID":"135a8573-b55a-4c5f-9cb2-a7c3adea9720","Type":"ContainerDied","Data":"74afad19ccfca1492a879f61489b7e643f3472e59cfcfff81bc9271163b2f5df"} Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.291261 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="74afad19ccfca1492a879f61489b7e643f3472e59cfcfff81bc9271163b2f5df" Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.291279 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-7gtcg" Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.293578 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"42d7c353-4ca7-4fa0-b4da-4793ce3e36a1","Type":"ContainerStarted","Data":"87edd33f33351d88d9cbd6b17c925a8893b337952321dc91c1ba77e5c174ae3c"} Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.293623 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"42d7c353-4ca7-4fa0-b4da-4793ce3e36a1","Type":"ContainerStarted","Data":"97c437ef91be90c480bbbc66844b2f1bf63fc92b3d5efc241a8ff7639e0f322c"} Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.297110 4824 generic.go:334] "Generic (PLEG): container finished" podID="f2ec0128-5881-4c67-8d78-d53d8be3a2d3" containerID="18927e4fff1f2c120dbfd5a267daf57c4c84dbc311e5fd297d4153df2a6f21f4" exitCode=0 Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.297137 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7756b9d78c-ftf8g" event={"ID":"f2ec0128-5881-4c67-8d78-d53d8be3a2d3","Type":"ContainerDied","Data":"18927e4fff1f2c120dbfd5a267daf57c4c84dbc311e5fd297d4153df2a6f21f4"} Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.299188 4824 generic.go:334] "Generic (PLEG): container finished" podID="9934f06c-6f15-4ca5-95f1-4c5bb27e72e8" containerID="2dc23f89a2ecb53e2543111bc5e25eb8fa216de246339731c0b2bddd341062d7" exitCode=0 Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.299217 4824 generic.go:334] "Generic (PLEG): container finished" podID="9934f06c-6f15-4ca5-95f1-4c5bb27e72e8" containerID="79798a638b185d61504b6917c46892e6c77377d55b27c826aa0f1424cea017bf" exitCode=143 Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.299357 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="9372ff5e-019e-47b0-b2b1-fd3f5cbd16f3" containerName="nova-metadata-log" containerID="cri-o://3b32312f5047bb2f4067be047f24a0224f9cb4437dee6fe43810999d5a0e1540" gracePeriod=30 Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.299649 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.300141 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"9934f06c-6f15-4ca5-95f1-4c5bb27e72e8","Type":"ContainerDied","Data":"2dc23f89a2ecb53e2543111bc5e25eb8fa216de246339731c0b2bddd341062d7"} Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.300189 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"9934f06c-6f15-4ca5-95f1-4c5bb27e72e8","Type":"ContainerDied","Data":"79798a638b185d61504b6917c46892e6c77377d55b27c826aa0f1424cea017bf"} Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.300202 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"9934f06c-6f15-4ca5-95f1-4c5bb27e72e8","Type":"ContainerDied","Data":"2c15246588a81cb1f14c8e607d836b282630b2c662037ff7639bacf65e5fa2ae"} Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.300222 4824 scope.go:117] "RemoveContainer" containerID="2dc23f89a2ecb53e2543111bc5e25eb8fa216de246339731c0b2bddd341062d7" Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.300714 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="9372ff5e-019e-47b0-b2b1-fd3f5cbd16f3" containerName="nova-metadata-metadata" containerID="cri-o://7db077deeddf26df005d9b8e08dcd67b955383aa1afae0547bd5d1b88562a3ff" gracePeriod=30 Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.324943 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7756b9d78c-ftf8g" Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.339082 4824 scope.go:117] "RemoveContainer" containerID="79798a638b185d61504b6917c46892e6c77377d55b27c826aa0f1424cea017bf" Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.351097 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.363844 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.374045 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Jan 21 11:28:15 crc kubenswrapper[4824]: E0121 11:28:15.374406 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="baa34da9-3711-4cd4-a5eb-ee5814ac51a7" containerName="nova-manage" Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.374424 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="baa34da9-3711-4cd4-a5eb-ee5814ac51a7" containerName="nova-manage" Jan 21 11:28:15 crc kubenswrapper[4824]: E0121 11:28:15.374445 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f2ec0128-5881-4c67-8d78-d53d8be3a2d3" containerName="init" Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.374451 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="f2ec0128-5881-4c67-8d78-d53d8be3a2d3" containerName="init" Jan 21 11:28:15 crc kubenswrapper[4824]: E0121 11:28:15.374459 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="135a8573-b55a-4c5f-9cb2-a7c3adea9720" containerName="nova-cell1-conductor-db-sync" Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.374465 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="135a8573-b55a-4c5f-9cb2-a7c3adea9720" containerName="nova-cell1-conductor-db-sync" Jan 21 11:28:15 crc kubenswrapper[4824]: E0121 11:28:15.374479 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9934f06c-6f15-4ca5-95f1-4c5bb27e72e8" containerName="nova-api-api" Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.374485 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="9934f06c-6f15-4ca5-95f1-4c5bb27e72e8" containerName="nova-api-api" Jan 21 11:28:15 crc kubenswrapper[4824]: E0121 11:28:15.374507 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f2ec0128-5881-4c67-8d78-d53d8be3a2d3" containerName="dnsmasq-dns" Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.374512 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="f2ec0128-5881-4c67-8d78-d53d8be3a2d3" containerName="dnsmasq-dns" Jan 21 11:28:15 crc kubenswrapper[4824]: E0121 11:28:15.374524 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9934f06c-6f15-4ca5-95f1-4c5bb27e72e8" containerName="nova-api-log" Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.374529 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="9934f06c-6f15-4ca5-95f1-4c5bb27e72e8" containerName="nova-api-log" Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.374706 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="9934f06c-6f15-4ca5-95f1-4c5bb27e72e8" containerName="nova-api-api" Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.374720 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="baa34da9-3711-4cd4-a5eb-ee5814ac51a7" containerName="nova-manage" Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.374734 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="9934f06c-6f15-4ca5-95f1-4c5bb27e72e8" containerName="nova-api-log" Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.374743 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="135a8573-b55a-4c5f-9cb2-a7c3adea9720" containerName="nova-cell1-conductor-db-sync" Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.374752 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="f2ec0128-5881-4c67-8d78-d53d8be3a2d3" containerName="dnsmasq-dns" Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.375601 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.380244 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f2ec0128-5881-4c67-8d78-d53d8be3a2d3-ovsdbserver-sb\") pod \"f2ec0128-5881-4c67-8d78-d53d8be3a2d3\" (UID: \"f2ec0128-5881-4c67-8d78-d53d8be3a2d3\") " Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.380314 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f2ec0128-5881-4c67-8d78-d53d8be3a2d3-ovsdbserver-nb\") pod \"f2ec0128-5881-4c67-8d78-d53d8be3a2d3\" (UID: \"f2ec0128-5881-4c67-8d78-d53d8be3a2d3\") " Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.380464 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f2ec0128-5881-4c67-8d78-d53d8be3a2d3-dns-svc\") pod \"f2ec0128-5881-4c67-8d78-d53d8be3a2d3\" (UID: \"f2ec0128-5881-4c67-8d78-d53d8be3a2d3\") " Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.380500 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f2ec0128-5881-4c67-8d78-d53d8be3a2d3-dns-swift-storage-0\") pod \"f2ec0128-5881-4c67-8d78-d53d8be3a2d3\" (UID: \"f2ec0128-5881-4c67-8d78-d53d8be3a2d3\") " Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.380530 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f2ec0128-5881-4c67-8d78-d53d8be3a2d3-config\") pod \"f2ec0128-5881-4c67-8d78-d53d8be3a2d3\" (UID: \"f2ec0128-5881-4c67-8d78-d53d8be3a2d3\") " Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.380558 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hljdx\" (UniqueName: \"kubernetes.io/projected/f2ec0128-5881-4c67-8d78-d53d8be3a2d3-kube-api-access-hljdx\") pod \"f2ec0128-5881-4c67-8d78-d53d8be3a2d3\" (UID: \"f2ec0128-5881-4c67-8d78-d53d8be3a2d3\") " Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.385409 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.392137 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f2ec0128-5881-4c67-8d78-d53d8be3a2d3-kube-api-access-hljdx" (OuterVolumeSpecName: "kube-api-access-hljdx") pod "f2ec0128-5881-4c67-8d78-d53d8be3a2d3" (UID: "f2ec0128-5881-4c67-8d78-d53d8be3a2d3"). InnerVolumeSpecName "kube-api-access-hljdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.398456 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.437532 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.439192 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.441676 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f2ec0128-5881-4c67-8d78-d53d8be3a2d3-config" (OuterVolumeSpecName: "config") pod "f2ec0128-5881-4c67-8d78-d53d8be3a2d3" (UID: "f2ec0128-5881-4c67-8d78-d53d8be3a2d3"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.442263 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f2ec0128-5881-4c67-8d78-d53d8be3a2d3-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "f2ec0128-5881-4c67-8d78-d53d8be3a2d3" (UID: "f2ec0128-5881-4c67-8d78-d53d8be3a2d3"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.443901 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.446932 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f2ec0128-5881-4c67-8d78-d53d8be3a2d3-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "f2ec0128-5881-4c67-8d78-d53d8be3a2d3" (UID: "f2ec0128-5881-4c67-8d78-d53d8be3a2d3"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.459312 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f2ec0128-5881-4c67-8d78-d53d8be3a2d3-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "f2ec0128-5881-4c67-8d78-d53d8be3a2d3" (UID: "f2ec0128-5881-4c67-8d78-d53d8be3a2d3"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.463917 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.464548 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f2ec0128-5881-4c67-8d78-d53d8be3a2d3-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "f2ec0128-5881-4c67-8d78-d53d8be3a2d3" (UID: "f2ec0128-5881-4c67-8d78-d53d8be3a2d3"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.486031 4824 scope.go:117] "RemoveContainer" containerID="2dc23f89a2ecb53e2543111bc5e25eb8fa216de246339731c0b2bddd341062d7" Jan 21 11:28:15 crc kubenswrapper[4824]: E0121 11:28:15.486510 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2dc23f89a2ecb53e2543111bc5e25eb8fa216de246339731c0b2bddd341062d7\": container with ID starting with 2dc23f89a2ecb53e2543111bc5e25eb8fa216de246339731c0b2bddd341062d7 not found: ID does not exist" containerID="2dc23f89a2ecb53e2543111bc5e25eb8fa216de246339731c0b2bddd341062d7" Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.486543 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2dc23f89a2ecb53e2543111bc5e25eb8fa216de246339731c0b2bddd341062d7"} err="failed to get container status \"2dc23f89a2ecb53e2543111bc5e25eb8fa216de246339731c0b2bddd341062d7\": rpc error: code = NotFound desc = could not find container \"2dc23f89a2ecb53e2543111bc5e25eb8fa216de246339731c0b2bddd341062d7\": container with ID starting with 2dc23f89a2ecb53e2543111bc5e25eb8fa216de246339731c0b2bddd341062d7 not found: ID does not exist" Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.486567 4824 scope.go:117] "RemoveContainer" containerID="79798a638b185d61504b6917c46892e6c77377d55b27c826aa0f1424cea017bf" Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.486637 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/75cf0b57-bdd7-4b26-8d52-5acfbe9d520d-config-data\") pod \"nova-api-0\" (UID: \"75cf0b57-bdd7-4b26-8d52-5acfbe9d520d\") " pod="openstack/nova-api-0" Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.486834 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/75cf0b57-bdd7-4b26-8d52-5acfbe9d520d-logs\") pod \"nova-api-0\" (UID: \"75cf0b57-bdd7-4b26-8d52-5acfbe9d520d\") " pod="openstack/nova-api-0" Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.486915 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/75cf0b57-bdd7-4b26-8d52-5acfbe9d520d-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"75cf0b57-bdd7-4b26-8d52-5acfbe9d520d\") " pod="openstack/nova-api-0" Jan 21 11:28:15 crc kubenswrapper[4824]: E0121 11:28:15.487053 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"79798a638b185d61504b6917c46892e6c77377d55b27c826aa0f1424cea017bf\": container with ID starting with 79798a638b185d61504b6917c46892e6c77377d55b27c826aa0f1424cea017bf not found: ID does not exist" containerID="79798a638b185d61504b6917c46892e6c77377d55b27c826aa0f1424cea017bf" Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.487144 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n5q8r\" (UniqueName: \"kubernetes.io/projected/75cf0b57-bdd7-4b26-8d52-5acfbe9d520d-kube-api-access-n5q8r\") pod \"nova-api-0\" (UID: \"75cf0b57-bdd7-4b26-8d52-5acfbe9d520d\") " pod="openstack/nova-api-0" Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.487132 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"79798a638b185d61504b6917c46892e6c77377d55b27c826aa0f1424cea017bf"} err="failed to get container status \"79798a638b185d61504b6917c46892e6c77377d55b27c826aa0f1424cea017bf\": rpc error: code = NotFound desc = could not find container \"79798a638b185d61504b6917c46892e6c77377d55b27c826aa0f1424cea017bf\": container with ID starting with 79798a638b185d61504b6917c46892e6c77377d55b27c826aa0f1424cea017bf not found: ID does not exist" Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.487302 4824 scope.go:117] "RemoveContainer" containerID="2dc23f89a2ecb53e2543111bc5e25eb8fa216de246339731c0b2bddd341062d7" Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.487262 4824 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f2ec0128-5881-4c67-8d78-d53d8be3a2d3-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.487378 4824 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f2ec0128-5881-4c67-8d78-d53d8be3a2d3-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.487389 4824 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f2ec0128-5881-4c67-8d78-d53d8be3a2d3-dns-svc\") on node \"crc\" DevicePath \"\"" Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.487398 4824 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f2ec0128-5881-4c67-8d78-d53d8be3a2d3-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.487408 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f2ec0128-5881-4c67-8d78-d53d8be3a2d3-config\") on node \"crc\" DevicePath \"\"" Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.487415 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hljdx\" (UniqueName: \"kubernetes.io/projected/f2ec0128-5881-4c67-8d78-d53d8be3a2d3-kube-api-access-hljdx\") on node \"crc\" DevicePath \"\"" Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.487774 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2dc23f89a2ecb53e2543111bc5e25eb8fa216de246339731c0b2bddd341062d7"} err="failed to get container status \"2dc23f89a2ecb53e2543111bc5e25eb8fa216de246339731c0b2bddd341062d7\": rpc error: code = NotFound desc = could not find container \"2dc23f89a2ecb53e2543111bc5e25eb8fa216de246339731c0b2bddd341062d7\": container with ID starting with 2dc23f89a2ecb53e2543111bc5e25eb8fa216de246339731c0b2bddd341062d7 not found: ID does not exist" Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.487822 4824 scope.go:117] "RemoveContainer" containerID="79798a638b185d61504b6917c46892e6c77377d55b27c826aa0f1424cea017bf" Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.488232 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"79798a638b185d61504b6917c46892e6c77377d55b27c826aa0f1424cea017bf"} err="failed to get container status \"79798a638b185d61504b6917c46892e6c77377d55b27c826aa0f1424cea017bf\": rpc error: code = NotFound desc = could not find container \"79798a638b185d61504b6917c46892e6c77377d55b27c826aa0f1424cea017bf\": container with ID starting with 79798a638b185d61504b6917c46892e6c77377d55b27c826aa0f1424cea017bf not found: ID does not exist" Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.588768 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n5q8r\" (UniqueName: \"kubernetes.io/projected/75cf0b57-bdd7-4b26-8d52-5acfbe9d520d-kube-api-access-n5q8r\") pod \"nova-api-0\" (UID: \"75cf0b57-bdd7-4b26-8d52-5acfbe9d520d\") " pod="openstack/nova-api-0" Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.589047 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c398d1e9-13e1-400c-a609-bceea49aea51-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"c398d1e9-13e1-400c-a609-bceea49aea51\") " pod="openstack/nova-cell1-conductor-0" Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.589402 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c398d1e9-13e1-400c-a609-bceea49aea51-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"c398d1e9-13e1-400c-a609-bceea49aea51\") " pod="openstack/nova-cell1-conductor-0" Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.589504 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/75cf0b57-bdd7-4b26-8d52-5acfbe9d520d-config-data\") pod \"nova-api-0\" (UID: \"75cf0b57-bdd7-4b26-8d52-5acfbe9d520d\") " pod="openstack/nova-api-0" Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.589588 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mc5qr\" (UniqueName: \"kubernetes.io/projected/c398d1e9-13e1-400c-a609-bceea49aea51-kube-api-access-mc5qr\") pod \"nova-cell1-conductor-0\" (UID: \"c398d1e9-13e1-400c-a609-bceea49aea51\") " pod="openstack/nova-cell1-conductor-0" Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.589719 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/75cf0b57-bdd7-4b26-8d52-5acfbe9d520d-logs\") pod \"nova-api-0\" (UID: \"75cf0b57-bdd7-4b26-8d52-5acfbe9d520d\") " pod="openstack/nova-api-0" Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.589830 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/75cf0b57-bdd7-4b26-8d52-5acfbe9d520d-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"75cf0b57-bdd7-4b26-8d52-5acfbe9d520d\") " pod="openstack/nova-api-0" Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.590766 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/75cf0b57-bdd7-4b26-8d52-5acfbe9d520d-logs\") pod \"nova-api-0\" (UID: \"75cf0b57-bdd7-4b26-8d52-5acfbe9d520d\") " pod="openstack/nova-api-0" Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.592830 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/75cf0b57-bdd7-4b26-8d52-5acfbe9d520d-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"75cf0b57-bdd7-4b26-8d52-5acfbe9d520d\") " pod="openstack/nova-api-0" Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.593573 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/75cf0b57-bdd7-4b26-8d52-5acfbe9d520d-config-data\") pod \"nova-api-0\" (UID: \"75cf0b57-bdd7-4b26-8d52-5acfbe9d520d\") " pod="openstack/nova-api-0" Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.604677 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n5q8r\" (UniqueName: \"kubernetes.io/projected/75cf0b57-bdd7-4b26-8d52-5acfbe9d520d-kube-api-access-n5q8r\") pod \"nova-api-0\" (UID: \"75cf0b57-bdd7-4b26-8d52-5acfbe9d520d\") " pod="openstack/nova-api-0" Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.691911 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c398d1e9-13e1-400c-a609-bceea49aea51-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"c398d1e9-13e1-400c-a609-bceea49aea51\") " pod="openstack/nova-cell1-conductor-0" Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.692153 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c398d1e9-13e1-400c-a609-bceea49aea51-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"c398d1e9-13e1-400c-a609-bceea49aea51\") " pod="openstack/nova-cell1-conductor-0" Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.692327 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mc5qr\" (UniqueName: \"kubernetes.io/projected/c398d1e9-13e1-400c-a609-bceea49aea51-kube-api-access-mc5qr\") pod \"nova-cell1-conductor-0\" (UID: \"c398d1e9-13e1-400c-a609-bceea49aea51\") " pod="openstack/nova-cell1-conductor-0" Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.694920 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c398d1e9-13e1-400c-a609-bceea49aea51-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"c398d1e9-13e1-400c-a609-bceea49aea51\") " pod="openstack/nova-cell1-conductor-0" Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.695017 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c398d1e9-13e1-400c-a609-bceea49aea51-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"c398d1e9-13e1-400c-a609-bceea49aea51\") " pod="openstack/nova-cell1-conductor-0" Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.695335 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.712002 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mc5qr\" (UniqueName: \"kubernetes.io/projected/c398d1e9-13e1-400c-a609-bceea49aea51-kube-api-access-mc5qr\") pod \"nova-cell1-conductor-0\" (UID: \"c398d1e9-13e1-400c-a609-bceea49aea51\") " pod="openstack/nova-cell1-conductor-0" Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.762805 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Jan 21 11:28:15 crc kubenswrapper[4824]: I0121 11:28:15.937813 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Jan 21 11:28:16 crc kubenswrapper[4824]: I0121 11:28:16.057506 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9934f06c-6f15-4ca5-95f1-4c5bb27e72e8" path="/var/lib/kubelet/pods/9934f06c-6f15-4ca5-95f1-4c5bb27e72e8/volumes" Jan 21 11:28:16 crc kubenswrapper[4824]: I0121 11:28:16.301375 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Jan 21 11:28:16 crc kubenswrapper[4824]: I0121 11:28:16.311395 4824 generic.go:334] "Generic (PLEG): container finished" podID="4a414d51-aaa8-43a6-88f5-f3ca59eb87a6" containerID="e871bc06d477443b88ec205254bfc60975ceca19ed6da8e295be1de5bf3e986e" exitCode=0 Jan 21 11:28:16 crc kubenswrapper[4824]: I0121 11:28:16.311474 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"4a414d51-aaa8-43a6-88f5-f3ca59eb87a6","Type":"ContainerDied","Data":"e871bc06d477443b88ec205254bfc60975ceca19ed6da8e295be1de5bf3e986e"} Jan 21 11:28:16 crc kubenswrapper[4824]: I0121 11:28:16.313083 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"75cf0b57-bdd7-4b26-8d52-5acfbe9d520d","Type":"ContainerStarted","Data":"7d8c20082288cf600e39dc7a0f1bd5ac8ba09165ca35559ccd6592d1b0add249"} Jan 21 11:28:16 crc kubenswrapper[4824]: I0121 11:28:16.313130 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"75cf0b57-bdd7-4b26-8d52-5acfbe9d520d","Type":"ContainerStarted","Data":"728845c124e990e3b5128ca1bd8dd69e1a52e3b9f1c5598a8738ea82598401a9"} Jan 21 11:28:16 crc kubenswrapper[4824]: I0121 11:28:16.320223 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Jan 21 11:28:16 crc kubenswrapper[4824]: I0121 11:28:16.320279 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"9372ff5e-019e-47b0-b2b1-fd3f5cbd16f3","Type":"ContainerDied","Data":"7db077deeddf26df005d9b8e08dcd67b955383aa1afae0547bd5d1b88562a3ff"} Jan 21 11:28:16 crc kubenswrapper[4824]: I0121 11:28:16.320304 4824 generic.go:334] "Generic (PLEG): container finished" podID="9372ff5e-019e-47b0-b2b1-fd3f5cbd16f3" containerID="7db077deeddf26df005d9b8e08dcd67b955383aa1afae0547bd5d1b88562a3ff" exitCode=0 Jan 21 11:28:16 crc kubenswrapper[4824]: I0121 11:28:16.320337 4824 scope.go:117] "RemoveContainer" containerID="7db077deeddf26df005d9b8e08dcd67b955383aa1afae0547bd5d1b88562a3ff" Jan 21 11:28:16 crc kubenswrapper[4824]: I0121 11:28:16.320350 4824 generic.go:334] "Generic (PLEG): container finished" podID="9372ff5e-019e-47b0-b2b1-fd3f5cbd16f3" containerID="3b32312f5047bb2f4067be047f24a0224f9cb4437dee6fe43810999d5a0e1540" exitCode=143 Jan 21 11:28:16 crc kubenswrapper[4824]: I0121 11:28:16.320422 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"9372ff5e-019e-47b0-b2b1-fd3f5cbd16f3","Type":"ContainerDied","Data":"3b32312f5047bb2f4067be047f24a0224f9cb4437dee6fe43810999d5a0e1540"} Jan 21 11:28:16 crc kubenswrapper[4824]: I0121 11:28:16.320445 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"9372ff5e-019e-47b0-b2b1-fd3f5cbd16f3","Type":"ContainerDied","Data":"3b22de5e8cf949114c3b9b5d7ecfe3c1b6ca7ce93636a1e49cddaaa4c2711814"} Jan 21 11:28:16 crc kubenswrapper[4824]: I0121 11:28:16.325947 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7756b9d78c-ftf8g" event={"ID":"f2ec0128-5881-4c67-8d78-d53d8be3a2d3","Type":"ContainerDied","Data":"76a9ebfca2b2e541f5b20e22a77a9c3eda4db32f3628a468b328bf434a816a25"} Jan 21 11:28:16 crc kubenswrapper[4824]: I0121 11:28:16.325987 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7756b9d78c-ftf8g" Jan 21 11:28:16 crc kubenswrapper[4824]: I0121 11:28:16.362932 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7756b9d78c-ftf8g"] Jan 21 11:28:16 crc kubenswrapper[4824]: I0121 11:28:16.373168 4824 scope.go:117] "RemoveContainer" containerID="3b32312f5047bb2f4067be047f24a0224f9cb4437dee6fe43810999d5a0e1540" Jan 21 11:28:16 crc kubenswrapper[4824]: I0121 11:28:16.373263 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7756b9d78c-ftf8g"] Jan 21 11:28:16 crc kubenswrapper[4824]: I0121 11:28:16.383558 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Jan 21 11:28:16 crc kubenswrapper[4824]: I0121 11:28:16.403278 4824 scope.go:117] "RemoveContainer" containerID="7db077deeddf26df005d9b8e08dcd67b955383aa1afae0547bd5d1b88562a3ff" Jan 21 11:28:16 crc kubenswrapper[4824]: E0121 11:28:16.404386 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7db077deeddf26df005d9b8e08dcd67b955383aa1afae0547bd5d1b88562a3ff\": container with ID starting with 7db077deeddf26df005d9b8e08dcd67b955383aa1afae0547bd5d1b88562a3ff not found: ID does not exist" containerID="7db077deeddf26df005d9b8e08dcd67b955383aa1afae0547bd5d1b88562a3ff" Jan 21 11:28:16 crc kubenswrapper[4824]: I0121 11:28:16.404437 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7db077deeddf26df005d9b8e08dcd67b955383aa1afae0547bd5d1b88562a3ff"} err="failed to get container status \"7db077deeddf26df005d9b8e08dcd67b955383aa1afae0547bd5d1b88562a3ff\": rpc error: code = NotFound desc = could not find container \"7db077deeddf26df005d9b8e08dcd67b955383aa1afae0547bd5d1b88562a3ff\": container with ID starting with 7db077deeddf26df005d9b8e08dcd67b955383aa1afae0547bd5d1b88562a3ff not found: ID does not exist" Jan 21 11:28:16 crc kubenswrapper[4824]: I0121 11:28:16.404458 4824 scope.go:117] "RemoveContainer" containerID="3b32312f5047bb2f4067be047f24a0224f9cb4437dee6fe43810999d5a0e1540" Jan 21 11:28:16 crc kubenswrapper[4824]: E0121 11:28:16.406069 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3b32312f5047bb2f4067be047f24a0224f9cb4437dee6fe43810999d5a0e1540\": container with ID starting with 3b32312f5047bb2f4067be047f24a0224f9cb4437dee6fe43810999d5a0e1540 not found: ID does not exist" containerID="3b32312f5047bb2f4067be047f24a0224f9cb4437dee6fe43810999d5a0e1540" Jan 21 11:28:16 crc kubenswrapper[4824]: I0121 11:28:16.406097 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3b32312f5047bb2f4067be047f24a0224f9cb4437dee6fe43810999d5a0e1540"} err="failed to get container status \"3b32312f5047bb2f4067be047f24a0224f9cb4437dee6fe43810999d5a0e1540\": rpc error: code = NotFound desc = could not find container \"3b32312f5047bb2f4067be047f24a0224f9cb4437dee6fe43810999d5a0e1540\": container with ID starting with 3b32312f5047bb2f4067be047f24a0224f9cb4437dee6fe43810999d5a0e1540 not found: ID does not exist" Jan 21 11:28:16 crc kubenswrapper[4824]: I0121 11:28:16.406123 4824 scope.go:117] "RemoveContainer" containerID="7db077deeddf26df005d9b8e08dcd67b955383aa1afae0547bd5d1b88562a3ff" Jan 21 11:28:16 crc kubenswrapper[4824]: I0121 11:28:16.406512 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7db077deeddf26df005d9b8e08dcd67b955383aa1afae0547bd5d1b88562a3ff"} err="failed to get container status \"7db077deeddf26df005d9b8e08dcd67b955383aa1afae0547bd5d1b88562a3ff\": rpc error: code = NotFound desc = could not find container \"7db077deeddf26df005d9b8e08dcd67b955383aa1afae0547bd5d1b88562a3ff\": container with ID starting with 7db077deeddf26df005d9b8e08dcd67b955383aa1afae0547bd5d1b88562a3ff not found: ID does not exist" Jan 21 11:28:16 crc kubenswrapper[4824]: I0121 11:28:16.406535 4824 scope.go:117] "RemoveContainer" containerID="3b32312f5047bb2f4067be047f24a0224f9cb4437dee6fe43810999d5a0e1540" Jan 21 11:28:16 crc kubenswrapper[4824]: I0121 11:28:16.407387 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3b32312f5047bb2f4067be047f24a0224f9cb4437dee6fe43810999d5a0e1540"} err="failed to get container status \"3b32312f5047bb2f4067be047f24a0224f9cb4437dee6fe43810999d5a0e1540\": rpc error: code = NotFound desc = could not find container \"3b32312f5047bb2f4067be047f24a0224f9cb4437dee6fe43810999d5a0e1540\": container with ID starting with 3b32312f5047bb2f4067be047f24a0224f9cb4437dee6fe43810999d5a0e1540 not found: ID does not exist" Jan 21 11:28:16 crc kubenswrapper[4824]: I0121 11:28:16.407421 4824 scope.go:117] "RemoveContainer" containerID="18927e4fff1f2c120dbfd5a267daf57c4c84dbc311e5fd297d4153df2a6f21f4" Jan 21 11:28:16 crc kubenswrapper[4824]: I0121 11:28:16.409079 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9372ff5e-019e-47b0-b2b1-fd3f5cbd16f3-config-data\") pod \"9372ff5e-019e-47b0-b2b1-fd3f5cbd16f3\" (UID: \"9372ff5e-019e-47b0-b2b1-fd3f5cbd16f3\") " Jan 21 11:28:16 crc kubenswrapper[4824]: I0121 11:28:16.409143 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9372ff5e-019e-47b0-b2b1-fd3f5cbd16f3-combined-ca-bundle\") pod \"9372ff5e-019e-47b0-b2b1-fd3f5cbd16f3\" (UID: \"9372ff5e-019e-47b0-b2b1-fd3f5cbd16f3\") " Jan 21 11:28:16 crc kubenswrapper[4824]: I0121 11:28:16.409217 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m24v8\" (UniqueName: \"kubernetes.io/projected/9372ff5e-019e-47b0-b2b1-fd3f5cbd16f3-kube-api-access-m24v8\") pod \"9372ff5e-019e-47b0-b2b1-fd3f5cbd16f3\" (UID: \"9372ff5e-019e-47b0-b2b1-fd3f5cbd16f3\") " Jan 21 11:28:16 crc kubenswrapper[4824]: I0121 11:28:16.409257 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/9372ff5e-019e-47b0-b2b1-fd3f5cbd16f3-nova-metadata-tls-certs\") pod \"9372ff5e-019e-47b0-b2b1-fd3f5cbd16f3\" (UID: \"9372ff5e-019e-47b0-b2b1-fd3f5cbd16f3\") " Jan 21 11:28:16 crc kubenswrapper[4824]: I0121 11:28:16.409295 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9372ff5e-019e-47b0-b2b1-fd3f5cbd16f3-logs\") pod \"9372ff5e-019e-47b0-b2b1-fd3f5cbd16f3\" (UID: \"9372ff5e-019e-47b0-b2b1-fd3f5cbd16f3\") " Jan 21 11:28:16 crc kubenswrapper[4824]: I0121 11:28:16.410256 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9372ff5e-019e-47b0-b2b1-fd3f5cbd16f3-logs" (OuterVolumeSpecName: "logs") pod "9372ff5e-019e-47b0-b2b1-fd3f5cbd16f3" (UID: "9372ff5e-019e-47b0-b2b1-fd3f5cbd16f3"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:28:16 crc kubenswrapper[4824]: I0121 11:28:16.420737 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9372ff5e-019e-47b0-b2b1-fd3f5cbd16f3-kube-api-access-m24v8" (OuterVolumeSpecName: "kube-api-access-m24v8") pod "9372ff5e-019e-47b0-b2b1-fd3f5cbd16f3" (UID: "9372ff5e-019e-47b0-b2b1-fd3f5cbd16f3"). InnerVolumeSpecName "kube-api-access-m24v8". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:28:16 crc kubenswrapper[4824]: I0121 11:28:16.424570 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Jan 21 11:28:16 crc kubenswrapper[4824]: I0121 11:28:16.444599 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9372ff5e-019e-47b0-b2b1-fd3f5cbd16f3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9372ff5e-019e-47b0-b2b1-fd3f5cbd16f3" (UID: "9372ff5e-019e-47b0-b2b1-fd3f5cbd16f3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:28:16 crc kubenswrapper[4824]: I0121 11:28:16.453759 4824 scope.go:117] "RemoveContainer" containerID="77516fe4a72158d203b645cac966a850920a66f17eb3fb68fcdc002cb9e2b607" Jan 21 11:28:16 crc kubenswrapper[4824]: I0121 11:28:16.454030 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9372ff5e-019e-47b0-b2b1-fd3f5cbd16f3-config-data" (OuterVolumeSpecName: "config-data") pod "9372ff5e-019e-47b0-b2b1-fd3f5cbd16f3" (UID: "9372ff5e-019e-47b0-b2b1-fd3f5cbd16f3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:28:16 crc kubenswrapper[4824]: I0121 11:28:16.471494 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9372ff5e-019e-47b0-b2b1-fd3f5cbd16f3-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "9372ff5e-019e-47b0-b2b1-fd3f5cbd16f3" (UID: "9372ff5e-019e-47b0-b2b1-fd3f5cbd16f3"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:28:16 crc kubenswrapper[4824]: I0121 11:28:16.510656 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4a414d51-aaa8-43a6-88f5-f3ca59eb87a6-config-data\") pod \"4a414d51-aaa8-43a6-88f5-f3ca59eb87a6\" (UID: \"4a414d51-aaa8-43a6-88f5-f3ca59eb87a6\") " Jan 21 11:28:16 crc kubenswrapper[4824]: I0121 11:28:16.510689 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-znk42\" (UniqueName: \"kubernetes.io/projected/4a414d51-aaa8-43a6-88f5-f3ca59eb87a6-kube-api-access-znk42\") pod \"4a414d51-aaa8-43a6-88f5-f3ca59eb87a6\" (UID: \"4a414d51-aaa8-43a6-88f5-f3ca59eb87a6\") " Jan 21 11:28:16 crc kubenswrapper[4824]: I0121 11:28:16.510802 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a414d51-aaa8-43a6-88f5-f3ca59eb87a6-combined-ca-bundle\") pod \"4a414d51-aaa8-43a6-88f5-f3ca59eb87a6\" (UID: \"4a414d51-aaa8-43a6-88f5-f3ca59eb87a6\") " Jan 21 11:28:16 crc kubenswrapper[4824]: I0121 11:28:16.511363 4824 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9372ff5e-019e-47b0-b2b1-fd3f5cbd16f3-logs\") on node \"crc\" DevicePath \"\"" Jan 21 11:28:16 crc kubenswrapper[4824]: I0121 11:28:16.511380 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9372ff5e-019e-47b0-b2b1-fd3f5cbd16f3-config-data\") on node \"crc\" DevicePath \"\"" Jan 21 11:28:16 crc kubenswrapper[4824]: I0121 11:28:16.511389 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9372ff5e-019e-47b0-b2b1-fd3f5cbd16f3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 21 11:28:16 crc kubenswrapper[4824]: I0121 11:28:16.511398 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m24v8\" (UniqueName: \"kubernetes.io/projected/9372ff5e-019e-47b0-b2b1-fd3f5cbd16f3-kube-api-access-m24v8\") on node \"crc\" DevicePath \"\"" Jan 21 11:28:16 crc kubenswrapper[4824]: I0121 11:28:16.511407 4824 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/9372ff5e-019e-47b0-b2b1-fd3f5cbd16f3-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Jan 21 11:28:16 crc kubenswrapper[4824]: I0121 11:28:16.514071 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4a414d51-aaa8-43a6-88f5-f3ca59eb87a6-kube-api-access-znk42" (OuterVolumeSpecName: "kube-api-access-znk42") pod "4a414d51-aaa8-43a6-88f5-f3ca59eb87a6" (UID: "4a414d51-aaa8-43a6-88f5-f3ca59eb87a6"). InnerVolumeSpecName "kube-api-access-znk42". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:28:16 crc kubenswrapper[4824]: I0121 11:28:16.547947 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4a414d51-aaa8-43a6-88f5-f3ca59eb87a6-config-data" (OuterVolumeSpecName: "config-data") pod "4a414d51-aaa8-43a6-88f5-f3ca59eb87a6" (UID: "4a414d51-aaa8-43a6-88f5-f3ca59eb87a6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:28:16 crc kubenswrapper[4824]: I0121 11:28:16.548007 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4a414d51-aaa8-43a6-88f5-f3ca59eb87a6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4a414d51-aaa8-43a6-88f5-f3ca59eb87a6" (UID: "4a414d51-aaa8-43a6-88f5-f3ca59eb87a6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:28:16 crc kubenswrapper[4824]: I0121 11:28:16.612609 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a414d51-aaa8-43a6-88f5-f3ca59eb87a6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 21 11:28:16 crc kubenswrapper[4824]: I0121 11:28:16.612640 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4a414d51-aaa8-43a6-88f5-f3ca59eb87a6-config-data\") on node \"crc\" DevicePath \"\"" Jan 21 11:28:16 crc kubenswrapper[4824]: I0121 11:28:16.612650 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-znk42\" (UniqueName: \"kubernetes.io/projected/4a414d51-aaa8-43a6-88f5-f3ca59eb87a6-kube-api-access-znk42\") on node \"crc\" DevicePath \"\"" Jan 21 11:28:16 crc kubenswrapper[4824]: I0121 11:28:16.643939 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Jan 21 11:28:16 crc kubenswrapper[4824]: I0121 11:28:16.651548 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Jan 21 11:28:16 crc kubenswrapper[4824]: I0121 11:28:16.689506 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Jan 21 11:28:16 crc kubenswrapper[4824]: E0121 11:28:16.690177 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9372ff5e-019e-47b0-b2b1-fd3f5cbd16f3" containerName="nova-metadata-metadata" Jan 21 11:28:16 crc kubenswrapper[4824]: I0121 11:28:16.690197 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="9372ff5e-019e-47b0-b2b1-fd3f5cbd16f3" containerName="nova-metadata-metadata" Jan 21 11:28:16 crc kubenswrapper[4824]: E0121 11:28:16.690253 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9372ff5e-019e-47b0-b2b1-fd3f5cbd16f3" containerName="nova-metadata-log" Jan 21 11:28:16 crc kubenswrapper[4824]: I0121 11:28:16.690260 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="9372ff5e-019e-47b0-b2b1-fd3f5cbd16f3" containerName="nova-metadata-log" Jan 21 11:28:16 crc kubenswrapper[4824]: E0121 11:28:16.690269 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4a414d51-aaa8-43a6-88f5-f3ca59eb87a6" containerName="nova-scheduler-scheduler" Jan 21 11:28:16 crc kubenswrapper[4824]: I0121 11:28:16.690276 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="4a414d51-aaa8-43a6-88f5-f3ca59eb87a6" containerName="nova-scheduler-scheduler" Jan 21 11:28:16 crc kubenswrapper[4824]: I0121 11:28:16.690516 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="4a414d51-aaa8-43a6-88f5-f3ca59eb87a6" containerName="nova-scheduler-scheduler" Jan 21 11:28:16 crc kubenswrapper[4824]: I0121 11:28:16.690539 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="9372ff5e-019e-47b0-b2b1-fd3f5cbd16f3" containerName="nova-metadata-log" Jan 21 11:28:16 crc kubenswrapper[4824]: I0121 11:28:16.690567 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="9372ff5e-019e-47b0-b2b1-fd3f5cbd16f3" containerName="nova-metadata-metadata" Jan 21 11:28:16 crc kubenswrapper[4824]: I0121 11:28:16.691935 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Jan 21 11:28:16 crc kubenswrapper[4824]: I0121 11:28:16.693461 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Jan 21 11:28:16 crc kubenswrapper[4824]: I0121 11:28:16.693842 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Jan 21 11:28:16 crc kubenswrapper[4824]: I0121 11:28:16.695900 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Jan 21 11:28:16 crc kubenswrapper[4824]: I0121 11:28:16.816548 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/3b9162ed-f3d2-4877-b8a3-6342a25200b7-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"3b9162ed-f3d2-4877-b8a3-6342a25200b7\") " pod="openstack/nova-metadata-0" Jan 21 11:28:16 crc kubenswrapper[4824]: I0121 11:28:16.816646 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3b9162ed-f3d2-4877-b8a3-6342a25200b7-logs\") pod \"nova-metadata-0\" (UID: \"3b9162ed-f3d2-4877-b8a3-6342a25200b7\") " pod="openstack/nova-metadata-0" Jan 21 11:28:16 crc kubenswrapper[4824]: I0121 11:28:16.816910 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lpvwd\" (UniqueName: \"kubernetes.io/projected/3b9162ed-f3d2-4877-b8a3-6342a25200b7-kube-api-access-lpvwd\") pod \"nova-metadata-0\" (UID: \"3b9162ed-f3d2-4877-b8a3-6342a25200b7\") " pod="openstack/nova-metadata-0" Jan 21 11:28:16 crc kubenswrapper[4824]: I0121 11:28:16.816952 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3b9162ed-f3d2-4877-b8a3-6342a25200b7-config-data\") pod \"nova-metadata-0\" (UID: \"3b9162ed-f3d2-4877-b8a3-6342a25200b7\") " pod="openstack/nova-metadata-0" Jan 21 11:28:16 crc kubenswrapper[4824]: I0121 11:28:16.817120 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b9162ed-f3d2-4877-b8a3-6342a25200b7-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"3b9162ed-f3d2-4877-b8a3-6342a25200b7\") " pod="openstack/nova-metadata-0" Jan 21 11:28:16 crc kubenswrapper[4824]: I0121 11:28:16.918653 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lpvwd\" (UniqueName: \"kubernetes.io/projected/3b9162ed-f3d2-4877-b8a3-6342a25200b7-kube-api-access-lpvwd\") pod \"nova-metadata-0\" (UID: \"3b9162ed-f3d2-4877-b8a3-6342a25200b7\") " pod="openstack/nova-metadata-0" Jan 21 11:28:16 crc kubenswrapper[4824]: I0121 11:28:16.918708 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3b9162ed-f3d2-4877-b8a3-6342a25200b7-config-data\") pod \"nova-metadata-0\" (UID: \"3b9162ed-f3d2-4877-b8a3-6342a25200b7\") " pod="openstack/nova-metadata-0" Jan 21 11:28:16 crc kubenswrapper[4824]: I0121 11:28:16.918779 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b9162ed-f3d2-4877-b8a3-6342a25200b7-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"3b9162ed-f3d2-4877-b8a3-6342a25200b7\") " pod="openstack/nova-metadata-0" Jan 21 11:28:16 crc kubenswrapper[4824]: I0121 11:28:16.918883 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/3b9162ed-f3d2-4877-b8a3-6342a25200b7-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"3b9162ed-f3d2-4877-b8a3-6342a25200b7\") " pod="openstack/nova-metadata-0" Jan 21 11:28:16 crc kubenswrapper[4824]: I0121 11:28:16.918920 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3b9162ed-f3d2-4877-b8a3-6342a25200b7-logs\") pod \"nova-metadata-0\" (UID: \"3b9162ed-f3d2-4877-b8a3-6342a25200b7\") " pod="openstack/nova-metadata-0" Jan 21 11:28:16 crc kubenswrapper[4824]: I0121 11:28:16.919679 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3b9162ed-f3d2-4877-b8a3-6342a25200b7-logs\") pod \"nova-metadata-0\" (UID: \"3b9162ed-f3d2-4877-b8a3-6342a25200b7\") " pod="openstack/nova-metadata-0" Jan 21 11:28:16 crc kubenswrapper[4824]: I0121 11:28:16.924143 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/3b9162ed-f3d2-4877-b8a3-6342a25200b7-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"3b9162ed-f3d2-4877-b8a3-6342a25200b7\") " pod="openstack/nova-metadata-0" Jan 21 11:28:16 crc kubenswrapper[4824]: I0121 11:28:16.924603 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3b9162ed-f3d2-4877-b8a3-6342a25200b7-config-data\") pod \"nova-metadata-0\" (UID: \"3b9162ed-f3d2-4877-b8a3-6342a25200b7\") " pod="openstack/nova-metadata-0" Jan 21 11:28:16 crc kubenswrapper[4824]: I0121 11:28:16.927386 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b9162ed-f3d2-4877-b8a3-6342a25200b7-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"3b9162ed-f3d2-4877-b8a3-6342a25200b7\") " pod="openstack/nova-metadata-0" Jan 21 11:28:16 crc kubenswrapper[4824]: I0121 11:28:16.935436 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lpvwd\" (UniqueName: \"kubernetes.io/projected/3b9162ed-f3d2-4877-b8a3-6342a25200b7-kube-api-access-lpvwd\") pod \"nova-metadata-0\" (UID: \"3b9162ed-f3d2-4877-b8a3-6342a25200b7\") " pod="openstack/nova-metadata-0" Jan 21 11:28:17 crc kubenswrapper[4824]: I0121 11:28:17.010404 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Jan 21 11:28:17 crc kubenswrapper[4824]: I0121 11:28:17.335508 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"42d7c353-4ca7-4fa0-b4da-4793ce3e36a1","Type":"ContainerStarted","Data":"d91972673e1ac4ea40f718ccf4ab2ec2575e9319007c3a3cc55cff5c9b8fc08a"} Jan 21 11:28:17 crc kubenswrapper[4824]: I0121 11:28:17.336461 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Jan 21 11:28:17 crc kubenswrapper[4824]: I0121 11:28:17.338407 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"4a414d51-aaa8-43a6-88f5-f3ca59eb87a6","Type":"ContainerDied","Data":"720379547d053b3767f55293ad70e28fa03d8b901bf9feb0905f084febf1fbbe"} Jan 21 11:28:17 crc kubenswrapper[4824]: I0121 11:28:17.338439 4824 scope.go:117] "RemoveContainer" containerID="e871bc06d477443b88ec205254bfc60975ceca19ed6da8e295be1de5bf3e986e" Jan 21 11:28:17 crc kubenswrapper[4824]: I0121 11:28:17.338530 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Jan 21 11:28:17 crc kubenswrapper[4824]: I0121 11:28:17.343877 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"75cf0b57-bdd7-4b26-8d52-5acfbe9d520d","Type":"ContainerStarted","Data":"ca3b941babbbd07e235f5eab12e05e43e0ac2f0218981d4b14fc63386c42e254"} Jan 21 11:28:17 crc kubenswrapper[4824]: I0121 11:28:17.347924 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"c398d1e9-13e1-400c-a609-bceea49aea51","Type":"ContainerStarted","Data":"f51abad78825f4a03796cceaefb7a7b874eb98078ea107786fb273dcdc8422a5"} Jan 21 11:28:17 crc kubenswrapper[4824]: I0121 11:28:17.348051 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"c398d1e9-13e1-400c-a609-bceea49aea51","Type":"ContainerStarted","Data":"4275ae8b0ade24135f7edd7238e9a5b6244c1754d3bdfca05db8d8222ab9befc"} Jan 21 11:28:17 crc kubenswrapper[4824]: I0121 11:28:17.348530 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Jan 21 11:28:17 crc kubenswrapper[4824]: I0121 11:28:17.359362 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.148877432 podStartE2EDuration="5.359350477s" podCreationTimestamp="2026-01-21 11:28:12 +0000 UTC" firstStartedPulling="2026-01-21 11:28:13.04601383 +0000 UTC m=+1035.339043122" lastFinishedPulling="2026-01-21 11:28:16.256486875 +0000 UTC m=+1038.549516167" observedRunningTime="2026-01-21 11:28:17.357813489 +0000 UTC m=+1039.650842781" watchObservedRunningTime="2026-01-21 11:28:17.359350477 +0000 UTC m=+1039.652379769" Jan 21 11:28:17 crc kubenswrapper[4824]: I0121 11:28:17.385580 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.385538965 podStartE2EDuration="2.385538965s" podCreationTimestamp="2026-01-21 11:28:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:28:17.381660139 +0000 UTC m=+1039.674689431" watchObservedRunningTime="2026-01-21 11:28:17.385538965 +0000 UTC m=+1039.678568257" Jan 21 11:28:17 crc kubenswrapper[4824]: W0121 11:28:17.410119 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3b9162ed_f3d2_4877_b8a3_6342a25200b7.slice/crio-cf896b02cb7f72cd22b94afd5cb645eb3dc94f2424b20aa748a3b1e35979c8a8 WatchSource:0}: Error finding container cf896b02cb7f72cd22b94afd5cb645eb3dc94f2424b20aa748a3b1e35979c8a8: Status 404 returned error can't find the container with id cf896b02cb7f72cd22b94afd5cb645eb3dc94f2424b20aa748a3b1e35979c8a8 Jan 21 11:28:17 crc kubenswrapper[4824]: I0121 11:28:17.413365 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Jan 21 11:28:17 crc kubenswrapper[4824]: I0121 11:28:17.427231 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Jan 21 11:28:17 crc kubenswrapper[4824]: I0121 11:28:17.440341 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Jan 21 11:28:17 crc kubenswrapper[4824]: I0121 11:28:17.446838 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Jan 21 11:28:17 crc kubenswrapper[4824]: I0121 11:28:17.447970 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Jan 21 11:28:17 crc kubenswrapper[4824]: I0121 11:28:17.449372 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Jan 21 11:28:17 crc kubenswrapper[4824]: I0121 11:28:17.450867 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.45085198 podStartE2EDuration="2.45085198s" podCreationTimestamp="2026-01-21 11:28:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:28:17.410184682 +0000 UTC m=+1039.703213974" watchObservedRunningTime="2026-01-21 11:28:17.45085198 +0000 UTC m=+1039.743881273" Jan 21 11:28:17 crc kubenswrapper[4824]: I0121 11:28:17.463698 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Jan 21 11:28:17 crc kubenswrapper[4824]: I0121 11:28:17.528571 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nsphm\" (UniqueName: \"kubernetes.io/projected/a8b9cd0d-c7af-455d-a170-1521870a07f6-kube-api-access-nsphm\") pod \"nova-scheduler-0\" (UID: \"a8b9cd0d-c7af-455d-a170-1521870a07f6\") " pod="openstack/nova-scheduler-0" Jan 21 11:28:17 crc kubenswrapper[4824]: I0121 11:28:17.528846 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a8b9cd0d-c7af-455d-a170-1521870a07f6-config-data\") pod \"nova-scheduler-0\" (UID: \"a8b9cd0d-c7af-455d-a170-1521870a07f6\") " pod="openstack/nova-scheduler-0" Jan 21 11:28:17 crc kubenswrapper[4824]: I0121 11:28:17.529011 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8b9cd0d-c7af-455d-a170-1521870a07f6-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"a8b9cd0d-c7af-455d-a170-1521870a07f6\") " pod="openstack/nova-scheduler-0" Jan 21 11:28:17 crc kubenswrapper[4824]: I0121 11:28:17.630564 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8b9cd0d-c7af-455d-a170-1521870a07f6-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"a8b9cd0d-c7af-455d-a170-1521870a07f6\") " pod="openstack/nova-scheduler-0" Jan 21 11:28:17 crc kubenswrapper[4824]: I0121 11:28:17.630786 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nsphm\" (UniqueName: \"kubernetes.io/projected/a8b9cd0d-c7af-455d-a170-1521870a07f6-kube-api-access-nsphm\") pod \"nova-scheduler-0\" (UID: \"a8b9cd0d-c7af-455d-a170-1521870a07f6\") " pod="openstack/nova-scheduler-0" Jan 21 11:28:17 crc kubenswrapper[4824]: I0121 11:28:17.631007 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a8b9cd0d-c7af-455d-a170-1521870a07f6-config-data\") pod \"nova-scheduler-0\" (UID: \"a8b9cd0d-c7af-455d-a170-1521870a07f6\") " pod="openstack/nova-scheduler-0" Jan 21 11:28:17 crc kubenswrapper[4824]: I0121 11:28:17.633565 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8b9cd0d-c7af-455d-a170-1521870a07f6-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"a8b9cd0d-c7af-455d-a170-1521870a07f6\") " pod="openstack/nova-scheduler-0" Jan 21 11:28:17 crc kubenswrapper[4824]: I0121 11:28:17.633809 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a8b9cd0d-c7af-455d-a170-1521870a07f6-config-data\") pod \"nova-scheduler-0\" (UID: \"a8b9cd0d-c7af-455d-a170-1521870a07f6\") " pod="openstack/nova-scheduler-0" Jan 21 11:28:17 crc kubenswrapper[4824]: I0121 11:28:17.644935 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nsphm\" (UniqueName: \"kubernetes.io/projected/a8b9cd0d-c7af-455d-a170-1521870a07f6-kube-api-access-nsphm\") pod \"nova-scheduler-0\" (UID: \"a8b9cd0d-c7af-455d-a170-1521870a07f6\") " pod="openstack/nova-scheduler-0" Jan 21 11:28:17 crc kubenswrapper[4824]: I0121 11:28:17.763274 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Jan 21 11:28:18 crc kubenswrapper[4824]: I0121 11:28:18.063723 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4a414d51-aaa8-43a6-88f5-f3ca59eb87a6" path="/var/lib/kubelet/pods/4a414d51-aaa8-43a6-88f5-f3ca59eb87a6/volumes" Jan 21 11:28:18 crc kubenswrapper[4824]: I0121 11:28:18.064594 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9372ff5e-019e-47b0-b2b1-fd3f5cbd16f3" path="/var/lib/kubelet/pods/9372ff5e-019e-47b0-b2b1-fd3f5cbd16f3/volumes" Jan 21 11:28:18 crc kubenswrapper[4824]: I0121 11:28:18.065152 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f2ec0128-5881-4c67-8d78-d53d8be3a2d3" path="/var/lib/kubelet/pods/f2ec0128-5881-4c67-8d78-d53d8be3a2d3/volumes" Jan 21 11:28:18 crc kubenswrapper[4824]: W0121 11:28:18.139674 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda8b9cd0d_c7af_455d_a170_1521870a07f6.slice/crio-bcd965629c724abfbdcc4781034ea86d257b188520a919034221d6a8bf2c0f86 WatchSource:0}: Error finding container bcd965629c724abfbdcc4781034ea86d257b188520a919034221d6a8bf2c0f86: Status 404 returned error can't find the container with id bcd965629c724abfbdcc4781034ea86d257b188520a919034221d6a8bf2c0f86 Jan 21 11:28:18 crc kubenswrapper[4824]: I0121 11:28:18.140762 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Jan 21 11:28:18 crc kubenswrapper[4824]: I0121 11:28:18.358139 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"a8b9cd0d-c7af-455d-a170-1521870a07f6","Type":"ContainerStarted","Data":"be161c0756cab08e44d6825e362adf099303c412379798ba94c4f4bf705f4830"} Jan 21 11:28:18 crc kubenswrapper[4824]: I0121 11:28:18.358344 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"a8b9cd0d-c7af-455d-a170-1521870a07f6","Type":"ContainerStarted","Data":"bcd965629c724abfbdcc4781034ea86d257b188520a919034221d6a8bf2c0f86"} Jan 21 11:28:18 crc kubenswrapper[4824]: I0121 11:28:18.361691 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"3b9162ed-f3d2-4877-b8a3-6342a25200b7","Type":"ContainerStarted","Data":"12872dcc363c76e366d89e82065e5f0f0eb8251e49a88c72fc419ac5a668148c"} Jan 21 11:28:18 crc kubenswrapper[4824]: I0121 11:28:18.361728 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"3b9162ed-f3d2-4877-b8a3-6342a25200b7","Type":"ContainerStarted","Data":"55a7e3a41e9be3e82e087f957b241f50bebac19bc452a77348ef96c1447373d9"} Jan 21 11:28:18 crc kubenswrapper[4824]: I0121 11:28:18.361741 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"3b9162ed-f3d2-4877-b8a3-6342a25200b7","Type":"ContainerStarted","Data":"cf896b02cb7f72cd22b94afd5cb645eb3dc94f2424b20aa748a3b1e35979c8a8"} Jan 21 11:28:18 crc kubenswrapper[4824]: I0121 11:28:18.369668 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=1.369655436 podStartE2EDuration="1.369655436s" podCreationTimestamp="2026-01-21 11:28:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:28:18.368504105 +0000 UTC m=+1040.661533397" watchObservedRunningTime="2026-01-21 11:28:18.369655436 +0000 UTC m=+1040.662684728" Jan 21 11:28:18 crc kubenswrapper[4824]: I0121 11:28:18.384846 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.384834546 podStartE2EDuration="2.384834546s" podCreationTimestamp="2026-01-21 11:28:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:28:18.383658219 +0000 UTC m=+1040.676687511" watchObservedRunningTime="2026-01-21 11:28:18.384834546 +0000 UTC m=+1040.677863837" Jan 21 11:28:22 crc kubenswrapper[4824]: I0121 11:28:22.010934 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Jan 21 11:28:22 crc kubenswrapper[4824]: I0121 11:28:22.011422 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Jan 21 11:28:22 crc kubenswrapper[4824]: I0121 11:28:22.764157 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Jan 21 11:28:25 crc kubenswrapper[4824]: I0121 11:28:25.695998 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Jan 21 11:28:25 crc kubenswrapper[4824]: I0121 11:28:25.696346 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Jan 21 11:28:25 crc kubenswrapper[4824]: I0121 11:28:25.783870 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Jan 21 11:28:26 crc kubenswrapper[4824]: I0121 11:28:26.779093 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="75cf0b57-bdd7-4b26-8d52-5acfbe9d520d" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.202:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Jan 21 11:28:26 crc kubenswrapper[4824]: I0121 11:28:26.779093 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="75cf0b57-bdd7-4b26-8d52-5acfbe9d520d" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.202:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Jan 21 11:28:27 crc kubenswrapper[4824]: I0121 11:28:27.011591 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Jan 21 11:28:27 crc kubenswrapper[4824]: I0121 11:28:27.011779 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Jan 21 11:28:27 crc kubenswrapper[4824]: I0121 11:28:27.763620 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Jan 21 11:28:27 crc kubenswrapper[4824]: I0121 11:28:27.788212 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Jan 21 11:28:28 crc kubenswrapper[4824]: I0121 11:28:28.025092 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="3b9162ed-f3d2-4877-b8a3-6342a25200b7" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.204:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Jan 21 11:28:28 crc kubenswrapper[4824]: I0121 11:28:28.025096 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="3b9162ed-f3d2-4877-b8a3-6342a25200b7" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.204:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Jan 21 11:28:28 crc kubenswrapper[4824]: I0121 11:28:28.460270 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Jan 21 11:28:35 crc kubenswrapper[4824]: I0121 11:28:35.698991 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Jan 21 11:28:35 crc kubenswrapper[4824]: I0121 11:28:35.700048 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Jan 21 11:28:35 crc kubenswrapper[4824]: I0121 11:28:35.700360 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Jan 21 11:28:35 crc kubenswrapper[4824]: I0121 11:28:35.700390 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Jan 21 11:28:35 crc kubenswrapper[4824]: I0121 11:28:35.703039 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Jan 21 11:28:35 crc kubenswrapper[4824]: I0121 11:28:35.703224 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Jan 21 11:28:35 crc kubenswrapper[4824]: I0121 11:28:35.854646 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6b7bbf7cf9-tmvm9"] Jan 21 11:28:35 crc kubenswrapper[4824]: I0121 11:28:35.866500 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b7bbf7cf9-tmvm9" Jan 21 11:28:35 crc kubenswrapper[4824]: I0121 11:28:35.886714 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6b7bbf7cf9-tmvm9"] Jan 21 11:28:35 crc kubenswrapper[4824]: I0121 11:28:35.957996 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8ad8dd9e-6fe4-462b-98a9-d0192072589d-ovsdbserver-sb\") pod \"dnsmasq-dns-6b7bbf7cf9-tmvm9\" (UID: \"8ad8dd9e-6fe4-462b-98a9-d0192072589d\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-tmvm9" Jan 21 11:28:35 crc kubenswrapper[4824]: I0121 11:28:35.958064 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8ad8dd9e-6fe4-462b-98a9-d0192072589d-config\") pod \"dnsmasq-dns-6b7bbf7cf9-tmvm9\" (UID: \"8ad8dd9e-6fe4-462b-98a9-d0192072589d\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-tmvm9" Jan 21 11:28:35 crc kubenswrapper[4824]: I0121 11:28:35.958103 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8ad8dd9e-6fe4-462b-98a9-d0192072589d-ovsdbserver-nb\") pod \"dnsmasq-dns-6b7bbf7cf9-tmvm9\" (UID: \"8ad8dd9e-6fe4-462b-98a9-d0192072589d\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-tmvm9" Jan 21 11:28:35 crc kubenswrapper[4824]: I0121 11:28:35.958484 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bkm2s\" (UniqueName: \"kubernetes.io/projected/8ad8dd9e-6fe4-462b-98a9-d0192072589d-kube-api-access-bkm2s\") pod \"dnsmasq-dns-6b7bbf7cf9-tmvm9\" (UID: \"8ad8dd9e-6fe4-462b-98a9-d0192072589d\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-tmvm9" Jan 21 11:28:35 crc kubenswrapper[4824]: I0121 11:28:35.958713 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8ad8dd9e-6fe4-462b-98a9-d0192072589d-dns-swift-storage-0\") pod \"dnsmasq-dns-6b7bbf7cf9-tmvm9\" (UID: \"8ad8dd9e-6fe4-462b-98a9-d0192072589d\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-tmvm9" Jan 21 11:28:35 crc kubenswrapper[4824]: I0121 11:28:35.958783 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8ad8dd9e-6fe4-462b-98a9-d0192072589d-dns-svc\") pod \"dnsmasq-dns-6b7bbf7cf9-tmvm9\" (UID: \"8ad8dd9e-6fe4-462b-98a9-d0192072589d\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-tmvm9" Jan 21 11:28:36 crc kubenswrapper[4824]: I0121 11:28:36.060072 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bkm2s\" (UniqueName: \"kubernetes.io/projected/8ad8dd9e-6fe4-462b-98a9-d0192072589d-kube-api-access-bkm2s\") pod \"dnsmasq-dns-6b7bbf7cf9-tmvm9\" (UID: \"8ad8dd9e-6fe4-462b-98a9-d0192072589d\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-tmvm9" Jan 21 11:28:36 crc kubenswrapper[4824]: I0121 11:28:36.060163 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8ad8dd9e-6fe4-462b-98a9-d0192072589d-dns-swift-storage-0\") pod \"dnsmasq-dns-6b7bbf7cf9-tmvm9\" (UID: \"8ad8dd9e-6fe4-462b-98a9-d0192072589d\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-tmvm9" Jan 21 11:28:36 crc kubenswrapper[4824]: I0121 11:28:36.060191 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8ad8dd9e-6fe4-462b-98a9-d0192072589d-dns-svc\") pod \"dnsmasq-dns-6b7bbf7cf9-tmvm9\" (UID: \"8ad8dd9e-6fe4-462b-98a9-d0192072589d\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-tmvm9" Jan 21 11:28:36 crc kubenswrapper[4824]: I0121 11:28:36.060236 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8ad8dd9e-6fe4-462b-98a9-d0192072589d-ovsdbserver-sb\") pod \"dnsmasq-dns-6b7bbf7cf9-tmvm9\" (UID: \"8ad8dd9e-6fe4-462b-98a9-d0192072589d\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-tmvm9" Jan 21 11:28:36 crc kubenswrapper[4824]: I0121 11:28:36.060264 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8ad8dd9e-6fe4-462b-98a9-d0192072589d-config\") pod \"dnsmasq-dns-6b7bbf7cf9-tmvm9\" (UID: \"8ad8dd9e-6fe4-462b-98a9-d0192072589d\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-tmvm9" Jan 21 11:28:36 crc kubenswrapper[4824]: I0121 11:28:36.060294 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8ad8dd9e-6fe4-462b-98a9-d0192072589d-ovsdbserver-nb\") pod \"dnsmasq-dns-6b7bbf7cf9-tmvm9\" (UID: \"8ad8dd9e-6fe4-462b-98a9-d0192072589d\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-tmvm9" Jan 21 11:28:36 crc kubenswrapper[4824]: I0121 11:28:36.061065 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8ad8dd9e-6fe4-462b-98a9-d0192072589d-ovsdbserver-nb\") pod \"dnsmasq-dns-6b7bbf7cf9-tmvm9\" (UID: \"8ad8dd9e-6fe4-462b-98a9-d0192072589d\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-tmvm9" Jan 21 11:28:36 crc kubenswrapper[4824]: I0121 11:28:36.061764 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8ad8dd9e-6fe4-462b-98a9-d0192072589d-dns-swift-storage-0\") pod \"dnsmasq-dns-6b7bbf7cf9-tmvm9\" (UID: \"8ad8dd9e-6fe4-462b-98a9-d0192072589d\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-tmvm9" Jan 21 11:28:36 crc kubenswrapper[4824]: I0121 11:28:36.062259 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8ad8dd9e-6fe4-462b-98a9-d0192072589d-dns-svc\") pod \"dnsmasq-dns-6b7bbf7cf9-tmvm9\" (UID: \"8ad8dd9e-6fe4-462b-98a9-d0192072589d\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-tmvm9" Jan 21 11:28:36 crc kubenswrapper[4824]: I0121 11:28:36.062265 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8ad8dd9e-6fe4-462b-98a9-d0192072589d-config\") pod \"dnsmasq-dns-6b7bbf7cf9-tmvm9\" (UID: \"8ad8dd9e-6fe4-462b-98a9-d0192072589d\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-tmvm9" Jan 21 11:28:36 crc kubenswrapper[4824]: I0121 11:28:36.062976 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8ad8dd9e-6fe4-462b-98a9-d0192072589d-ovsdbserver-sb\") pod \"dnsmasq-dns-6b7bbf7cf9-tmvm9\" (UID: \"8ad8dd9e-6fe4-462b-98a9-d0192072589d\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-tmvm9" Jan 21 11:28:36 crc kubenswrapper[4824]: I0121 11:28:36.082485 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bkm2s\" (UniqueName: \"kubernetes.io/projected/8ad8dd9e-6fe4-462b-98a9-d0192072589d-kube-api-access-bkm2s\") pod \"dnsmasq-dns-6b7bbf7cf9-tmvm9\" (UID: \"8ad8dd9e-6fe4-462b-98a9-d0192072589d\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-tmvm9" Jan 21 11:28:36 crc kubenswrapper[4824]: I0121 11:28:36.197998 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b7bbf7cf9-tmvm9" Jan 21 11:28:36 crc kubenswrapper[4824]: W0121 11:28:36.693782 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8ad8dd9e_6fe4_462b_98a9_d0192072589d.slice/crio-6738b1680c3c03cab4e1e7a4ccdc6941fbdb963c0035872cd1e033904215c51b WatchSource:0}: Error finding container 6738b1680c3c03cab4e1e7a4ccdc6941fbdb963c0035872cd1e033904215c51b: Status 404 returned error can't find the container with id 6738b1680c3c03cab4e1e7a4ccdc6941fbdb963c0035872cd1e033904215c51b Jan 21 11:28:36 crc kubenswrapper[4824]: I0121 11:28:36.694451 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6b7bbf7cf9-tmvm9"] Jan 21 11:28:37 crc kubenswrapper[4824]: I0121 11:28:37.017691 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Jan 21 11:28:37 crc kubenswrapper[4824]: I0121 11:28:37.023189 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Jan 21 11:28:37 crc kubenswrapper[4824]: I0121 11:28:37.023927 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Jan 21 11:28:37 crc kubenswrapper[4824]: I0121 11:28:37.499029 4824 generic.go:334] "Generic (PLEG): container finished" podID="8ad8dd9e-6fe4-462b-98a9-d0192072589d" containerID="ed4051f2d4931880d9659716eb3f4ebdfc337e399698e29d497ad42bb33cd14f" exitCode=0 Jan 21 11:28:37 crc kubenswrapper[4824]: I0121 11:28:37.499114 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7bbf7cf9-tmvm9" event={"ID":"8ad8dd9e-6fe4-462b-98a9-d0192072589d","Type":"ContainerDied","Data":"ed4051f2d4931880d9659716eb3f4ebdfc337e399698e29d497ad42bb33cd14f"} Jan 21 11:28:37 crc kubenswrapper[4824]: I0121 11:28:37.499157 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7bbf7cf9-tmvm9" event={"ID":"8ad8dd9e-6fe4-462b-98a9-d0192072589d","Type":"ContainerStarted","Data":"6738b1680c3c03cab4e1e7a4ccdc6941fbdb963c0035872cd1e033904215c51b"} Jan 21 11:28:37 crc kubenswrapper[4824]: I0121 11:28:37.540617 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Jan 21 11:28:37 crc kubenswrapper[4824]: I0121 11:28:37.686763 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Jan 21 11:28:37 crc kubenswrapper[4824]: I0121 11:28:37.687202 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="42d7c353-4ca7-4fa0-b4da-4793ce3e36a1" containerName="ceilometer-central-agent" containerID="cri-o://82b46e3d25bb50cb39acd7bbf34eb02ad735f71fcba98f606eba2559159f21c7" gracePeriod=30 Jan 21 11:28:37 crc kubenswrapper[4824]: I0121 11:28:37.687309 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="42d7c353-4ca7-4fa0-b4da-4793ce3e36a1" containerName="proxy-httpd" containerID="cri-o://d91972673e1ac4ea40f718ccf4ab2ec2575e9319007c3a3cc55cff5c9b8fc08a" gracePeriod=30 Jan 21 11:28:37 crc kubenswrapper[4824]: I0121 11:28:37.687345 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="42d7c353-4ca7-4fa0-b4da-4793ce3e36a1" containerName="sg-core" containerID="cri-o://87edd33f33351d88d9cbd6b17c925a8893b337952321dc91c1ba77e5c174ae3c" gracePeriod=30 Jan 21 11:28:37 crc kubenswrapper[4824]: I0121 11:28:37.687375 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="42d7c353-4ca7-4fa0-b4da-4793ce3e36a1" containerName="ceilometer-notification-agent" containerID="cri-o://97c437ef91be90c480bbbc66844b2f1bf63fc92b3d5efc241a8ff7639e0f322c" gracePeriod=30 Jan 21 11:28:37 crc kubenswrapper[4824]: I0121 11:28:37.787419 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="42d7c353-4ca7-4fa0-b4da-4793ce3e36a1" containerName="proxy-httpd" probeResult="failure" output="Get \"http://10.217.0.201:3000/\": read tcp 10.217.0.2:39062->10.217.0.201:3000: read: connection reset by peer" Jan 21 11:28:38 crc kubenswrapper[4824]: I0121 11:28:38.166506 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Jan 21 11:28:38 crc kubenswrapper[4824]: I0121 11:28:38.507365 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7bbf7cf9-tmvm9" event={"ID":"8ad8dd9e-6fe4-462b-98a9-d0192072589d","Type":"ContainerStarted","Data":"8687725a582d2593d699bc27c414dbcbeca025f725a9dfdc0c21cbe095d4f575"} Jan 21 11:28:38 crc kubenswrapper[4824]: I0121 11:28:38.507754 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6b7bbf7cf9-tmvm9" Jan 21 11:28:38 crc kubenswrapper[4824]: I0121 11:28:38.510114 4824 generic.go:334] "Generic (PLEG): container finished" podID="42d7c353-4ca7-4fa0-b4da-4793ce3e36a1" containerID="d91972673e1ac4ea40f718ccf4ab2ec2575e9319007c3a3cc55cff5c9b8fc08a" exitCode=0 Jan 21 11:28:38 crc kubenswrapper[4824]: I0121 11:28:38.510143 4824 generic.go:334] "Generic (PLEG): container finished" podID="42d7c353-4ca7-4fa0-b4da-4793ce3e36a1" containerID="87edd33f33351d88d9cbd6b17c925a8893b337952321dc91c1ba77e5c174ae3c" exitCode=2 Jan 21 11:28:38 crc kubenswrapper[4824]: I0121 11:28:38.510150 4824 generic.go:334] "Generic (PLEG): container finished" podID="42d7c353-4ca7-4fa0-b4da-4793ce3e36a1" containerID="82b46e3d25bb50cb39acd7bbf34eb02ad735f71fcba98f606eba2559159f21c7" exitCode=0 Jan 21 11:28:38 crc kubenswrapper[4824]: I0121 11:28:38.510220 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"42d7c353-4ca7-4fa0-b4da-4793ce3e36a1","Type":"ContainerDied","Data":"d91972673e1ac4ea40f718ccf4ab2ec2575e9319007c3a3cc55cff5c9b8fc08a"} Jan 21 11:28:38 crc kubenswrapper[4824]: I0121 11:28:38.510258 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"42d7c353-4ca7-4fa0-b4da-4793ce3e36a1","Type":"ContainerDied","Data":"87edd33f33351d88d9cbd6b17c925a8893b337952321dc91c1ba77e5c174ae3c"} Jan 21 11:28:38 crc kubenswrapper[4824]: I0121 11:28:38.510271 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"42d7c353-4ca7-4fa0-b4da-4793ce3e36a1","Type":"ContainerDied","Data":"82b46e3d25bb50cb39acd7bbf34eb02ad735f71fcba98f606eba2559159f21c7"} Jan 21 11:28:38 crc kubenswrapper[4824]: I0121 11:28:38.510313 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="75cf0b57-bdd7-4b26-8d52-5acfbe9d520d" containerName="nova-api-log" containerID="cri-o://7d8c20082288cf600e39dc7a0f1bd5ac8ba09165ca35559ccd6592d1b0add249" gracePeriod=30 Jan 21 11:28:38 crc kubenswrapper[4824]: I0121 11:28:38.510425 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="75cf0b57-bdd7-4b26-8d52-5acfbe9d520d" containerName="nova-api-api" containerID="cri-o://ca3b941babbbd07e235f5eab12e05e43e0ac2f0218981d4b14fc63386c42e254" gracePeriod=30 Jan 21 11:28:38 crc kubenswrapper[4824]: I0121 11:28:38.530283 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6b7bbf7cf9-tmvm9" podStartSLOduration=3.530267214 podStartE2EDuration="3.530267214s" podCreationTimestamp="2026-01-21 11:28:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:28:38.526274385 +0000 UTC m=+1060.819303678" watchObservedRunningTime="2026-01-21 11:28:38.530267214 +0000 UTC m=+1060.823296506" Jan 21 11:28:39 crc kubenswrapper[4824]: I0121 11:28:39.520087 4824 generic.go:334] "Generic (PLEG): container finished" podID="42d7c353-4ca7-4fa0-b4da-4793ce3e36a1" containerID="97c437ef91be90c480bbbc66844b2f1bf63fc92b3d5efc241a8ff7639e0f322c" exitCode=0 Jan 21 11:28:39 crc kubenswrapper[4824]: I0121 11:28:39.520330 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"42d7c353-4ca7-4fa0-b4da-4793ce3e36a1","Type":"ContainerDied","Data":"97c437ef91be90c480bbbc66844b2f1bf63fc92b3d5efc241a8ff7639e0f322c"} Jan 21 11:28:39 crc kubenswrapper[4824]: I0121 11:28:39.520353 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"42d7c353-4ca7-4fa0-b4da-4793ce3e36a1","Type":"ContainerDied","Data":"3debb3f0581abba1b7346533099c933cc5a8bb94ead0c540a6932f4be37b2b39"} Jan 21 11:28:39 crc kubenswrapper[4824]: I0121 11:28:39.520363 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3debb3f0581abba1b7346533099c933cc5a8bb94ead0c540a6932f4be37b2b39" Jan 21 11:28:39 crc kubenswrapper[4824]: I0121 11:28:39.522233 4824 generic.go:334] "Generic (PLEG): container finished" podID="75cf0b57-bdd7-4b26-8d52-5acfbe9d520d" containerID="7d8c20082288cf600e39dc7a0f1bd5ac8ba09165ca35559ccd6592d1b0add249" exitCode=143 Jan 21 11:28:39 crc kubenswrapper[4824]: I0121 11:28:39.522941 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"75cf0b57-bdd7-4b26-8d52-5acfbe9d520d","Type":"ContainerDied","Data":"7d8c20082288cf600e39dc7a0f1bd5ac8ba09165ca35559ccd6592d1b0add249"} Jan 21 11:28:39 crc kubenswrapper[4824]: I0121 11:28:39.528935 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 21 11:28:39 crc kubenswrapper[4824]: I0121 11:28:39.620459 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/42d7c353-4ca7-4fa0-b4da-4793ce3e36a1-run-httpd\") pod \"42d7c353-4ca7-4fa0-b4da-4793ce3e36a1\" (UID: \"42d7c353-4ca7-4fa0-b4da-4793ce3e36a1\") " Jan 21 11:28:39 crc kubenswrapper[4824]: I0121 11:28:39.620520 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dmqz4\" (UniqueName: \"kubernetes.io/projected/42d7c353-4ca7-4fa0-b4da-4793ce3e36a1-kube-api-access-dmqz4\") pod \"42d7c353-4ca7-4fa0-b4da-4793ce3e36a1\" (UID: \"42d7c353-4ca7-4fa0-b4da-4793ce3e36a1\") " Jan 21 11:28:39 crc kubenswrapper[4824]: I0121 11:28:39.620657 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/42d7c353-4ca7-4fa0-b4da-4793ce3e36a1-config-data\") pod \"42d7c353-4ca7-4fa0-b4da-4793ce3e36a1\" (UID: \"42d7c353-4ca7-4fa0-b4da-4793ce3e36a1\") " Jan 21 11:28:39 crc kubenswrapper[4824]: I0121 11:28:39.620695 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/42d7c353-4ca7-4fa0-b4da-4793ce3e36a1-sg-core-conf-yaml\") pod \"42d7c353-4ca7-4fa0-b4da-4793ce3e36a1\" (UID: \"42d7c353-4ca7-4fa0-b4da-4793ce3e36a1\") " Jan 21 11:28:39 crc kubenswrapper[4824]: I0121 11:28:39.620754 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/42d7c353-4ca7-4fa0-b4da-4793ce3e36a1-scripts\") pod \"42d7c353-4ca7-4fa0-b4da-4793ce3e36a1\" (UID: \"42d7c353-4ca7-4fa0-b4da-4793ce3e36a1\") " Jan 21 11:28:39 crc kubenswrapper[4824]: I0121 11:28:39.620787 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/42d7c353-4ca7-4fa0-b4da-4793ce3e36a1-log-httpd\") pod \"42d7c353-4ca7-4fa0-b4da-4793ce3e36a1\" (UID: \"42d7c353-4ca7-4fa0-b4da-4793ce3e36a1\") " Jan 21 11:28:39 crc kubenswrapper[4824]: I0121 11:28:39.620813 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42d7c353-4ca7-4fa0-b4da-4793ce3e36a1-combined-ca-bundle\") pod \"42d7c353-4ca7-4fa0-b4da-4793ce3e36a1\" (UID: \"42d7c353-4ca7-4fa0-b4da-4793ce3e36a1\") " Jan 21 11:28:39 crc kubenswrapper[4824]: I0121 11:28:39.621504 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/42d7c353-4ca7-4fa0-b4da-4793ce3e36a1-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "42d7c353-4ca7-4fa0-b4da-4793ce3e36a1" (UID: "42d7c353-4ca7-4fa0-b4da-4793ce3e36a1"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:28:39 crc kubenswrapper[4824]: I0121 11:28:39.621842 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/42d7c353-4ca7-4fa0-b4da-4793ce3e36a1-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "42d7c353-4ca7-4fa0-b4da-4793ce3e36a1" (UID: "42d7c353-4ca7-4fa0-b4da-4793ce3e36a1"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:28:39 crc kubenswrapper[4824]: I0121 11:28:39.625355 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/42d7c353-4ca7-4fa0-b4da-4793ce3e36a1-scripts" (OuterVolumeSpecName: "scripts") pod "42d7c353-4ca7-4fa0-b4da-4793ce3e36a1" (UID: "42d7c353-4ca7-4fa0-b4da-4793ce3e36a1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:28:39 crc kubenswrapper[4824]: I0121 11:28:39.633035 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/42d7c353-4ca7-4fa0-b4da-4793ce3e36a1-kube-api-access-dmqz4" (OuterVolumeSpecName: "kube-api-access-dmqz4") pod "42d7c353-4ca7-4fa0-b4da-4793ce3e36a1" (UID: "42d7c353-4ca7-4fa0-b4da-4793ce3e36a1"). InnerVolumeSpecName "kube-api-access-dmqz4". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:28:39 crc kubenswrapper[4824]: I0121 11:28:39.645607 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/42d7c353-4ca7-4fa0-b4da-4793ce3e36a1-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "42d7c353-4ca7-4fa0-b4da-4793ce3e36a1" (UID: "42d7c353-4ca7-4fa0-b4da-4793ce3e36a1"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:28:39 crc kubenswrapper[4824]: I0121 11:28:39.686428 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/42d7c353-4ca7-4fa0-b4da-4793ce3e36a1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "42d7c353-4ca7-4fa0-b4da-4793ce3e36a1" (UID: "42d7c353-4ca7-4fa0-b4da-4793ce3e36a1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:28:39 crc kubenswrapper[4824]: I0121 11:28:39.704672 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/42d7c353-4ca7-4fa0-b4da-4793ce3e36a1-config-data" (OuterVolumeSpecName: "config-data") pod "42d7c353-4ca7-4fa0-b4da-4793ce3e36a1" (UID: "42d7c353-4ca7-4fa0-b4da-4793ce3e36a1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:28:39 crc kubenswrapper[4824]: I0121 11:28:39.722638 4824 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/42d7c353-4ca7-4fa0-b4da-4793ce3e36a1-scripts\") on node \"crc\" DevicePath \"\"" Jan 21 11:28:39 crc kubenswrapper[4824]: I0121 11:28:39.722664 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42d7c353-4ca7-4fa0-b4da-4793ce3e36a1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 21 11:28:39 crc kubenswrapper[4824]: I0121 11:28:39.722675 4824 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/42d7c353-4ca7-4fa0-b4da-4793ce3e36a1-log-httpd\") on node \"crc\" DevicePath \"\"" Jan 21 11:28:39 crc kubenswrapper[4824]: I0121 11:28:39.722683 4824 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/42d7c353-4ca7-4fa0-b4da-4793ce3e36a1-run-httpd\") on node \"crc\" DevicePath \"\"" Jan 21 11:28:39 crc kubenswrapper[4824]: I0121 11:28:39.722691 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dmqz4\" (UniqueName: \"kubernetes.io/projected/42d7c353-4ca7-4fa0-b4da-4793ce3e36a1-kube-api-access-dmqz4\") on node \"crc\" DevicePath \"\"" Jan 21 11:28:39 crc kubenswrapper[4824]: I0121 11:28:39.722699 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/42d7c353-4ca7-4fa0-b4da-4793ce3e36a1-config-data\") on node \"crc\" DevicePath \"\"" Jan 21 11:28:39 crc kubenswrapper[4824]: I0121 11:28:39.722707 4824 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/42d7c353-4ca7-4fa0-b4da-4793ce3e36a1-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Jan 21 11:28:40 crc kubenswrapper[4824]: I0121 11:28:40.518392 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Jan 21 11:28:40 crc kubenswrapper[4824]: I0121 11:28:40.551740 4824 generic.go:334] "Generic (PLEG): container finished" podID="d373f159-2b64-4dd6-9169-a96a01afcfce" containerID="82cb1335346c23ca78526b5fd5db5f17dcd0b8161c12a767fa8321e0f3ab845f" exitCode=137 Jan 21 11:28:40 crc kubenswrapper[4824]: I0121 11:28:40.551804 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Jan 21 11:28:40 crc kubenswrapper[4824]: I0121 11:28:40.552031 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 21 11:28:40 crc kubenswrapper[4824]: I0121 11:28:40.551825 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"d373f159-2b64-4dd6-9169-a96a01afcfce","Type":"ContainerDied","Data":"82cb1335346c23ca78526b5fd5db5f17dcd0b8161c12a767fa8321e0f3ab845f"} Jan 21 11:28:40 crc kubenswrapper[4824]: I0121 11:28:40.552138 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"d373f159-2b64-4dd6-9169-a96a01afcfce","Type":"ContainerDied","Data":"7669fbc7cdadc946c7f940b38d316891febbd77d056792c26182131fc73631b1"} Jan 21 11:28:40 crc kubenswrapper[4824]: I0121 11:28:40.552164 4824 scope.go:117] "RemoveContainer" containerID="82cb1335346c23ca78526b5fd5db5f17dcd0b8161c12a767fa8321e0f3ab845f" Jan 21 11:28:40 crc kubenswrapper[4824]: I0121 11:28:40.577564 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Jan 21 11:28:40 crc kubenswrapper[4824]: I0121 11:28:40.583002 4824 scope.go:117] "RemoveContainer" containerID="82cb1335346c23ca78526b5fd5db5f17dcd0b8161c12a767fa8321e0f3ab845f" Jan 21 11:28:40 crc kubenswrapper[4824]: E0121 11:28:40.583303 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"82cb1335346c23ca78526b5fd5db5f17dcd0b8161c12a767fa8321e0f3ab845f\": container with ID starting with 82cb1335346c23ca78526b5fd5db5f17dcd0b8161c12a767fa8321e0f3ab845f not found: ID does not exist" containerID="82cb1335346c23ca78526b5fd5db5f17dcd0b8161c12a767fa8321e0f3ab845f" Jan 21 11:28:40 crc kubenswrapper[4824]: I0121 11:28:40.583330 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"82cb1335346c23ca78526b5fd5db5f17dcd0b8161c12a767fa8321e0f3ab845f"} err="failed to get container status \"82cb1335346c23ca78526b5fd5db5f17dcd0b8161c12a767fa8321e0f3ab845f\": rpc error: code = NotFound desc = could not find container \"82cb1335346c23ca78526b5fd5db5f17dcd0b8161c12a767fa8321e0f3ab845f\": container with ID starting with 82cb1335346c23ca78526b5fd5db5f17dcd0b8161c12a767fa8321e0f3ab845f not found: ID does not exist" Jan 21 11:28:40 crc kubenswrapper[4824]: I0121 11:28:40.589449 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Jan 21 11:28:40 crc kubenswrapper[4824]: I0121 11:28:40.596423 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Jan 21 11:28:40 crc kubenswrapper[4824]: E0121 11:28:40.596766 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="42d7c353-4ca7-4fa0-b4da-4793ce3e36a1" containerName="ceilometer-notification-agent" Jan 21 11:28:40 crc kubenswrapper[4824]: I0121 11:28:40.596778 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="42d7c353-4ca7-4fa0-b4da-4793ce3e36a1" containerName="ceilometer-notification-agent" Jan 21 11:28:40 crc kubenswrapper[4824]: E0121 11:28:40.596791 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d373f159-2b64-4dd6-9169-a96a01afcfce" containerName="nova-cell1-novncproxy-novncproxy" Jan 21 11:28:40 crc kubenswrapper[4824]: I0121 11:28:40.596809 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="d373f159-2b64-4dd6-9169-a96a01afcfce" containerName="nova-cell1-novncproxy-novncproxy" Jan 21 11:28:40 crc kubenswrapper[4824]: E0121 11:28:40.596838 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="42d7c353-4ca7-4fa0-b4da-4793ce3e36a1" containerName="proxy-httpd" Jan 21 11:28:40 crc kubenswrapper[4824]: I0121 11:28:40.596843 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="42d7c353-4ca7-4fa0-b4da-4793ce3e36a1" containerName="proxy-httpd" Jan 21 11:28:40 crc kubenswrapper[4824]: E0121 11:28:40.596856 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="42d7c353-4ca7-4fa0-b4da-4793ce3e36a1" containerName="sg-core" Jan 21 11:28:40 crc kubenswrapper[4824]: I0121 11:28:40.596861 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="42d7c353-4ca7-4fa0-b4da-4793ce3e36a1" containerName="sg-core" Jan 21 11:28:40 crc kubenswrapper[4824]: E0121 11:28:40.596879 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="42d7c353-4ca7-4fa0-b4da-4793ce3e36a1" containerName="ceilometer-central-agent" Jan 21 11:28:40 crc kubenswrapper[4824]: I0121 11:28:40.596884 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="42d7c353-4ca7-4fa0-b4da-4793ce3e36a1" containerName="ceilometer-central-agent" Jan 21 11:28:40 crc kubenswrapper[4824]: I0121 11:28:40.597107 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="42d7c353-4ca7-4fa0-b4da-4793ce3e36a1" containerName="ceilometer-notification-agent" Jan 21 11:28:40 crc kubenswrapper[4824]: I0121 11:28:40.597124 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="42d7c353-4ca7-4fa0-b4da-4793ce3e36a1" containerName="proxy-httpd" Jan 21 11:28:40 crc kubenswrapper[4824]: I0121 11:28:40.597130 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="d373f159-2b64-4dd6-9169-a96a01afcfce" containerName="nova-cell1-novncproxy-novncproxy" Jan 21 11:28:40 crc kubenswrapper[4824]: I0121 11:28:40.597147 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="42d7c353-4ca7-4fa0-b4da-4793ce3e36a1" containerName="ceilometer-central-agent" Jan 21 11:28:40 crc kubenswrapper[4824]: I0121 11:28:40.597155 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="42d7c353-4ca7-4fa0-b4da-4793ce3e36a1" containerName="sg-core" Jan 21 11:28:40 crc kubenswrapper[4824]: I0121 11:28:40.598970 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 21 11:28:40 crc kubenswrapper[4824]: I0121 11:28:40.600471 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Jan 21 11:28:40 crc kubenswrapper[4824]: I0121 11:28:40.600545 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Jan 21 11:28:40 crc kubenswrapper[4824]: I0121 11:28:40.605250 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Jan 21 11:28:40 crc kubenswrapper[4824]: I0121 11:28:40.637636 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d373f159-2b64-4dd6-9169-a96a01afcfce-combined-ca-bundle\") pod \"d373f159-2b64-4dd6-9169-a96a01afcfce\" (UID: \"d373f159-2b64-4dd6-9169-a96a01afcfce\") " Jan 21 11:28:40 crc kubenswrapper[4824]: I0121 11:28:40.637685 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z6l4s\" (UniqueName: \"kubernetes.io/projected/d373f159-2b64-4dd6-9169-a96a01afcfce-kube-api-access-z6l4s\") pod \"d373f159-2b64-4dd6-9169-a96a01afcfce\" (UID: \"d373f159-2b64-4dd6-9169-a96a01afcfce\") " Jan 21 11:28:40 crc kubenswrapper[4824]: I0121 11:28:40.637761 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d373f159-2b64-4dd6-9169-a96a01afcfce-config-data\") pod \"d373f159-2b64-4dd6-9169-a96a01afcfce\" (UID: \"d373f159-2b64-4dd6-9169-a96a01afcfce\") " Jan 21 11:28:40 crc kubenswrapper[4824]: I0121 11:28:40.648362 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d373f159-2b64-4dd6-9169-a96a01afcfce-kube-api-access-z6l4s" (OuterVolumeSpecName: "kube-api-access-z6l4s") pod "d373f159-2b64-4dd6-9169-a96a01afcfce" (UID: "d373f159-2b64-4dd6-9169-a96a01afcfce"). InnerVolumeSpecName "kube-api-access-z6l4s". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:28:40 crc kubenswrapper[4824]: I0121 11:28:40.659138 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d373f159-2b64-4dd6-9169-a96a01afcfce-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d373f159-2b64-4dd6-9169-a96a01afcfce" (UID: "d373f159-2b64-4dd6-9169-a96a01afcfce"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:28:40 crc kubenswrapper[4824]: I0121 11:28:40.661129 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d373f159-2b64-4dd6-9169-a96a01afcfce-config-data" (OuterVolumeSpecName: "config-data") pod "d373f159-2b64-4dd6-9169-a96a01afcfce" (UID: "d373f159-2b64-4dd6-9169-a96a01afcfce"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:28:40 crc kubenswrapper[4824]: I0121 11:28:40.740342 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/55916065-772b-45ee-8871-37c229777deb-log-httpd\") pod \"ceilometer-0\" (UID: \"55916065-772b-45ee-8871-37c229777deb\") " pod="openstack/ceilometer-0" Jan 21 11:28:40 crc kubenswrapper[4824]: I0121 11:28:40.740385 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/55916065-772b-45ee-8871-37c229777deb-run-httpd\") pod \"ceilometer-0\" (UID: \"55916065-772b-45ee-8871-37c229777deb\") " pod="openstack/ceilometer-0" Jan 21 11:28:40 crc kubenswrapper[4824]: I0121 11:28:40.740544 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/55916065-772b-45ee-8871-37c229777deb-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"55916065-772b-45ee-8871-37c229777deb\") " pod="openstack/ceilometer-0" Jan 21 11:28:40 crc kubenswrapper[4824]: I0121 11:28:40.740581 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/55916065-772b-45ee-8871-37c229777deb-scripts\") pod \"ceilometer-0\" (UID: \"55916065-772b-45ee-8871-37c229777deb\") " pod="openstack/ceilometer-0" Jan 21 11:28:40 crc kubenswrapper[4824]: I0121 11:28:40.740718 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/55916065-772b-45ee-8871-37c229777deb-config-data\") pod \"ceilometer-0\" (UID: \"55916065-772b-45ee-8871-37c229777deb\") " pod="openstack/ceilometer-0" Jan 21 11:28:40 crc kubenswrapper[4824]: I0121 11:28:40.740906 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gvljc\" (UniqueName: \"kubernetes.io/projected/55916065-772b-45ee-8871-37c229777deb-kube-api-access-gvljc\") pod \"ceilometer-0\" (UID: \"55916065-772b-45ee-8871-37c229777deb\") " pod="openstack/ceilometer-0" Jan 21 11:28:40 crc kubenswrapper[4824]: I0121 11:28:40.741036 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/55916065-772b-45ee-8871-37c229777deb-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"55916065-772b-45ee-8871-37c229777deb\") " pod="openstack/ceilometer-0" Jan 21 11:28:40 crc kubenswrapper[4824]: I0121 11:28:40.741149 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d373f159-2b64-4dd6-9169-a96a01afcfce-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 21 11:28:40 crc kubenswrapper[4824]: I0121 11:28:40.741165 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z6l4s\" (UniqueName: \"kubernetes.io/projected/d373f159-2b64-4dd6-9169-a96a01afcfce-kube-api-access-z6l4s\") on node \"crc\" DevicePath \"\"" Jan 21 11:28:40 crc kubenswrapper[4824]: I0121 11:28:40.741176 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d373f159-2b64-4dd6-9169-a96a01afcfce-config-data\") on node \"crc\" DevicePath \"\"" Jan 21 11:28:40 crc kubenswrapper[4824]: I0121 11:28:40.843042 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/55916065-772b-45ee-8871-37c229777deb-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"55916065-772b-45ee-8871-37c229777deb\") " pod="openstack/ceilometer-0" Jan 21 11:28:40 crc kubenswrapper[4824]: I0121 11:28:40.843092 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/55916065-772b-45ee-8871-37c229777deb-log-httpd\") pod \"ceilometer-0\" (UID: \"55916065-772b-45ee-8871-37c229777deb\") " pod="openstack/ceilometer-0" Jan 21 11:28:40 crc kubenswrapper[4824]: I0121 11:28:40.843119 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/55916065-772b-45ee-8871-37c229777deb-run-httpd\") pod \"ceilometer-0\" (UID: \"55916065-772b-45ee-8871-37c229777deb\") " pod="openstack/ceilometer-0" Jan 21 11:28:40 crc kubenswrapper[4824]: I0121 11:28:40.843197 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/55916065-772b-45ee-8871-37c229777deb-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"55916065-772b-45ee-8871-37c229777deb\") " pod="openstack/ceilometer-0" Jan 21 11:28:40 crc kubenswrapper[4824]: I0121 11:28:40.843214 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/55916065-772b-45ee-8871-37c229777deb-scripts\") pod \"ceilometer-0\" (UID: \"55916065-772b-45ee-8871-37c229777deb\") " pod="openstack/ceilometer-0" Jan 21 11:28:40 crc kubenswrapper[4824]: I0121 11:28:40.843269 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/55916065-772b-45ee-8871-37c229777deb-config-data\") pod \"ceilometer-0\" (UID: \"55916065-772b-45ee-8871-37c229777deb\") " pod="openstack/ceilometer-0" Jan 21 11:28:40 crc kubenswrapper[4824]: I0121 11:28:40.843354 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gvljc\" (UniqueName: \"kubernetes.io/projected/55916065-772b-45ee-8871-37c229777deb-kube-api-access-gvljc\") pod \"ceilometer-0\" (UID: \"55916065-772b-45ee-8871-37c229777deb\") " pod="openstack/ceilometer-0" Jan 21 11:28:40 crc kubenswrapper[4824]: I0121 11:28:40.843504 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/55916065-772b-45ee-8871-37c229777deb-log-httpd\") pod \"ceilometer-0\" (UID: \"55916065-772b-45ee-8871-37c229777deb\") " pod="openstack/ceilometer-0" Jan 21 11:28:40 crc kubenswrapper[4824]: I0121 11:28:40.843578 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/55916065-772b-45ee-8871-37c229777deb-run-httpd\") pod \"ceilometer-0\" (UID: \"55916065-772b-45ee-8871-37c229777deb\") " pod="openstack/ceilometer-0" Jan 21 11:28:40 crc kubenswrapper[4824]: I0121 11:28:40.846275 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/55916065-772b-45ee-8871-37c229777deb-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"55916065-772b-45ee-8871-37c229777deb\") " pod="openstack/ceilometer-0" Jan 21 11:28:40 crc kubenswrapper[4824]: I0121 11:28:40.846464 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/55916065-772b-45ee-8871-37c229777deb-scripts\") pod \"ceilometer-0\" (UID: \"55916065-772b-45ee-8871-37c229777deb\") " pod="openstack/ceilometer-0" Jan 21 11:28:40 crc kubenswrapper[4824]: I0121 11:28:40.846524 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/55916065-772b-45ee-8871-37c229777deb-config-data\") pod \"ceilometer-0\" (UID: \"55916065-772b-45ee-8871-37c229777deb\") " pod="openstack/ceilometer-0" Jan 21 11:28:40 crc kubenswrapper[4824]: I0121 11:28:40.846536 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/55916065-772b-45ee-8871-37c229777deb-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"55916065-772b-45ee-8871-37c229777deb\") " pod="openstack/ceilometer-0" Jan 21 11:28:40 crc kubenswrapper[4824]: I0121 11:28:40.856222 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gvljc\" (UniqueName: \"kubernetes.io/projected/55916065-772b-45ee-8871-37c229777deb-kube-api-access-gvljc\") pod \"ceilometer-0\" (UID: \"55916065-772b-45ee-8871-37c229777deb\") " pod="openstack/ceilometer-0" Jan 21 11:28:40 crc kubenswrapper[4824]: I0121 11:28:40.876517 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Jan 21 11:28:40 crc kubenswrapper[4824]: I0121 11:28:40.886384 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Jan 21 11:28:40 crc kubenswrapper[4824]: I0121 11:28:40.906733 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Jan 21 11:28:40 crc kubenswrapper[4824]: I0121 11:28:40.908573 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Jan 21 11:28:40 crc kubenswrapper[4824]: I0121 11:28:40.915478 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Jan 21 11:28:40 crc kubenswrapper[4824]: I0121 11:28:40.915622 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Jan 21 11:28:40 crc kubenswrapper[4824]: I0121 11:28:40.915890 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Jan 21 11:28:40 crc kubenswrapper[4824]: I0121 11:28:40.916086 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Jan 21 11:28:40 crc kubenswrapper[4824]: I0121 11:28:40.922705 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 21 11:28:41 crc kubenswrapper[4824]: I0121 11:28:41.047270 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8e42e9d6-485f-4b97-b049-316cd47af30e-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"8e42e9d6-485f-4b97-b049-316cd47af30e\") " pod="openstack/nova-cell1-novncproxy-0" Jan 21 11:28:41 crc kubenswrapper[4824]: I0121 11:28:41.047555 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/8e42e9d6-485f-4b97-b049-316cd47af30e-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"8e42e9d6-485f-4b97-b049-316cd47af30e\") " pod="openstack/nova-cell1-novncproxy-0" Jan 21 11:28:41 crc kubenswrapper[4824]: I0121 11:28:41.047639 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8e42e9d6-485f-4b97-b049-316cd47af30e-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"8e42e9d6-485f-4b97-b049-316cd47af30e\") " pod="openstack/nova-cell1-novncproxy-0" Jan 21 11:28:41 crc kubenswrapper[4824]: I0121 11:28:41.047840 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m2dmc\" (UniqueName: \"kubernetes.io/projected/8e42e9d6-485f-4b97-b049-316cd47af30e-kube-api-access-m2dmc\") pod \"nova-cell1-novncproxy-0\" (UID: \"8e42e9d6-485f-4b97-b049-316cd47af30e\") " pod="openstack/nova-cell1-novncproxy-0" Jan 21 11:28:41 crc kubenswrapper[4824]: I0121 11:28:41.047966 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/8e42e9d6-485f-4b97-b049-316cd47af30e-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"8e42e9d6-485f-4b97-b049-316cd47af30e\") " pod="openstack/nova-cell1-novncproxy-0" Jan 21 11:28:41 crc kubenswrapper[4824]: I0121 11:28:41.149265 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m2dmc\" (UniqueName: \"kubernetes.io/projected/8e42e9d6-485f-4b97-b049-316cd47af30e-kube-api-access-m2dmc\") pod \"nova-cell1-novncproxy-0\" (UID: \"8e42e9d6-485f-4b97-b049-316cd47af30e\") " pod="openstack/nova-cell1-novncproxy-0" Jan 21 11:28:41 crc kubenswrapper[4824]: I0121 11:28:41.149328 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/8e42e9d6-485f-4b97-b049-316cd47af30e-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"8e42e9d6-485f-4b97-b049-316cd47af30e\") " pod="openstack/nova-cell1-novncproxy-0" Jan 21 11:28:41 crc kubenswrapper[4824]: I0121 11:28:41.149378 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8e42e9d6-485f-4b97-b049-316cd47af30e-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"8e42e9d6-485f-4b97-b049-316cd47af30e\") " pod="openstack/nova-cell1-novncproxy-0" Jan 21 11:28:41 crc kubenswrapper[4824]: I0121 11:28:41.149450 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/8e42e9d6-485f-4b97-b049-316cd47af30e-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"8e42e9d6-485f-4b97-b049-316cd47af30e\") " pod="openstack/nova-cell1-novncproxy-0" Jan 21 11:28:41 crc kubenswrapper[4824]: I0121 11:28:41.149510 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8e42e9d6-485f-4b97-b049-316cd47af30e-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"8e42e9d6-485f-4b97-b049-316cd47af30e\") " pod="openstack/nova-cell1-novncproxy-0" Jan 21 11:28:41 crc kubenswrapper[4824]: I0121 11:28:41.152563 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/8e42e9d6-485f-4b97-b049-316cd47af30e-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"8e42e9d6-485f-4b97-b049-316cd47af30e\") " pod="openstack/nova-cell1-novncproxy-0" Jan 21 11:28:41 crc kubenswrapper[4824]: I0121 11:28:41.152635 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8e42e9d6-485f-4b97-b049-316cd47af30e-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"8e42e9d6-485f-4b97-b049-316cd47af30e\") " pod="openstack/nova-cell1-novncproxy-0" Jan 21 11:28:41 crc kubenswrapper[4824]: I0121 11:28:41.152852 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/8e42e9d6-485f-4b97-b049-316cd47af30e-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"8e42e9d6-485f-4b97-b049-316cd47af30e\") " pod="openstack/nova-cell1-novncproxy-0" Jan 21 11:28:41 crc kubenswrapper[4824]: I0121 11:28:41.153758 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8e42e9d6-485f-4b97-b049-316cd47af30e-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"8e42e9d6-485f-4b97-b049-316cd47af30e\") " pod="openstack/nova-cell1-novncproxy-0" Jan 21 11:28:41 crc kubenswrapper[4824]: I0121 11:28:41.161925 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m2dmc\" (UniqueName: \"kubernetes.io/projected/8e42e9d6-485f-4b97-b049-316cd47af30e-kube-api-access-m2dmc\") pod \"nova-cell1-novncproxy-0\" (UID: \"8e42e9d6-485f-4b97-b049-316cd47af30e\") " pod="openstack/nova-cell1-novncproxy-0" Jan 21 11:28:41 crc kubenswrapper[4824]: I0121 11:28:41.230752 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Jan 21 11:28:41 crc kubenswrapper[4824]: W0121 11:28:41.318586 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod55916065_772b_45ee_8871_37c229777deb.slice/crio-65e4d9ce199d66a5d4b888b55edee4667e5ba88065fc68ccd26d1ad79dbffbb0 WatchSource:0}: Error finding container 65e4d9ce199d66a5d4b888b55edee4667e5ba88065fc68ccd26d1ad79dbffbb0: Status 404 returned error can't find the container with id 65e4d9ce199d66a5d4b888b55edee4667e5ba88065fc68ccd26d1ad79dbffbb0 Jan 21 11:28:41 crc kubenswrapper[4824]: I0121 11:28:41.319893 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Jan 21 11:28:41 crc kubenswrapper[4824]: I0121 11:28:41.559513 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"55916065-772b-45ee-8871-37c229777deb","Type":"ContainerStarted","Data":"65e4d9ce199d66a5d4b888b55edee4667e5ba88065fc68ccd26d1ad79dbffbb0"} Jan 21 11:28:41 crc kubenswrapper[4824]: I0121 11:28:41.628614 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Jan 21 11:28:41 crc kubenswrapper[4824]: I0121 11:28:41.948595 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Jan 21 11:28:42 crc kubenswrapper[4824]: I0121 11:28:42.057986 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="42d7c353-4ca7-4fa0-b4da-4793ce3e36a1" path="/var/lib/kubelet/pods/42d7c353-4ca7-4fa0-b4da-4793ce3e36a1/volumes" Jan 21 11:28:42 crc kubenswrapper[4824]: I0121 11:28:42.058818 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d373f159-2b64-4dd6-9169-a96a01afcfce" path="/var/lib/kubelet/pods/d373f159-2b64-4dd6-9169-a96a01afcfce/volumes" Jan 21 11:28:42 crc kubenswrapper[4824]: I0121 11:28:42.065784 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/75cf0b57-bdd7-4b26-8d52-5acfbe9d520d-logs\") pod \"75cf0b57-bdd7-4b26-8d52-5acfbe9d520d\" (UID: \"75cf0b57-bdd7-4b26-8d52-5acfbe9d520d\") " Jan 21 11:28:42 crc kubenswrapper[4824]: I0121 11:28:42.065839 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/75cf0b57-bdd7-4b26-8d52-5acfbe9d520d-config-data\") pod \"75cf0b57-bdd7-4b26-8d52-5acfbe9d520d\" (UID: \"75cf0b57-bdd7-4b26-8d52-5acfbe9d520d\") " Jan 21 11:28:42 crc kubenswrapper[4824]: I0121 11:28:42.065895 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/75cf0b57-bdd7-4b26-8d52-5acfbe9d520d-combined-ca-bundle\") pod \"75cf0b57-bdd7-4b26-8d52-5acfbe9d520d\" (UID: \"75cf0b57-bdd7-4b26-8d52-5acfbe9d520d\") " Jan 21 11:28:42 crc kubenswrapper[4824]: I0121 11:28:42.065925 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n5q8r\" (UniqueName: \"kubernetes.io/projected/75cf0b57-bdd7-4b26-8d52-5acfbe9d520d-kube-api-access-n5q8r\") pod \"75cf0b57-bdd7-4b26-8d52-5acfbe9d520d\" (UID: \"75cf0b57-bdd7-4b26-8d52-5acfbe9d520d\") " Jan 21 11:28:42 crc kubenswrapper[4824]: I0121 11:28:42.066410 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/75cf0b57-bdd7-4b26-8d52-5acfbe9d520d-logs" (OuterVolumeSpecName: "logs") pod "75cf0b57-bdd7-4b26-8d52-5acfbe9d520d" (UID: "75cf0b57-bdd7-4b26-8d52-5acfbe9d520d"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:28:42 crc kubenswrapper[4824]: I0121 11:28:42.066688 4824 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/75cf0b57-bdd7-4b26-8d52-5acfbe9d520d-logs\") on node \"crc\" DevicePath \"\"" Jan 21 11:28:42 crc kubenswrapper[4824]: I0121 11:28:42.070046 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/75cf0b57-bdd7-4b26-8d52-5acfbe9d520d-kube-api-access-n5q8r" (OuterVolumeSpecName: "kube-api-access-n5q8r") pod "75cf0b57-bdd7-4b26-8d52-5acfbe9d520d" (UID: "75cf0b57-bdd7-4b26-8d52-5acfbe9d520d"). InnerVolumeSpecName "kube-api-access-n5q8r". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:28:42 crc kubenswrapper[4824]: I0121 11:28:42.087096 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/75cf0b57-bdd7-4b26-8d52-5acfbe9d520d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "75cf0b57-bdd7-4b26-8d52-5acfbe9d520d" (UID: "75cf0b57-bdd7-4b26-8d52-5acfbe9d520d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:28:42 crc kubenswrapper[4824]: I0121 11:28:42.088110 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/75cf0b57-bdd7-4b26-8d52-5acfbe9d520d-config-data" (OuterVolumeSpecName: "config-data") pod "75cf0b57-bdd7-4b26-8d52-5acfbe9d520d" (UID: "75cf0b57-bdd7-4b26-8d52-5acfbe9d520d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:28:42 crc kubenswrapper[4824]: I0121 11:28:42.168193 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/75cf0b57-bdd7-4b26-8d52-5acfbe9d520d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 21 11:28:42 crc kubenswrapper[4824]: I0121 11:28:42.168221 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n5q8r\" (UniqueName: \"kubernetes.io/projected/75cf0b57-bdd7-4b26-8d52-5acfbe9d520d-kube-api-access-n5q8r\") on node \"crc\" DevicePath \"\"" Jan 21 11:28:42 crc kubenswrapper[4824]: I0121 11:28:42.168233 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/75cf0b57-bdd7-4b26-8d52-5acfbe9d520d-config-data\") on node \"crc\" DevicePath \"\"" Jan 21 11:28:42 crc kubenswrapper[4824]: I0121 11:28:42.569164 4824 generic.go:334] "Generic (PLEG): container finished" podID="75cf0b57-bdd7-4b26-8d52-5acfbe9d520d" containerID="ca3b941babbbd07e235f5eab12e05e43e0ac2f0218981d4b14fc63386c42e254" exitCode=0 Jan 21 11:28:42 crc kubenswrapper[4824]: I0121 11:28:42.569230 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Jan 21 11:28:42 crc kubenswrapper[4824]: I0121 11:28:42.569255 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"75cf0b57-bdd7-4b26-8d52-5acfbe9d520d","Type":"ContainerDied","Data":"ca3b941babbbd07e235f5eab12e05e43e0ac2f0218981d4b14fc63386c42e254"} Jan 21 11:28:42 crc kubenswrapper[4824]: I0121 11:28:42.570267 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"75cf0b57-bdd7-4b26-8d52-5acfbe9d520d","Type":"ContainerDied","Data":"728845c124e990e3b5128ca1bd8dd69e1a52e3b9f1c5598a8738ea82598401a9"} Jan 21 11:28:42 crc kubenswrapper[4824]: I0121 11:28:42.570292 4824 scope.go:117] "RemoveContainer" containerID="ca3b941babbbd07e235f5eab12e05e43e0ac2f0218981d4b14fc63386c42e254" Jan 21 11:28:42 crc kubenswrapper[4824]: I0121 11:28:42.571419 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"55916065-772b-45ee-8871-37c229777deb","Type":"ContainerStarted","Data":"f6009500618f69ecbe25f69878a4c52dce4019a903ece88ba0ef9f2b52fe0e5f"} Jan 21 11:28:42 crc kubenswrapper[4824]: I0121 11:28:42.573161 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"8e42e9d6-485f-4b97-b049-316cd47af30e","Type":"ContainerStarted","Data":"c50e2f4bbf17ec122e748d38d9b57650b3e16cc7f0d18cca04bd8713f5473e55"} Jan 21 11:28:42 crc kubenswrapper[4824]: I0121 11:28:42.573191 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"8e42e9d6-485f-4b97-b049-316cd47af30e","Type":"ContainerStarted","Data":"faf17c84aa6350cbbfe7b1a2a465fe70e978a750b6dbf0fa539472cf64b43203"} Jan 21 11:28:42 crc kubenswrapper[4824]: I0121 11:28:42.591727 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.591702482 podStartE2EDuration="2.591702482s" podCreationTimestamp="2026-01-21 11:28:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:28:42.587873542 +0000 UTC m=+1064.880902825" watchObservedRunningTime="2026-01-21 11:28:42.591702482 +0000 UTC m=+1064.884731774" Jan 21 11:28:42 crc kubenswrapper[4824]: I0121 11:28:42.598360 4824 scope.go:117] "RemoveContainer" containerID="7d8c20082288cf600e39dc7a0f1bd5ac8ba09165ca35559ccd6592d1b0add249" Jan 21 11:28:42 crc kubenswrapper[4824]: I0121 11:28:42.603968 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Jan 21 11:28:42 crc kubenswrapper[4824]: I0121 11:28:42.621374 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Jan 21 11:28:42 crc kubenswrapper[4824]: I0121 11:28:42.623971 4824 scope.go:117] "RemoveContainer" containerID="ca3b941babbbd07e235f5eab12e05e43e0ac2f0218981d4b14fc63386c42e254" Jan 21 11:28:42 crc kubenswrapper[4824]: E0121 11:28:42.624792 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ca3b941babbbd07e235f5eab12e05e43e0ac2f0218981d4b14fc63386c42e254\": container with ID starting with ca3b941babbbd07e235f5eab12e05e43e0ac2f0218981d4b14fc63386c42e254 not found: ID does not exist" containerID="ca3b941babbbd07e235f5eab12e05e43e0ac2f0218981d4b14fc63386c42e254" Jan 21 11:28:42 crc kubenswrapper[4824]: I0121 11:28:42.624914 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ca3b941babbbd07e235f5eab12e05e43e0ac2f0218981d4b14fc63386c42e254"} err="failed to get container status \"ca3b941babbbd07e235f5eab12e05e43e0ac2f0218981d4b14fc63386c42e254\": rpc error: code = NotFound desc = could not find container \"ca3b941babbbd07e235f5eab12e05e43e0ac2f0218981d4b14fc63386c42e254\": container with ID starting with ca3b941babbbd07e235f5eab12e05e43e0ac2f0218981d4b14fc63386c42e254 not found: ID does not exist" Jan 21 11:28:42 crc kubenswrapper[4824]: I0121 11:28:42.625017 4824 scope.go:117] "RemoveContainer" containerID="7d8c20082288cf600e39dc7a0f1bd5ac8ba09165ca35559ccd6592d1b0add249" Jan 21 11:28:42 crc kubenswrapper[4824]: E0121 11:28:42.625469 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7d8c20082288cf600e39dc7a0f1bd5ac8ba09165ca35559ccd6592d1b0add249\": container with ID starting with 7d8c20082288cf600e39dc7a0f1bd5ac8ba09165ca35559ccd6592d1b0add249 not found: ID does not exist" containerID="7d8c20082288cf600e39dc7a0f1bd5ac8ba09165ca35559ccd6592d1b0add249" Jan 21 11:28:42 crc kubenswrapper[4824]: I0121 11:28:42.625545 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7d8c20082288cf600e39dc7a0f1bd5ac8ba09165ca35559ccd6592d1b0add249"} err="failed to get container status \"7d8c20082288cf600e39dc7a0f1bd5ac8ba09165ca35559ccd6592d1b0add249\": rpc error: code = NotFound desc = could not find container \"7d8c20082288cf600e39dc7a0f1bd5ac8ba09165ca35559ccd6592d1b0add249\": container with ID starting with 7d8c20082288cf600e39dc7a0f1bd5ac8ba09165ca35559ccd6592d1b0add249 not found: ID does not exist" Jan 21 11:28:42 crc kubenswrapper[4824]: I0121 11:28:42.634379 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Jan 21 11:28:42 crc kubenswrapper[4824]: E0121 11:28:42.634750 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="75cf0b57-bdd7-4b26-8d52-5acfbe9d520d" containerName="nova-api-api" Jan 21 11:28:42 crc kubenswrapper[4824]: I0121 11:28:42.634767 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="75cf0b57-bdd7-4b26-8d52-5acfbe9d520d" containerName="nova-api-api" Jan 21 11:28:42 crc kubenswrapper[4824]: E0121 11:28:42.634784 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="75cf0b57-bdd7-4b26-8d52-5acfbe9d520d" containerName="nova-api-log" Jan 21 11:28:42 crc kubenswrapper[4824]: I0121 11:28:42.634791 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="75cf0b57-bdd7-4b26-8d52-5acfbe9d520d" containerName="nova-api-log" Jan 21 11:28:42 crc kubenswrapper[4824]: I0121 11:28:42.634990 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="75cf0b57-bdd7-4b26-8d52-5acfbe9d520d" containerName="nova-api-api" Jan 21 11:28:42 crc kubenswrapper[4824]: I0121 11:28:42.635013 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="75cf0b57-bdd7-4b26-8d52-5acfbe9d520d" containerName="nova-api-log" Jan 21 11:28:42 crc kubenswrapper[4824]: I0121 11:28:42.635888 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Jan 21 11:28:42 crc kubenswrapper[4824]: I0121 11:28:42.638380 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Jan 21 11:28:42 crc kubenswrapper[4824]: I0121 11:28:42.638545 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Jan 21 11:28:42 crc kubenswrapper[4824]: I0121 11:28:42.638664 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Jan 21 11:28:42 crc kubenswrapper[4824]: I0121 11:28:42.640196 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Jan 21 11:28:42 crc kubenswrapper[4824]: I0121 11:28:42.777316 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ad597758-1f2c-4f84-8035-59fa20e68084-internal-tls-certs\") pod \"nova-api-0\" (UID: \"ad597758-1f2c-4f84-8035-59fa20e68084\") " pod="openstack/nova-api-0" Jan 21 11:28:42 crc kubenswrapper[4824]: I0121 11:28:42.777384 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-22h96\" (UniqueName: \"kubernetes.io/projected/ad597758-1f2c-4f84-8035-59fa20e68084-kube-api-access-22h96\") pod \"nova-api-0\" (UID: \"ad597758-1f2c-4f84-8035-59fa20e68084\") " pod="openstack/nova-api-0" Jan 21 11:28:42 crc kubenswrapper[4824]: I0121 11:28:42.777432 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ad597758-1f2c-4f84-8035-59fa20e68084-config-data\") pod \"nova-api-0\" (UID: \"ad597758-1f2c-4f84-8035-59fa20e68084\") " pod="openstack/nova-api-0" Jan 21 11:28:42 crc kubenswrapper[4824]: I0121 11:28:42.777454 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ad597758-1f2c-4f84-8035-59fa20e68084-public-tls-certs\") pod \"nova-api-0\" (UID: \"ad597758-1f2c-4f84-8035-59fa20e68084\") " pod="openstack/nova-api-0" Jan 21 11:28:42 crc kubenswrapper[4824]: I0121 11:28:42.777472 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ad597758-1f2c-4f84-8035-59fa20e68084-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"ad597758-1f2c-4f84-8035-59fa20e68084\") " pod="openstack/nova-api-0" Jan 21 11:28:42 crc kubenswrapper[4824]: I0121 11:28:42.777547 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ad597758-1f2c-4f84-8035-59fa20e68084-logs\") pod \"nova-api-0\" (UID: \"ad597758-1f2c-4f84-8035-59fa20e68084\") " pod="openstack/nova-api-0" Jan 21 11:28:42 crc kubenswrapper[4824]: I0121 11:28:42.878942 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ad597758-1f2c-4f84-8035-59fa20e68084-internal-tls-certs\") pod \"nova-api-0\" (UID: \"ad597758-1f2c-4f84-8035-59fa20e68084\") " pod="openstack/nova-api-0" Jan 21 11:28:42 crc kubenswrapper[4824]: I0121 11:28:42.879018 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-22h96\" (UniqueName: \"kubernetes.io/projected/ad597758-1f2c-4f84-8035-59fa20e68084-kube-api-access-22h96\") pod \"nova-api-0\" (UID: \"ad597758-1f2c-4f84-8035-59fa20e68084\") " pod="openstack/nova-api-0" Jan 21 11:28:42 crc kubenswrapper[4824]: I0121 11:28:42.879070 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ad597758-1f2c-4f84-8035-59fa20e68084-config-data\") pod \"nova-api-0\" (UID: \"ad597758-1f2c-4f84-8035-59fa20e68084\") " pod="openstack/nova-api-0" Jan 21 11:28:42 crc kubenswrapper[4824]: I0121 11:28:42.879093 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ad597758-1f2c-4f84-8035-59fa20e68084-public-tls-certs\") pod \"nova-api-0\" (UID: \"ad597758-1f2c-4f84-8035-59fa20e68084\") " pod="openstack/nova-api-0" Jan 21 11:28:42 crc kubenswrapper[4824]: I0121 11:28:42.879113 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ad597758-1f2c-4f84-8035-59fa20e68084-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"ad597758-1f2c-4f84-8035-59fa20e68084\") " pod="openstack/nova-api-0" Jan 21 11:28:42 crc kubenswrapper[4824]: I0121 11:28:42.879189 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ad597758-1f2c-4f84-8035-59fa20e68084-logs\") pod \"nova-api-0\" (UID: \"ad597758-1f2c-4f84-8035-59fa20e68084\") " pod="openstack/nova-api-0" Jan 21 11:28:42 crc kubenswrapper[4824]: I0121 11:28:42.879540 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ad597758-1f2c-4f84-8035-59fa20e68084-logs\") pod \"nova-api-0\" (UID: \"ad597758-1f2c-4f84-8035-59fa20e68084\") " pod="openstack/nova-api-0" Jan 21 11:28:42 crc kubenswrapper[4824]: I0121 11:28:42.882198 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ad597758-1f2c-4f84-8035-59fa20e68084-internal-tls-certs\") pod \"nova-api-0\" (UID: \"ad597758-1f2c-4f84-8035-59fa20e68084\") " pod="openstack/nova-api-0" Jan 21 11:28:42 crc kubenswrapper[4824]: I0121 11:28:42.882763 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ad597758-1f2c-4f84-8035-59fa20e68084-config-data\") pod \"nova-api-0\" (UID: \"ad597758-1f2c-4f84-8035-59fa20e68084\") " pod="openstack/nova-api-0" Jan 21 11:28:42 crc kubenswrapper[4824]: I0121 11:28:42.882891 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ad597758-1f2c-4f84-8035-59fa20e68084-public-tls-certs\") pod \"nova-api-0\" (UID: \"ad597758-1f2c-4f84-8035-59fa20e68084\") " pod="openstack/nova-api-0" Jan 21 11:28:42 crc kubenswrapper[4824]: I0121 11:28:42.883222 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ad597758-1f2c-4f84-8035-59fa20e68084-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"ad597758-1f2c-4f84-8035-59fa20e68084\") " pod="openstack/nova-api-0" Jan 21 11:28:42 crc kubenswrapper[4824]: I0121 11:28:42.891771 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-22h96\" (UniqueName: \"kubernetes.io/projected/ad597758-1f2c-4f84-8035-59fa20e68084-kube-api-access-22h96\") pod \"nova-api-0\" (UID: \"ad597758-1f2c-4f84-8035-59fa20e68084\") " pod="openstack/nova-api-0" Jan 21 11:28:42 crc kubenswrapper[4824]: I0121 11:28:42.948995 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Jan 21 11:28:43 crc kubenswrapper[4824]: I0121 11:28:43.348527 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Jan 21 11:28:43 crc kubenswrapper[4824]: W0121 11:28:43.351291 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podad597758_1f2c_4f84_8035_59fa20e68084.slice/crio-04682cea27407b0d473ba48a7b6b0c26d36ec7ff8680fb81116bbb44f989542f WatchSource:0}: Error finding container 04682cea27407b0d473ba48a7b6b0c26d36ec7ff8680fb81116bbb44f989542f: Status 404 returned error can't find the container with id 04682cea27407b0d473ba48a7b6b0c26d36ec7ff8680fb81116bbb44f989542f Jan 21 11:28:43 crc kubenswrapper[4824]: I0121 11:28:43.599398 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"55916065-772b-45ee-8871-37c229777deb","Type":"ContainerStarted","Data":"ceacb0570b30b00232645caa9a72628c88bd5083b2072630b2e8ba9033c27f4a"} Jan 21 11:28:43 crc kubenswrapper[4824]: I0121 11:28:43.599436 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"55916065-772b-45ee-8871-37c229777deb","Type":"ContainerStarted","Data":"ee73eb7a1ab0880b735d27e1db094d3fe0d66679a0480addcb5e49e67621c253"} Jan 21 11:28:43 crc kubenswrapper[4824]: I0121 11:28:43.601440 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ad597758-1f2c-4f84-8035-59fa20e68084","Type":"ContainerStarted","Data":"4a99cc0942c83c20256fb6754c20ac3db4e17b12130bc22c771bd7cc4942d218"} Jan 21 11:28:43 crc kubenswrapper[4824]: I0121 11:28:43.601488 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ad597758-1f2c-4f84-8035-59fa20e68084","Type":"ContainerStarted","Data":"04682cea27407b0d473ba48a7b6b0c26d36ec7ff8680fb81116bbb44f989542f"} Jan 21 11:28:44 crc kubenswrapper[4824]: I0121 11:28:44.056698 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="75cf0b57-bdd7-4b26-8d52-5acfbe9d520d" path="/var/lib/kubelet/pods/75cf0b57-bdd7-4b26-8d52-5acfbe9d520d/volumes" Jan 21 11:28:44 crc kubenswrapper[4824]: I0121 11:28:44.610272 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"55916065-772b-45ee-8871-37c229777deb","Type":"ContainerStarted","Data":"0949e2b629f2a233cd433291c41e4a65cbb4c871e7cd93788379c3836b2e6e74"} Jan 21 11:28:44 crc kubenswrapper[4824]: I0121 11:28:44.610573 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Jan 21 11:28:44 crc kubenswrapper[4824]: I0121 11:28:44.612343 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ad597758-1f2c-4f84-8035-59fa20e68084","Type":"ContainerStarted","Data":"ec2b52af19a4c108e2ffb7356d1ff797d5745c69b97673e2b9ad650f81683331"} Jan 21 11:28:44 crc kubenswrapper[4824]: I0121 11:28:44.631163 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.527669825 podStartE2EDuration="4.631148042s" podCreationTimestamp="2026-01-21 11:28:40 +0000 UTC" firstStartedPulling="2026-01-21 11:28:41.320567052 +0000 UTC m=+1063.613596344" lastFinishedPulling="2026-01-21 11:28:44.424045258 +0000 UTC m=+1066.717074561" observedRunningTime="2026-01-21 11:28:44.625335391 +0000 UTC m=+1066.918364684" watchObservedRunningTime="2026-01-21 11:28:44.631148042 +0000 UTC m=+1066.924177333" Jan 21 11:28:44 crc kubenswrapper[4824]: I0121 11:28:44.637226 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.637213428 podStartE2EDuration="2.637213428s" podCreationTimestamp="2026-01-21 11:28:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:28:44.636417998 +0000 UTC m=+1066.929447290" watchObservedRunningTime="2026-01-21 11:28:44.637213428 +0000 UTC m=+1066.930242720" Jan 21 11:28:46 crc kubenswrapper[4824]: I0121 11:28:46.199093 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6b7bbf7cf9-tmvm9" Jan 21 11:28:46 crc kubenswrapper[4824]: I0121 11:28:46.233844 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Jan 21 11:28:46 crc kubenswrapper[4824]: I0121 11:28:46.245319 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-9b86998b5-lwrg9"] Jan 21 11:28:46 crc kubenswrapper[4824]: I0121 11:28:46.245490 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-9b86998b5-lwrg9" podUID="8247185f-987d-4f6e-b633-042907078e39" containerName="dnsmasq-dns" containerID="cri-o://32fb68e8168ce229b5eee2857a8b6f19fa0565f1c232d572bde25ae83e70fb37" gracePeriod=10 Jan 21 11:28:46 crc kubenswrapper[4824]: I0121 11:28:46.627075 4824 generic.go:334] "Generic (PLEG): container finished" podID="8247185f-987d-4f6e-b633-042907078e39" containerID="32fb68e8168ce229b5eee2857a8b6f19fa0565f1c232d572bde25ae83e70fb37" exitCode=0 Jan 21 11:28:46 crc kubenswrapper[4824]: I0121 11:28:46.627290 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-9b86998b5-lwrg9" event={"ID":"8247185f-987d-4f6e-b633-042907078e39","Type":"ContainerDied","Data":"32fb68e8168ce229b5eee2857a8b6f19fa0565f1c232d572bde25ae83e70fb37"} Jan 21 11:28:46 crc kubenswrapper[4824]: I0121 11:28:46.627480 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-9b86998b5-lwrg9" event={"ID":"8247185f-987d-4f6e-b633-042907078e39","Type":"ContainerDied","Data":"cee4d42d6e8f113c3a920c0a14aa2caa944d9803b04e59596b4d760c6b1d3e91"} Jan 21 11:28:46 crc kubenswrapper[4824]: I0121 11:28:46.627566 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cee4d42d6e8f113c3a920c0a14aa2caa944d9803b04e59596b4d760c6b1d3e91" Jan 21 11:28:46 crc kubenswrapper[4824]: I0121 11:28:46.659485 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-9b86998b5-lwrg9" Jan 21 11:28:46 crc kubenswrapper[4824]: I0121 11:28:46.745108 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8247185f-987d-4f6e-b633-042907078e39-dns-svc\") pod \"8247185f-987d-4f6e-b633-042907078e39\" (UID: \"8247185f-987d-4f6e-b633-042907078e39\") " Jan 21 11:28:46 crc kubenswrapper[4824]: I0121 11:28:46.745919 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8247185f-987d-4f6e-b633-042907078e39-ovsdbserver-sb\") pod \"8247185f-987d-4f6e-b633-042907078e39\" (UID: \"8247185f-987d-4f6e-b633-042907078e39\") " Jan 21 11:28:46 crc kubenswrapper[4824]: I0121 11:28:46.746095 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9kz5l\" (UniqueName: \"kubernetes.io/projected/8247185f-987d-4f6e-b633-042907078e39-kube-api-access-9kz5l\") pod \"8247185f-987d-4f6e-b633-042907078e39\" (UID: \"8247185f-987d-4f6e-b633-042907078e39\") " Jan 21 11:28:46 crc kubenswrapper[4824]: I0121 11:28:46.746196 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8247185f-987d-4f6e-b633-042907078e39-dns-swift-storage-0\") pod \"8247185f-987d-4f6e-b633-042907078e39\" (UID: \"8247185f-987d-4f6e-b633-042907078e39\") " Jan 21 11:28:46 crc kubenswrapper[4824]: I0121 11:28:46.746283 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8247185f-987d-4f6e-b633-042907078e39-config\") pod \"8247185f-987d-4f6e-b633-042907078e39\" (UID: \"8247185f-987d-4f6e-b633-042907078e39\") " Jan 21 11:28:46 crc kubenswrapper[4824]: I0121 11:28:46.746423 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8247185f-987d-4f6e-b633-042907078e39-ovsdbserver-nb\") pod \"8247185f-987d-4f6e-b633-042907078e39\" (UID: \"8247185f-987d-4f6e-b633-042907078e39\") " Jan 21 11:28:46 crc kubenswrapper[4824]: I0121 11:28:46.754136 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8247185f-987d-4f6e-b633-042907078e39-kube-api-access-9kz5l" (OuterVolumeSpecName: "kube-api-access-9kz5l") pod "8247185f-987d-4f6e-b633-042907078e39" (UID: "8247185f-987d-4f6e-b633-042907078e39"). InnerVolumeSpecName "kube-api-access-9kz5l". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:28:46 crc kubenswrapper[4824]: I0121 11:28:46.789709 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8247185f-987d-4f6e-b633-042907078e39-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "8247185f-987d-4f6e-b633-042907078e39" (UID: "8247185f-987d-4f6e-b633-042907078e39"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:28:46 crc kubenswrapper[4824]: I0121 11:28:46.791873 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8247185f-987d-4f6e-b633-042907078e39-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "8247185f-987d-4f6e-b633-042907078e39" (UID: "8247185f-987d-4f6e-b633-042907078e39"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:28:46 crc kubenswrapper[4824]: I0121 11:28:46.793064 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8247185f-987d-4f6e-b633-042907078e39-config" (OuterVolumeSpecName: "config") pod "8247185f-987d-4f6e-b633-042907078e39" (UID: "8247185f-987d-4f6e-b633-042907078e39"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:28:46 crc kubenswrapper[4824]: I0121 11:28:46.796752 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8247185f-987d-4f6e-b633-042907078e39-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "8247185f-987d-4f6e-b633-042907078e39" (UID: "8247185f-987d-4f6e-b633-042907078e39"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:28:46 crc kubenswrapper[4824]: I0121 11:28:46.798712 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8247185f-987d-4f6e-b633-042907078e39-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "8247185f-987d-4f6e-b633-042907078e39" (UID: "8247185f-987d-4f6e-b633-042907078e39"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:28:46 crc kubenswrapper[4824]: I0121 11:28:46.848382 4824 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8247185f-987d-4f6e-b633-042907078e39-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Jan 21 11:28:46 crc kubenswrapper[4824]: I0121 11:28:46.848619 4824 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8247185f-987d-4f6e-b633-042907078e39-dns-svc\") on node \"crc\" DevicePath \"\"" Jan 21 11:28:46 crc kubenswrapper[4824]: I0121 11:28:46.848632 4824 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8247185f-987d-4f6e-b633-042907078e39-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Jan 21 11:28:46 crc kubenswrapper[4824]: I0121 11:28:46.848642 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9kz5l\" (UniqueName: \"kubernetes.io/projected/8247185f-987d-4f6e-b633-042907078e39-kube-api-access-9kz5l\") on node \"crc\" DevicePath \"\"" Jan 21 11:28:46 crc kubenswrapper[4824]: I0121 11:28:46.848651 4824 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8247185f-987d-4f6e-b633-042907078e39-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Jan 21 11:28:46 crc kubenswrapper[4824]: I0121 11:28:46.848659 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8247185f-987d-4f6e-b633-042907078e39-config\") on node \"crc\" DevicePath \"\"" Jan 21 11:28:47 crc kubenswrapper[4824]: I0121 11:28:47.633854 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-9b86998b5-lwrg9" Jan 21 11:28:47 crc kubenswrapper[4824]: I0121 11:28:47.665486 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-9b86998b5-lwrg9"] Jan 21 11:28:47 crc kubenswrapper[4824]: I0121 11:28:47.672551 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-9b86998b5-lwrg9"] Jan 21 11:28:48 crc kubenswrapper[4824]: I0121 11:28:48.057627 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8247185f-987d-4f6e-b633-042907078e39" path="/var/lib/kubelet/pods/8247185f-987d-4f6e-b633-042907078e39/volumes" Jan 21 11:28:51 crc kubenswrapper[4824]: I0121 11:28:51.232002 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Jan 21 11:28:51 crc kubenswrapper[4824]: I0121 11:28:51.246478 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Jan 21 11:28:51 crc kubenswrapper[4824]: I0121 11:28:51.678245 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Jan 21 11:28:51 crc kubenswrapper[4824]: I0121 11:28:51.786161 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-6c642"] Jan 21 11:28:51 crc kubenswrapper[4824]: E0121 11:28:51.786478 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8247185f-987d-4f6e-b633-042907078e39" containerName="init" Jan 21 11:28:51 crc kubenswrapper[4824]: I0121 11:28:51.786494 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="8247185f-987d-4f6e-b633-042907078e39" containerName="init" Jan 21 11:28:51 crc kubenswrapper[4824]: E0121 11:28:51.786509 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8247185f-987d-4f6e-b633-042907078e39" containerName="dnsmasq-dns" Jan 21 11:28:51 crc kubenswrapper[4824]: I0121 11:28:51.786516 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="8247185f-987d-4f6e-b633-042907078e39" containerName="dnsmasq-dns" Jan 21 11:28:51 crc kubenswrapper[4824]: I0121 11:28:51.786696 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="8247185f-987d-4f6e-b633-042907078e39" containerName="dnsmasq-dns" Jan 21 11:28:51 crc kubenswrapper[4824]: I0121 11:28:51.788713 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-6c642" Jan 21 11:28:51 crc kubenswrapper[4824]: I0121 11:28:51.795057 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-6c642"] Jan 21 11:28:51 crc kubenswrapper[4824]: I0121 11:28:51.795203 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Jan 21 11:28:51 crc kubenswrapper[4824]: I0121 11:28:51.795545 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Jan 21 11:28:51 crc kubenswrapper[4824]: I0121 11:28:51.930457 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/12c162ff-82c2-4166-8a44-92e0455cae39-config-data\") pod \"nova-cell1-cell-mapping-6c642\" (UID: \"12c162ff-82c2-4166-8a44-92e0455cae39\") " pod="openstack/nova-cell1-cell-mapping-6c642" Jan 21 11:28:51 crc kubenswrapper[4824]: I0121 11:28:51.930770 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/12c162ff-82c2-4166-8a44-92e0455cae39-scripts\") pod \"nova-cell1-cell-mapping-6c642\" (UID: \"12c162ff-82c2-4166-8a44-92e0455cae39\") " pod="openstack/nova-cell1-cell-mapping-6c642" Jan 21 11:28:51 crc kubenswrapper[4824]: I0121 11:28:51.930789 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12c162ff-82c2-4166-8a44-92e0455cae39-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-6c642\" (UID: \"12c162ff-82c2-4166-8a44-92e0455cae39\") " pod="openstack/nova-cell1-cell-mapping-6c642" Jan 21 11:28:51 crc kubenswrapper[4824]: I0121 11:28:51.930853 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lcvgk\" (UniqueName: \"kubernetes.io/projected/12c162ff-82c2-4166-8a44-92e0455cae39-kube-api-access-lcvgk\") pod \"nova-cell1-cell-mapping-6c642\" (UID: \"12c162ff-82c2-4166-8a44-92e0455cae39\") " pod="openstack/nova-cell1-cell-mapping-6c642" Jan 21 11:28:52 crc kubenswrapper[4824]: I0121 11:28:52.032246 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/12c162ff-82c2-4166-8a44-92e0455cae39-scripts\") pod \"nova-cell1-cell-mapping-6c642\" (UID: \"12c162ff-82c2-4166-8a44-92e0455cae39\") " pod="openstack/nova-cell1-cell-mapping-6c642" Jan 21 11:28:52 crc kubenswrapper[4824]: I0121 11:28:52.032285 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12c162ff-82c2-4166-8a44-92e0455cae39-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-6c642\" (UID: \"12c162ff-82c2-4166-8a44-92e0455cae39\") " pod="openstack/nova-cell1-cell-mapping-6c642" Jan 21 11:28:52 crc kubenswrapper[4824]: I0121 11:28:52.032353 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lcvgk\" (UniqueName: \"kubernetes.io/projected/12c162ff-82c2-4166-8a44-92e0455cae39-kube-api-access-lcvgk\") pod \"nova-cell1-cell-mapping-6c642\" (UID: \"12c162ff-82c2-4166-8a44-92e0455cae39\") " pod="openstack/nova-cell1-cell-mapping-6c642" Jan 21 11:28:52 crc kubenswrapper[4824]: I0121 11:28:52.032461 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/12c162ff-82c2-4166-8a44-92e0455cae39-config-data\") pod \"nova-cell1-cell-mapping-6c642\" (UID: \"12c162ff-82c2-4166-8a44-92e0455cae39\") " pod="openstack/nova-cell1-cell-mapping-6c642" Jan 21 11:28:52 crc kubenswrapper[4824]: I0121 11:28:52.037007 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/12c162ff-82c2-4166-8a44-92e0455cae39-config-data\") pod \"nova-cell1-cell-mapping-6c642\" (UID: \"12c162ff-82c2-4166-8a44-92e0455cae39\") " pod="openstack/nova-cell1-cell-mapping-6c642" Jan 21 11:28:52 crc kubenswrapper[4824]: I0121 11:28:52.037165 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12c162ff-82c2-4166-8a44-92e0455cae39-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-6c642\" (UID: \"12c162ff-82c2-4166-8a44-92e0455cae39\") " pod="openstack/nova-cell1-cell-mapping-6c642" Jan 21 11:28:52 crc kubenswrapper[4824]: I0121 11:28:52.037882 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/12c162ff-82c2-4166-8a44-92e0455cae39-scripts\") pod \"nova-cell1-cell-mapping-6c642\" (UID: \"12c162ff-82c2-4166-8a44-92e0455cae39\") " pod="openstack/nova-cell1-cell-mapping-6c642" Jan 21 11:28:52 crc kubenswrapper[4824]: I0121 11:28:52.044719 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lcvgk\" (UniqueName: \"kubernetes.io/projected/12c162ff-82c2-4166-8a44-92e0455cae39-kube-api-access-lcvgk\") pod \"nova-cell1-cell-mapping-6c642\" (UID: \"12c162ff-82c2-4166-8a44-92e0455cae39\") " pod="openstack/nova-cell1-cell-mapping-6c642" Jan 21 11:28:52 crc kubenswrapper[4824]: I0121 11:28:52.106321 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-6c642" Jan 21 11:28:52 crc kubenswrapper[4824]: I0121 11:28:52.470707 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-6c642"] Jan 21 11:28:52 crc kubenswrapper[4824]: I0121 11:28:52.673604 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-6c642" event={"ID":"12c162ff-82c2-4166-8a44-92e0455cae39","Type":"ContainerStarted","Data":"7468f8bf7c17d51a36aabd9a1f3d9c19a0cb1f2b474c4bfcd676598418dd86f7"} Jan 21 11:28:52 crc kubenswrapper[4824]: I0121 11:28:52.673652 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-6c642" event={"ID":"12c162ff-82c2-4166-8a44-92e0455cae39","Type":"ContainerStarted","Data":"7a2f6e5181f30abf986b63626214d85313d1465ba029627b22f64bfd58a63370"} Jan 21 11:28:52 crc kubenswrapper[4824]: I0121 11:28:52.690976 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-6c642" podStartSLOduration=1.690948538 podStartE2EDuration="1.690948538s" podCreationTimestamp="2026-01-21 11:28:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:28:52.685937899 +0000 UTC m=+1074.978967191" watchObservedRunningTime="2026-01-21 11:28:52.690948538 +0000 UTC m=+1074.983977829" Jan 21 11:28:52 crc kubenswrapper[4824]: I0121 11:28:52.950137 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Jan 21 11:28:52 crc kubenswrapper[4824]: I0121 11:28:52.950174 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Jan 21 11:28:53 crc kubenswrapper[4824]: I0121 11:28:53.965061 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="ad597758-1f2c-4f84-8035-59fa20e68084" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.209:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Jan 21 11:28:53 crc kubenswrapper[4824]: I0121 11:28:53.965094 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="ad597758-1f2c-4f84-8035-59fa20e68084" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.209:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Jan 21 11:28:56 crc kubenswrapper[4824]: I0121 11:28:56.702869 4824 generic.go:334] "Generic (PLEG): container finished" podID="12c162ff-82c2-4166-8a44-92e0455cae39" containerID="7468f8bf7c17d51a36aabd9a1f3d9c19a0cb1f2b474c4bfcd676598418dd86f7" exitCode=0 Jan 21 11:28:56 crc kubenswrapper[4824]: I0121 11:28:56.702975 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-6c642" event={"ID":"12c162ff-82c2-4166-8a44-92e0455cae39","Type":"ContainerDied","Data":"7468f8bf7c17d51a36aabd9a1f3d9c19a0cb1f2b474c4bfcd676598418dd86f7"} Jan 21 11:28:57 crc kubenswrapper[4824]: I0121 11:28:57.930156 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-6c642" Jan 21 11:28:58 crc kubenswrapper[4824]: I0121 11:28:58.029602 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/12c162ff-82c2-4166-8a44-92e0455cae39-config-data\") pod \"12c162ff-82c2-4166-8a44-92e0455cae39\" (UID: \"12c162ff-82c2-4166-8a44-92e0455cae39\") " Jan 21 11:28:58 crc kubenswrapper[4824]: I0121 11:28:58.029665 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12c162ff-82c2-4166-8a44-92e0455cae39-combined-ca-bundle\") pod \"12c162ff-82c2-4166-8a44-92e0455cae39\" (UID: \"12c162ff-82c2-4166-8a44-92e0455cae39\") " Jan 21 11:28:58 crc kubenswrapper[4824]: I0121 11:28:58.029693 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lcvgk\" (UniqueName: \"kubernetes.io/projected/12c162ff-82c2-4166-8a44-92e0455cae39-kube-api-access-lcvgk\") pod \"12c162ff-82c2-4166-8a44-92e0455cae39\" (UID: \"12c162ff-82c2-4166-8a44-92e0455cae39\") " Jan 21 11:28:58 crc kubenswrapper[4824]: I0121 11:28:58.029720 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/12c162ff-82c2-4166-8a44-92e0455cae39-scripts\") pod \"12c162ff-82c2-4166-8a44-92e0455cae39\" (UID: \"12c162ff-82c2-4166-8a44-92e0455cae39\") " Jan 21 11:28:58 crc kubenswrapper[4824]: I0121 11:28:58.033653 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/12c162ff-82c2-4166-8a44-92e0455cae39-scripts" (OuterVolumeSpecName: "scripts") pod "12c162ff-82c2-4166-8a44-92e0455cae39" (UID: "12c162ff-82c2-4166-8a44-92e0455cae39"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:28:58 crc kubenswrapper[4824]: I0121 11:28:58.033909 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/12c162ff-82c2-4166-8a44-92e0455cae39-kube-api-access-lcvgk" (OuterVolumeSpecName: "kube-api-access-lcvgk") pod "12c162ff-82c2-4166-8a44-92e0455cae39" (UID: "12c162ff-82c2-4166-8a44-92e0455cae39"). InnerVolumeSpecName "kube-api-access-lcvgk". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:28:58 crc kubenswrapper[4824]: E0121 11:28:58.047430 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/12c162ff-82c2-4166-8a44-92e0455cae39-config-data podName:12c162ff-82c2-4166-8a44-92e0455cae39 nodeName:}" failed. No retries permitted until 2026-01-21 11:28:58.547413035 +0000 UTC m=+1080.840442327 (durationBeforeRetry 500ms). Error: error cleaning subPath mounts for volume "config-data" (UniqueName: "kubernetes.io/secret/12c162ff-82c2-4166-8a44-92e0455cae39-config-data") pod "12c162ff-82c2-4166-8a44-92e0455cae39" (UID: "12c162ff-82c2-4166-8a44-92e0455cae39") : error deleting /var/lib/kubelet/pods/12c162ff-82c2-4166-8a44-92e0455cae39/volume-subpaths: remove /var/lib/kubelet/pods/12c162ff-82c2-4166-8a44-92e0455cae39/volume-subpaths: no such file or directory Jan 21 11:28:58 crc kubenswrapper[4824]: I0121 11:28:58.049260 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/12c162ff-82c2-4166-8a44-92e0455cae39-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "12c162ff-82c2-4166-8a44-92e0455cae39" (UID: "12c162ff-82c2-4166-8a44-92e0455cae39"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:28:58 crc kubenswrapper[4824]: I0121 11:28:58.132138 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12c162ff-82c2-4166-8a44-92e0455cae39-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 21 11:28:58 crc kubenswrapper[4824]: I0121 11:28:58.132165 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lcvgk\" (UniqueName: \"kubernetes.io/projected/12c162ff-82c2-4166-8a44-92e0455cae39-kube-api-access-lcvgk\") on node \"crc\" DevicePath \"\"" Jan 21 11:28:58 crc kubenswrapper[4824]: I0121 11:28:58.132177 4824 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/12c162ff-82c2-4166-8a44-92e0455cae39-scripts\") on node \"crc\" DevicePath \"\"" Jan 21 11:28:58 crc kubenswrapper[4824]: I0121 11:28:58.640848 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/12c162ff-82c2-4166-8a44-92e0455cae39-config-data\") pod \"12c162ff-82c2-4166-8a44-92e0455cae39\" (UID: \"12c162ff-82c2-4166-8a44-92e0455cae39\") " Jan 21 11:28:58 crc kubenswrapper[4824]: I0121 11:28:58.644742 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/12c162ff-82c2-4166-8a44-92e0455cae39-config-data" (OuterVolumeSpecName: "config-data") pod "12c162ff-82c2-4166-8a44-92e0455cae39" (UID: "12c162ff-82c2-4166-8a44-92e0455cae39"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:28:58 crc kubenswrapper[4824]: I0121 11:28:58.719855 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-6c642" event={"ID":"12c162ff-82c2-4166-8a44-92e0455cae39","Type":"ContainerDied","Data":"7a2f6e5181f30abf986b63626214d85313d1465ba029627b22f64bfd58a63370"} Jan 21 11:28:58 crc kubenswrapper[4824]: I0121 11:28:58.719888 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7a2f6e5181f30abf986b63626214d85313d1465ba029627b22f64bfd58a63370" Jan 21 11:28:58 crc kubenswrapper[4824]: I0121 11:28:58.719916 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-6c642" Jan 21 11:28:58 crc kubenswrapper[4824]: I0121 11:28:58.743162 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/12c162ff-82c2-4166-8a44-92e0455cae39-config-data\") on node \"crc\" DevicePath \"\"" Jan 21 11:28:58 crc kubenswrapper[4824]: I0121 11:28:58.870597 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Jan 21 11:28:58 crc kubenswrapper[4824]: I0121 11:28:58.870853 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="ad597758-1f2c-4f84-8035-59fa20e68084" containerName="nova-api-log" containerID="cri-o://4a99cc0942c83c20256fb6754c20ac3db4e17b12130bc22c771bd7cc4942d218" gracePeriod=30 Jan 21 11:28:58 crc kubenswrapper[4824]: I0121 11:28:58.870907 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="ad597758-1f2c-4f84-8035-59fa20e68084" containerName="nova-api-api" containerID="cri-o://ec2b52af19a4c108e2ffb7356d1ff797d5745c69b97673e2b9ad650f81683331" gracePeriod=30 Jan 21 11:28:58 crc kubenswrapper[4824]: I0121 11:28:58.879576 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Jan 21 11:28:58 crc kubenswrapper[4824]: I0121 11:28:58.879743 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="a8b9cd0d-c7af-455d-a170-1521870a07f6" containerName="nova-scheduler-scheduler" containerID="cri-o://be161c0756cab08e44d6825e362adf099303c412379798ba94c4f4bf705f4830" gracePeriod=30 Jan 21 11:28:58 crc kubenswrapper[4824]: I0121 11:28:58.895901 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Jan 21 11:28:58 crc kubenswrapper[4824]: I0121 11:28:58.896170 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="3b9162ed-f3d2-4877-b8a3-6342a25200b7" containerName="nova-metadata-log" containerID="cri-o://55a7e3a41e9be3e82e087f957b241f50bebac19bc452a77348ef96c1447373d9" gracePeriod=30 Jan 21 11:28:58 crc kubenswrapper[4824]: I0121 11:28:58.896221 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="3b9162ed-f3d2-4877-b8a3-6342a25200b7" containerName="nova-metadata-metadata" containerID="cri-o://12872dcc363c76e366d89e82065e5f0f0eb8251e49a88c72fc419ac5a668148c" gracePeriod=30 Jan 21 11:28:59 crc kubenswrapper[4824]: I0121 11:28:59.727633 4824 generic.go:334] "Generic (PLEG): container finished" podID="ad597758-1f2c-4f84-8035-59fa20e68084" containerID="4a99cc0942c83c20256fb6754c20ac3db4e17b12130bc22c771bd7cc4942d218" exitCode=143 Jan 21 11:28:59 crc kubenswrapper[4824]: I0121 11:28:59.727697 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ad597758-1f2c-4f84-8035-59fa20e68084","Type":"ContainerDied","Data":"4a99cc0942c83c20256fb6754c20ac3db4e17b12130bc22c771bd7cc4942d218"} Jan 21 11:28:59 crc kubenswrapper[4824]: I0121 11:28:59.729731 4824 generic.go:334] "Generic (PLEG): container finished" podID="3b9162ed-f3d2-4877-b8a3-6342a25200b7" containerID="55a7e3a41e9be3e82e087f957b241f50bebac19bc452a77348ef96c1447373d9" exitCode=143 Jan 21 11:28:59 crc kubenswrapper[4824]: I0121 11:28:59.729757 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"3b9162ed-f3d2-4877-b8a3-6342a25200b7","Type":"ContainerDied","Data":"55a7e3a41e9be3e82e087f957b241f50bebac19bc452a77348ef96c1447373d9"} Jan 21 11:29:00 crc kubenswrapper[4824]: I0121 11:29:00.737367 4824 generic.go:334] "Generic (PLEG): container finished" podID="a8b9cd0d-c7af-455d-a170-1521870a07f6" containerID="be161c0756cab08e44d6825e362adf099303c412379798ba94c4f4bf705f4830" exitCode=0 Jan 21 11:29:00 crc kubenswrapper[4824]: I0121 11:29:00.737446 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"a8b9cd0d-c7af-455d-a170-1521870a07f6","Type":"ContainerDied","Data":"be161c0756cab08e44d6825e362adf099303c412379798ba94c4f4bf705f4830"} Jan 21 11:29:00 crc kubenswrapper[4824]: I0121 11:29:00.957234 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Jan 21 11:29:01 crc kubenswrapper[4824]: I0121 11:29:01.078919 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a8b9cd0d-c7af-455d-a170-1521870a07f6-config-data\") pod \"a8b9cd0d-c7af-455d-a170-1521870a07f6\" (UID: \"a8b9cd0d-c7af-455d-a170-1521870a07f6\") " Jan 21 11:29:01 crc kubenswrapper[4824]: I0121 11:29:01.079085 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nsphm\" (UniqueName: \"kubernetes.io/projected/a8b9cd0d-c7af-455d-a170-1521870a07f6-kube-api-access-nsphm\") pod \"a8b9cd0d-c7af-455d-a170-1521870a07f6\" (UID: \"a8b9cd0d-c7af-455d-a170-1521870a07f6\") " Jan 21 11:29:01 crc kubenswrapper[4824]: I0121 11:29:01.079165 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8b9cd0d-c7af-455d-a170-1521870a07f6-combined-ca-bundle\") pod \"a8b9cd0d-c7af-455d-a170-1521870a07f6\" (UID: \"a8b9cd0d-c7af-455d-a170-1521870a07f6\") " Jan 21 11:29:01 crc kubenswrapper[4824]: I0121 11:29:01.083253 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a8b9cd0d-c7af-455d-a170-1521870a07f6-kube-api-access-nsphm" (OuterVolumeSpecName: "kube-api-access-nsphm") pod "a8b9cd0d-c7af-455d-a170-1521870a07f6" (UID: "a8b9cd0d-c7af-455d-a170-1521870a07f6"). InnerVolumeSpecName "kube-api-access-nsphm". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:29:01 crc kubenswrapper[4824]: I0121 11:29:01.099861 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a8b9cd0d-c7af-455d-a170-1521870a07f6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a8b9cd0d-c7af-455d-a170-1521870a07f6" (UID: "a8b9cd0d-c7af-455d-a170-1521870a07f6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:29:01 crc kubenswrapper[4824]: I0121 11:29:01.115359 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a8b9cd0d-c7af-455d-a170-1521870a07f6-config-data" (OuterVolumeSpecName: "config-data") pod "a8b9cd0d-c7af-455d-a170-1521870a07f6" (UID: "a8b9cd0d-c7af-455d-a170-1521870a07f6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:29:01 crc kubenswrapper[4824]: I0121 11:29:01.182794 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a8b9cd0d-c7af-455d-a170-1521870a07f6-config-data\") on node \"crc\" DevicePath \"\"" Jan 21 11:29:01 crc kubenswrapper[4824]: I0121 11:29:01.183286 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nsphm\" (UniqueName: \"kubernetes.io/projected/a8b9cd0d-c7af-455d-a170-1521870a07f6-kube-api-access-nsphm\") on node \"crc\" DevicePath \"\"" Jan 21 11:29:01 crc kubenswrapper[4824]: I0121 11:29:01.183318 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8b9cd0d-c7af-455d-a170-1521870a07f6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 21 11:29:01 crc kubenswrapper[4824]: I0121 11:29:01.744533 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"a8b9cd0d-c7af-455d-a170-1521870a07f6","Type":"ContainerDied","Data":"bcd965629c724abfbdcc4781034ea86d257b188520a919034221d6a8bf2c0f86"} Jan 21 11:29:01 crc kubenswrapper[4824]: I0121 11:29:01.744583 4824 scope.go:117] "RemoveContainer" containerID="be161c0756cab08e44d6825e362adf099303c412379798ba94c4f4bf705f4830" Jan 21 11:29:01 crc kubenswrapper[4824]: I0121 11:29:01.744592 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Jan 21 11:29:01 crc kubenswrapper[4824]: I0121 11:29:01.768257 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Jan 21 11:29:01 crc kubenswrapper[4824]: I0121 11:29:01.776717 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Jan 21 11:29:01 crc kubenswrapper[4824]: I0121 11:29:01.784451 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Jan 21 11:29:01 crc kubenswrapper[4824]: E0121 11:29:01.784824 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="12c162ff-82c2-4166-8a44-92e0455cae39" containerName="nova-manage" Jan 21 11:29:01 crc kubenswrapper[4824]: I0121 11:29:01.784841 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="12c162ff-82c2-4166-8a44-92e0455cae39" containerName="nova-manage" Jan 21 11:29:01 crc kubenswrapper[4824]: E0121 11:29:01.784873 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a8b9cd0d-c7af-455d-a170-1521870a07f6" containerName="nova-scheduler-scheduler" Jan 21 11:29:01 crc kubenswrapper[4824]: I0121 11:29:01.784881 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="a8b9cd0d-c7af-455d-a170-1521870a07f6" containerName="nova-scheduler-scheduler" Jan 21 11:29:01 crc kubenswrapper[4824]: I0121 11:29:01.785057 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="12c162ff-82c2-4166-8a44-92e0455cae39" containerName="nova-manage" Jan 21 11:29:01 crc kubenswrapper[4824]: I0121 11:29:01.785092 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="a8b9cd0d-c7af-455d-a170-1521870a07f6" containerName="nova-scheduler-scheduler" Jan 21 11:29:01 crc kubenswrapper[4824]: I0121 11:29:01.785637 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Jan 21 11:29:01 crc kubenswrapper[4824]: I0121 11:29:01.787325 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Jan 21 11:29:01 crc kubenswrapper[4824]: I0121 11:29:01.794382 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Jan 21 11:29:01 crc kubenswrapper[4824]: I0121 11:29:01.894844 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7860ffbb-b435-4736-b2d3-f8871e70dc7a-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"7860ffbb-b435-4736-b2d3-f8871e70dc7a\") " pod="openstack/nova-scheduler-0" Jan 21 11:29:01 crc kubenswrapper[4824]: I0121 11:29:01.895036 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7860ffbb-b435-4736-b2d3-f8871e70dc7a-config-data\") pod \"nova-scheduler-0\" (UID: \"7860ffbb-b435-4736-b2d3-f8871e70dc7a\") " pod="openstack/nova-scheduler-0" Jan 21 11:29:01 crc kubenswrapper[4824]: I0121 11:29:01.895092 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lnbn7\" (UniqueName: \"kubernetes.io/projected/7860ffbb-b435-4736-b2d3-f8871e70dc7a-kube-api-access-lnbn7\") pod \"nova-scheduler-0\" (UID: \"7860ffbb-b435-4736-b2d3-f8871e70dc7a\") " pod="openstack/nova-scheduler-0" Jan 21 11:29:01 crc kubenswrapper[4824]: I0121 11:29:01.996580 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7860ffbb-b435-4736-b2d3-f8871e70dc7a-config-data\") pod \"nova-scheduler-0\" (UID: \"7860ffbb-b435-4736-b2d3-f8871e70dc7a\") " pod="openstack/nova-scheduler-0" Jan 21 11:29:01 crc kubenswrapper[4824]: I0121 11:29:01.997162 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lnbn7\" (UniqueName: \"kubernetes.io/projected/7860ffbb-b435-4736-b2d3-f8871e70dc7a-kube-api-access-lnbn7\") pod \"nova-scheduler-0\" (UID: \"7860ffbb-b435-4736-b2d3-f8871e70dc7a\") " pod="openstack/nova-scheduler-0" Jan 21 11:29:01 crc kubenswrapper[4824]: I0121 11:29:01.997241 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7860ffbb-b435-4736-b2d3-f8871e70dc7a-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"7860ffbb-b435-4736-b2d3-f8871e70dc7a\") " pod="openstack/nova-scheduler-0" Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.001166 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7860ffbb-b435-4736-b2d3-f8871e70dc7a-config-data\") pod \"nova-scheduler-0\" (UID: \"7860ffbb-b435-4736-b2d3-f8871e70dc7a\") " pod="openstack/nova-scheduler-0" Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.001450 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7860ffbb-b435-4736-b2d3-f8871e70dc7a-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"7860ffbb-b435-4736-b2d3-f8871e70dc7a\") " pod="openstack/nova-scheduler-0" Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.010208 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lnbn7\" (UniqueName: \"kubernetes.io/projected/7860ffbb-b435-4736-b2d3-f8871e70dc7a-kube-api-access-lnbn7\") pod \"nova-scheduler-0\" (UID: \"7860ffbb-b435-4736-b2d3-f8871e70dc7a\") " pod="openstack/nova-scheduler-0" Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.011226 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="3b9162ed-f3d2-4877-b8a3-6342a25200b7" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.204:8775/\": dial tcp 10.217.0.204:8775: connect: connection refused" Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.011229 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="3b9162ed-f3d2-4877-b8a3-6342a25200b7" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.204:8775/\": dial tcp 10.217.0.204:8775: connect: connection refused" Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.057664 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a8b9cd0d-c7af-455d-a170-1521870a07f6" path="/var/lib/kubelet/pods/a8b9cd0d-c7af-455d-a170-1521870a07f6/volumes" Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.101899 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.398903 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.415153 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.506032 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ad597758-1f2c-4f84-8035-59fa20e68084-combined-ca-bundle\") pod \"ad597758-1f2c-4f84-8035-59fa20e68084\" (UID: \"ad597758-1f2c-4f84-8035-59fa20e68084\") " Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.506114 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ad597758-1f2c-4f84-8035-59fa20e68084-logs\") pod \"ad597758-1f2c-4f84-8035-59fa20e68084\" (UID: \"ad597758-1f2c-4f84-8035-59fa20e68084\") " Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.506267 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-22h96\" (UniqueName: \"kubernetes.io/projected/ad597758-1f2c-4f84-8035-59fa20e68084-kube-api-access-22h96\") pod \"ad597758-1f2c-4f84-8035-59fa20e68084\" (UID: \"ad597758-1f2c-4f84-8035-59fa20e68084\") " Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.506293 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ad597758-1f2c-4f84-8035-59fa20e68084-config-data\") pod \"ad597758-1f2c-4f84-8035-59fa20e68084\" (UID: \"ad597758-1f2c-4f84-8035-59fa20e68084\") " Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.506312 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ad597758-1f2c-4f84-8035-59fa20e68084-internal-tls-certs\") pod \"ad597758-1f2c-4f84-8035-59fa20e68084\" (UID: \"ad597758-1f2c-4f84-8035-59fa20e68084\") " Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.506350 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ad597758-1f2c-4f84-8035-59fa20e68084-public-tls-certs\") pod \"ad597758-1f2c-4f84-8035-59fa20e68084\" (UID: \"ad597758-1f2c-4f84-8035-59fa20e68084\") " Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.506922 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ad597758-1f2c-4f84-8035-59fa20e68084-logs" (OuterVolumeSpecName: "logs") pod "ad597758-1f2c-4f84-8035-59fa20e68084" (UID: "ad597758-1f2c-4f84-8035-59fa20e68084"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.509178 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ad597758-1f2c-4f84-8035-59fa20e68084-kube-api-access-22h96" (OuterVolumeSpecName: "kube-api-access-22h96") pod "ad597758-1f2c-4f84-8035-59fa20e68084" (UID: "ad597758-1f2c-4f84-8035-59fa20e68084"). InnerVolumeSpecName "kube-api-access-22h96". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.530539 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ad597758-1f2c-4f84-8035-59fa20e68084-config-data" (OuterVolumeSpecName: "config-data") pod "ad597758-1f2c-4f84-8035-59fa20e68084" (UID: "ad597758-1f2c-4f84-8035-59fa20e68084"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.530612 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ad597758-1f2c-4f84-8035-59fa20e68084-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ad597758-1f2c-4f84-8035-59fa20e68084" (UID: "ad597758-1f2c-4f84-8035-59fa20e68084"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.536852 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Jan 21 11:29:02 crc kubenswrapper[4824]: W0121 11:29:02.537448 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7860ffbb_b435_4736_b2d3_f8871e70dc7a.slice/crio-54b1920e6073c5b64a316ce4f6144c773b3d456175a5da934941d89af91d6ff5 WatchSource:0}: Error finding container 54b1920e6073c5b64a316ce4f6144c773b3d456175a5da934941d89af91d6ff5: Status 404 returned error can't find the container with id 54b1920e6073c5b64a316ce4f6144c773b3d456175a5da934941d89af91d6ff5 Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.546631 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ad597758-1f2c-4f84-8035-59fa20e68084-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "ad597758-1f2c-4f84-8035-59fa20e68084" (UID: "ad597758-1f2c-4f84-8035-59fa20e68084"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.550309 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ad597758-1f2c-4f84-8035-59fa20e68084-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "ad597758-1f2c-4f84-8035-59fa20e68084" (UID: "ad597758-1f2c-4f84-8035-59fa20e68084"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.608065 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lpvwd\" (UniqueName: \"kubernetes.io/projected/3b9162ed-f3d2-4877-b8a3-6342a25200b7-kube-api-access-lpvwd\") pod \"3b9162ed-f3d2-4877-b8a3-6342a25200b7\" (UID: \"3b9162ed-f3d2-4877-b8a3-6342a25200b7\") " Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.608098 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3b9162ed-f3d2-4877-b8a3-6342a25200b7-config-data\") pod \"3b9162ed-f3d2-4877-b8a3-6342a25200b7\" (UID: \"3b9162ed-f3d2-4877-b8a3-6342a25200b7\") " Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.608151 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/3b9162ed-f3d2-4877-b8a3-6342a25200b7-nova-metadata-tls-certs\") pod \"3b9162ed-f3d2-4877-b8a3-6342a25200b7\" (UID: \"3b9162ed-f3d2-4877-b8a3-6342a25200b7\") " Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.608244 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3b9162ed-f3d2-4877-b8a3-6342a25200b7-logs\") pod \"3b9162ed-f3d2-4877-b8a3-6342a25200b7\" (UID: \"3b9162ed-f3d2-4877-b8a3-6342a25200b7\") " Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.608982 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3b9162ed-f3d2-4877-b8a3-6342a25200b7-logs" (OuterVolumeSpecName: "logs") pod "3b9162ed-f3d2-4877-b8a3-6342a25200b7" (UID: "3b9162ed-f3d2-4877-b8a3-6342a25200b7"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.609115 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b9162ed-f3d2-4877-b8a3-6342a25200b7-combined-ca-bundle\") pod \"3b9162ed-f3d2-4877-b8a3-6342a25200b7\" (UID: \"3b9162ed-f3d2-4877-b8a3-6342a25200b7\") " Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.609947 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-22h96\" (UniqueName: \"kubernetes.io/projected/ad597758-1f2c-4f84-8035-59fa20e68084-kube-api-access-22h96\") on node \"crc\" DevicePath \"\"" Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.609992 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ad597758-1f2c-4f84-8035-59fa20e68084-config-data\") on node \"crc\" DevicePath \"\"" Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.610009 4824 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ad597758-1f2c-4f84-8035-59fa20e68084-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.610018 4824 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ad597758-1f2c-4f84-8035-59fa20e68084-public-tls-certs\") on node \"crc\" DevicePath \"\"" Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.610025 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ad597758-1f2c-4f84-8035-59fa20e68084-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.610034 4824 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ad597758-1f2c-4f84-8035-59fa20e68084-logs\") on node \"crc\" DevicePath \"\"" Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.610103 4824 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3b9162ed-f3d2-4877-b8a3-6342a25200b7-logs\") on node \"crc\" DevicePath \"\"" Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.610835 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3b9162ed-f3d2-4877-b8a3-6342a25200b7-kube-api-access-lpvwd" (OuterVolumeSpecName: "kube-api-access-lpvwd") pod "3b9162ed-f3d2-4877-b8a3-6342a25200b7" (UID: "3b9162ed-f3d2-4877-b8a3-6342a25200b7"). InnerVolumeSpecName "kube-api-access-lpvwd". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.626675 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3b9162ed-f3d2-4877-b8a3-6342a25200b7-config-data" (OuterVolumeSpecName: "config-data") pod "3b9162ed-f3d2-4877-b8a3-6342a25200b7" (UID: "3b9162ed-f3d2-4877-b8a3-6342a25200b7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.628363 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3b9162ed-f3d2-4877-b8a3-6342a25200b7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3b9162ed-f3d2-4877-b8a3-6342a25200b7" (UID: "3b9162ed-f3d2-4877-b8a3-6342a25200b7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.650822 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3b9162ed-f3d2-4877-b8a3-6342a25200b7-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "3b9162ed-f3d2-4877-b8a3-6342a25200b7" (UID: "3b9162ed-f3d2-4877-b8a3-6342a25200b7"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.712111 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b9162ed-f3d2-4877-b8a3-6342a25200b7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.712140 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lpvwd\" (UniqueName: \"kubernetes.io/projected/3b9162ed-f3d2-4877-b8a3-6342a25200b7-kube-api-access-lpvwd\") on node \"crc\" DevicePath \"\"" Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.712151 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3b9162ed-f3d2-4877-b8a3-6342a25200b7-config-data\") on node \"crc\" DevicePath \"\"" Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.712159 4824 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/3b9162ed-f3d2-4877-b8a3-6342a25200b7-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.752864 4824 generic.go:334] "Generic (PLEG): container finished" podID="ad597758-1f2c-4f84-8035-59fa20e68084" containerID="ec2b52af19a4c108e2ffb7356d1ff797d5745c69b97673e2b9ad650f81683331" exitCode=0 Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.752922 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ad597758-1f2c-4f84-8035-59fa20e68084","Type":"ContainerDied","Data":"ec2b52af19a4c108e2ffb7356d1ff797d5745c69b97673e2b9ad650f81683331"} Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.752975 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ad597758-1f2c-4f84-8035-59fa20e68084","Type":"ContainerDied","Data":"04682cea27407b0d473ba48a7b6b0c26d36ec7ff8680fb81116bbb44f989542f"} Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.752993 4824 scope.go:117] "RemoveContainer" containerID="ec2b52af19a4c108e2ffb7356d1ff797d5745c69b97673e2b9ad650f81683331" Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.753702 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.756571 4824 generic.go:334] "Generic (PLEG): container finished" podID="3b9162ed-f3d2-4877-b8a3-6342a25200b7" containerID="12872dcc363c76e366d89e82065e5f0f0eb8251e49a88c72fc419ac5a668148c" exitCode=0 Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.756617 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.756631 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"3b9162ed-f3d2-4877-b8a3-6342a25200b7","Type":"ContainerDied","Data":"12872dcc363c76e366d89e82065e5f0f0eb8251e49a88c72fc419ac5a668148c"} Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.756655 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"3b9162ed-f3d2-4877-b8a3-6342a25200b7","Type":"ContainerDied","Data":"cf896b02cb7f72cd22b94afd5cb645eb3dc94f2424b20aa748a3b1e35979c8a8"} Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.758646 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"7860ffbb-b435-4736-b2d3-f8871e70dc7a","Type":"ContainerStarted","Data":"92316441f6f75de082981736928c3af2413970490f85031826e288c70a591cca"} Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.758671 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"7860ffbb-b435-4736-b2d3-f8871e70dc7a","Type":"ContainerStarted","Data":"54b1920e6073c5b64a316ce4f6144c773b3d456175a5da934941d89af91d6ff5"} Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.769447 4824 scope.go:117] "RemoveContainer" containerID="4a99cc0942c83c20256fb6754c20ac3db4e17b12130bc22c771bd7cc4942d218" Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.770915 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=1.7709049430000001 podStartE2EDuration="1.770904943s" podCreationTimestamp="2026-01-21 11:29:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:29:02.768465102 +0000 UTC m=+1085.061494394" watchObservedRunningTime="2026-01-21 11:29:02.770904943 +0000 UTC m=+1085.063934235" Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.785667 4824 scope.go:117] "RemoveContainer" containerID="ec2b52af19a4c108e2ffb7356d1ff797d5745c69b97673e2b9ad650f81683331" Jan 21 11:29:02 crc kubenswrapper[4824]: E0121 11:29:02.786389 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ec2b52af19a4c108e2ffb7356d1ff797d5745c69b97673e2b9ad650f81683331\": container with ID starting with ec2b52af19a4c108e2ffb7356d1ff797d5745c69b97673e2b9ad650f81683331 not found: ID does not exist" containerID="ec2b52af19a4c108e2ffb7356d1ff797d5745c69b97673e2b9ad650f81683331" Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.786422 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ec2b52af19a4c108e2ffb7356d1ff797d5745c69b97673e2b9ad650f81683331"} err="failed to get container status \"ec2b52af19a4c108e2ffb7356d1ff797d5745c69b97673e2b9ad650f81683331\": rpc error: code = NotFound desc = could not find container \"ec2b52af19a4c108e2ffb7356d1ff797d5745c69b97673e2b9ad650f81683331\": container with ID starting with ec2b52af19a4c108e2ffb7356d1ff797d5745c69b97673e2b9ad650f81683331 not found: ID does not exist" Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.786442 4824 scope.go:117] "RemoveContainer" containerID="4a99cc0942c83c20256fb6754c20ac3db4e17b12130bc22c771bd7cc4942d218" Jan 21 11:29:02 crc kubenswrapper[4824]: E0121 11:29:02.786713 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4a99cc0942c83c20256fb6754c20ac3db4e17b12130bc22c771bd7cc4942d218\": container with ID starting with 4a99cc0942c83c20256fb6754c20ac3db4e17b12130bc22c771bd7cc4942d218 not found: ID does not exist" containerID="4a99cc0942c83c20256fb6754c20ac3db4e17b12130bc22c771bd7cc4942d218" Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.786788 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4a99cc0942c83c20256fb6754c20ac3db4e17b12130bc22c771bd7cc4942d218"} err="failed to get container status \"4a99cc0942c83c20256fb6754c20ac3db4e17b12130bc22c771bd7cc4942d218\": rpc error: code = NotFound desc = could not find container \"4a99cc0942c83c20256fb6754c20ac3db4e17b12130bc22c771bd7cc4942d218\": container with ID starting with 4a99cc0942c83c20256fb6754c20ac3db4e17b12130bc22c771bd7cc4942d218 not found: ID does not exist" Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.786877 4824 scope.go:117] "RemoveContainer" containerID="12872dcc363c76e366d89e82065e5f0f0eb8251e49a88c72fc419ac5a668148c" Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.787302 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.797520 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.810668 4824 scope.go:117] "RemoveContainer" containerID="55a7e3a41e9be3e82e087f957b241f50bebac19bc452a77348ef96c1447373d9" Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.810830 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.826146 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.839565 4824 scope.go:117] "RemoveContainer" containerID="12872dcc363c76e366d89e82065e5f0f0eb8251e49a88c72fc419ac5a668148c" Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.839663 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Jan 21 11:29:02 crc kubenswrapper[4824]: E0121 11:29:02.840045 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad597758-1f2c-4f84-8035-59fa20e68084" containerName="nova-api-api" Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.840062 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad597758-1f2c-4f84-8035-59fa20e68084" containerName="nova-api-api" Jan 21 11:29:02 crc kubenswrapper[4824]: E0121 11:29:02.840084 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b9162ed-f3d2-4877-b8a3-6342a25200b7" containerName="nova-metadata-log" Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.840090 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b9162ed-f3d2-4877-b8a3-6342a25200b7" containerName="nova-metadata-log" Jan 21 11:29:02 crc kubenswrapper[4824]: E0121 11:29:02.840096 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad597758-1f2c-4f84-8035-59fa20e68084" containerName="nova-api-log" Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.840101 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad597758-1f2c-4f84-8035-59fa20e68084" containerName="nova-api-log" Jan 21 11:29:02 crc kubenswrapper[4824]: E0121 11:29:02.840115 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b9162ed-f3d2-4877-b8a3-6342a25200b7" containerName="nova-metadata-metadata" Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.840121 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b9162ed-f3d2-4877-b8a3-6342a25200b7" containerName="nova-metadata-metadata" Jan 21 11:29:02 crc kubenswrapper[4824]: E0121 11:29:02.840119 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"12872dcc363c76e366d89e82065e5f0f0eb8251e49a88c72fc419ac5a668148c\": container with ID starting with 12872dcc363c76e366d89e82065e5f0f0eb8251e49a88c72fc419ac5a668148c not found: ID does not exist" containerID="12872dcc363c76e366d89e82065e5f0f0eb8251e49a88c72fc419ac5a668148c" Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.840154 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"12872dcc363c76e366d89e82065e5f0f0eb8251e49a88c72fc419ac5a668148c"} err="failed to get container status \"12872dcc363c76e366d89e82065e5f0f0eb8251e49a88c72fc419ac5a668148c\": rpc error: code = NotFound desc = could not find container \"12872dcc363c76e366d89e82065e5f0f0eb8251e49a88c72fc419ac5a668148c\": container with ID starting with 12872dcc363c76e366d89e82065e5f0f0eb8251e49a88c72fc419ac5a668148c not found: ID does not exist" Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.840176 4824 scope.go:117] "RemoveContainer" containerID="55a7e3a41e9be3e82e087f957b241f50bebac19bc452a77348ef96c1447373d9" Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.840282 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="ad597758-1f2c-4f84-8035-59fa20e68084" containerName="nova-api-api" Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.840303 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="3b9162ed-f3d2-4877-b8a3-6342a25200b7" containerName="nova-metadata-metadata" Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.840315 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="3b9162ed-f3d2-4877-b8a3-6342a25200b7" containerName="nova-metadata-log" Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.840321 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="ad597758-1f2c-4f84-8035-59fa20e68084" containerName="nova-api-log" Jan 21 11:29:02 crc kubenswrapper[4824]: E0121 11:29:02.840652 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"55a7e3a41e9be3e82e087f957b241f50bebac19bc452a77348ef96c1447373d9\": container with ID starting with 55a7e3a41e9be3e82e087f957b241f50bebac19bc452a77348ef96c1447373d9 not found: ID does not exist" containerID="55a7e3a41e9be3e82e087f957b241f50bebac19bc452a77348ef96c1447373d9" Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.840672 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"55a7e3a41e9be3e82e087f957b241f50bebac19bc452a77348ef96c1447373d9"} err="failed to get container status \"55a7e3a41e9be3e82e087f957b241f50bebac19bc452a77348ef96c1447373d9\": rpc error: code = NotFound desc = could not find container \"55a7e3a41e9be3e82e087f957b241f50bebac19bc452a77348ef96c1447373d9\": container with ID starting with 55a7e3a41e9be3e82e087f957b241f50bebac19bc452a77348ef96c1447373d9 not found: ID does not exist" Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.841192 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.842891 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.843085 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.844082 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.845995 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.852534 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.853867 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.857250 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.857399 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Jan 21 11:29:02 crc kubenswrapper[4824]: I0121 11:29:02.858472 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Jan 21 11:29:03 crc kubenswrapper[4824]: I0121 11:29:03.018152 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/55525bd3-ae2b-494d-9a33-dd7d00c576b2-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"55525bd3-ae2b-494d-9a33-dd7d00c576b2\") " pod="openstack/nova-api-0" Jan 21 11:29:03 crc kubenswrapper[4824]: I0121 11:29:03.018243 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/654a41d0-b44a-4194-a514-a3c60126d0d9-logs\") pod \"nova-metadata-0\" (UID: \"654a41d0-b44a-4194-a514-a3c60126d0d9\") " pod="openstack/nova-metadata-0" Jan 21 11:29:03 crc kubenswrapper[4824]: I0121 11:29:03.018725 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/55525bd3-ae2b-494d-9a33-dd7d00c576b2-internal-tls-certs\") pod \"nova-api-0\" (UID: \"55525bd3-ae2b-494d-9a33-dd7d00c576b2\") " pod="openstack/nova-api-0" Jan 21 11:29:03 crc kubenswrapper[4824]: I0121 11:29:03.018754 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/55525bd3-ae2b-494d-9a33-dd7d00c576b2-public-tls-certs\") pod \"nova-api-0\" (UID: \"55525bd3-ae2b-494d-9a33-dd7d00c576b2\") " pod="openstack/nova-api-0" Jan 21 11:29:03 crc kubenswrapper[4824]: I0121 11:29:03.018844 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/654a41d0-b44a-4194-a514-a3c60126d0d9-config-data\") pod \"nova-metadata-0\" (UID: \"654a41d0-b44a-4194-a514-a3c60126d0d9\") " pod="openstack/nova-metadata-0" Jan 21 11:29:03 crc kubenswrapper[4824]: I0121 11:29:03.018994 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/654a41d0-b44a-4194-a514-a3c60126d0d9-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"654a41d0-b44a-4194-a514-a3c60126d0d9\") " pod="openstack/nova-metadata-0" Jan 21 11:29:03 crc kubenswrapper[4824]: I0121 11:29:03.019022 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/654a41d0-b44a-4194-a514-a3c60126d0d9-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"654a41d0-b44a-4194-a514-a3c60126d0d9\") " pod="openstack/nova-metadata-0" Jan 21 11:29:03 crc kubenswrapper[4824]: I0121 11:29:03.019043 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f6vg4\" (UniqueName: \"kubernetes.io/projected/654a41d0-b44a-4194-a514-a3c60126d0d9-kube-api-access-f6vg4\") pod \"nova-metadata-0\" (UID: \"654a41d0-b44a-4194-a514-a3c60126d0d9\") " pod="openstack/nova-metadata-0" Jan 21 11:29:03 crc kubenswrapper[4824]: I0121 11:29:03.019057 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/55525bd3-ae2b-494d-9a33-dd7d00c576b2-config-data\") pod \"nova-api-0\" (UID: \"55525bd3-ae2b-494d-9a33-dd7d00c576b2\") " pod="openstack/nova-api-0" Jan 21 11:29:03 crc kubenswrapper[4824]: I0121 11:29:03.019077 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/55525bd3-ae2b-494d-9a33-dd7d00c576b2-logs\") pod \"nova-api-0\" (UID: \"55525bd3-ae2b-494d-9a33-dd7d00c576b2\") " pod="openstack/nova-api-0" Jan 21 11:29:03 crc kubenswrapper[4824]: I0121 11:29:03.019205 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-96dc4\" (UniqueName: \"kubernetes.io/projected/55525bd3-ae2b-494d-9a33-dd7d00c576b2-kube-api-access-96dc4\") pod \"nova-api-0\" (UID: \"55525bd3-ae2b-494d-9a33-dd7d00c576b2\") " pod="openstack/nova-api-0" Jan 21 11:29:03 crc kubenswrapper[4824]: I0121 11:29:03.121428 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/654a41d0-b44a-4194-a514-a3c60126d0d9-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"654a41d0-b44a-4194-a514-a3c60126d0d9\") " pod="openstack/nova-metadata-0" Jan 21 11:29:03 crc kubenswrapper[4824]: I0121 11:29:03.121482 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/55525bd3-ae2b-494d-9a33-dd7d00c576b2-config-data\") pod \"nova-api-0\" (UID: \"55525bd3-ae2b-494d-9a33-dd7d00c576b2\") " pod="openstack/nova-api-0" Jan 21 11:29:03 crc kubenswrapper[4824]: I0121 11:29:03.121503 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/654a41d0-b44a-4194-a514-a3c60126d0d9-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"654a41d0-b44a-4194-a514-a3c60126d0d9\") " pod="openstack/nova-metadata-0" Jan 21 11:29:03 crc kubenswrapper[4824]: I0121 11:29:03.121521 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f6vg4\" (UniqueName: \"kubernetes.io/projected/654a41d0-b44a-4194-a514-a3c60126d0d9-kube-api-access-f6vg4\") pod \"nova-metadata-0\" (UID: \"654a41d0-b44a-4194-a514-a3c60126d0d9\") " pod="openstack/nova-metadata-0" Jan 21 11:29:03 crc kubenswrapper[4824]: I0121 11:29:03.121541 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/55525bd3-ae2b-494d-9a33-dd7d00c576b2-logs\") pod \"nova-api-0\" (UID: \"55525bd3-ae2b-494d-9a33-dd7d00c576b2\") " pod="openstack/nova-api-0" Jan 21 11:29:03 crc kubenswrapper[4824]: I0121 11:29:03.121569 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-96dc4\" (UniqueName: \"kubernetes.io/projected/55525bd3-ae2b-494d-9a33-dd7d00c576b2-kube-api-access-96dc4\") pod \"nova-api-0\" (UID: \"55525bd3-ae2b-494d-9a33-dd7d00c576b2\") " pod="openstack/nova-api-0" Jan 21 11:29:03 crc kubenswrapper[4824]: I0121 11:29:03.121693 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/55525bd3-ae2b-494d-9a33-dd7d00c576b2-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"55525bd3-ae2b-494d-9a33-dd7d00c576b2\") " pod="openstack/nova-api-0" Jan 21 11:29:03 crc kubenswrapper[4824]: I0121 11:29:03.121779 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/654a41d0-b44a-4194-a514-a3c60126d0d9-logs\") pod \"nova-metadata-0\" (UID: \"654a41d0-b44a-4194-a514-a3c60126d0d9\") " pod="openstack/nova-metadata-0" Jan 21 11:29:03 crc kubenswrapper[4824]: I0121 11:29:03.122265 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/55525bd3-ae2b-494d-9a33-dd7d00c576b2-logs\") pod \"nova-api-0\" (UID: \"55525bd3-ae2b-494d-9a33-dd7d00c576b2\") " pod="openstack/nova-api-0" Jan 21 11:29:03 crc kubenswrapper[4824]: I0121 11:29:03.121840 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/55525bd3-ae2b-494d-9a33-dd7d00c576b2-internal-tls-certs\") pod \"nova-api-0\" (UID: \"55525bd3-ae2b-494d-9a33-dd7d00c576b2\") " pod="openstack/nova-api-0" Jan 21 11:29:03 crc kubenswrapper[4824]: I0121 11:29:03.122395 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/55525bd3-ae2b-494d-9a33-dd7d00c576b2-public-tls-certs\") pod \"nova-api-0\" (UID: \"55525bd3-ae2b-494d-9a33-dd7d00c576b2\") " pod="openstack/nova-api-0" Jan 21 11:29:03 crc kubenswrapper[4824]: I0121 11:29:03.122455 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/654a41d0-b44a-4194-a514-a3c60126d0d9-config-data\") pod \"nova-metadata-0\" (UID: \"654a41d0-b44a-4194-a514-a3c60126d0d9\") " pod="openstack/nova-metadata-0" Jan 21 11:29:03 crc kubenswrapper[4824]: I0121 11:29:03.122553 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/654a41d0-b44a-4194-a514-a3c60126d0d9-logs\") pod \"nova-metadata-0\" (UID: \"654a41d0-b44a-4194-a514-a3c60126d0d9\") " pod="openstack/nova-metadata-0" Jan 21 11:29:03 crc kubenswrapper[4824]: I0121 11:29:03.132441 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/55525bd3-ae2b-494d-9a33-dd7d00c576b2-config-data\") pod \"nova-api-0\" (UID: \"55525bd3-ae2b-494d-9a33-dd7d00c576b2\") " pod="openstack/nova-api-0" Jan 21 11:29:03 crc kubenswrapper[4824]: I0121 11:29:03.139129 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/55525bd3-ae2b-494d-9a33-dd7d00c576b2-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"55525bd3-ae2b-494d-9a33-dd7d00c576b2\") " pod="openstack/nova-api-0" Jan 21 11:29:03 crc kubenswrapper[4824]: I0121 11:29:03.140898 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/55525bd3-ae2b-494d-9a33-dd7d00c576b2-internal-tls-certs\") pod \"nova-api-0\" (UID: \"55525bd3-ae2b-494d-9a33-dd7d00c576b2\") " pod="openstack/nova-api-0" Jan 21 11:29:03 crc kubenswrapper[4824]: I0121 11:29:03.141057 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/654a41d0-b44a-4194-a514-a3c60126d0d9-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"654a41d0-b44a-4194-a514-a3c60126d0d9\") " pod="openstack/nova-metadata-0" Jan 21 11:29:03 crc kubenswrapper[4824]: I0121 11:29:03.141163 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/654a41d0-b44a-4194-a514-a3c60126d0d9-config-data\") pod \"nova-metadata-0\" (UID: \"654a41d0-b44a-4194-a514-a3c60126d0d9\") " pod="openstack/nova-metadata-0" Jan 21 11:29:03 crc kubenswrapper[4824]: I0121 11:29:03.141281 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/55525bd3-ae2b-494d-9a33-dd7d00c576b2-public-tls-certs\") pod \"nova-api-0\" (UID: \"55525bd3-ae2b-494d-9a33-dd7d00c576b2\") " pod="openstack/nova-api-0" Jan 21 11:29:03 crc kubenswrapper[4824]: I0121 11:29:03.142695 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-96dc4\" (UniqueName: \"kubernetes.io/projected/55525bd3-ae2b-494d-9a33-dd7d00c576b2-kube-api-access-96dc4\") pod \"nova-api-0\" (UID: \"55525bd3-ae2b-494d-9a33-dd7d00c576b2\") " pod="openstack/nova-api-0" Jan 21 11:29:03 crc kubenswrapper[4824]: I0121 11:29:03.143817 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f6vg4\" (UniqueName: \"kubernetes.io/projected/654a41d0-b44a-4194-a514-a3c60126d0d9-kube-api-access-f6vg4\") pod \"nova-metadata-0\" (UID: \"654a41d0-b44a-4194-a514-a3c60126d0d9\") " pod="openstack/nova-metadata-0" Jan 21 11:29:03 crc kubenswrapper[4824]: I0121 11:29:03.151753 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/654a41d0-b44a-4194-a514-a3c60126d0d9-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"654a41d0-b44a-4194-a514-a3c60126d0d9\") " pod="openstack/nova-metadata-0" Jan 21 11:29:03 crc kubenswrapper[4824]: I0121 11:29:03.157396 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Jan 21 11:29:03 crc kubenswrapper[4824]: I0121 11:29:03.165157 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Jan 21 11:29:03 crc kubenswrapper[4824]: I0121 11:29:03.531105 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Jan 21 11:29:03 crc kubenswrapper[4824]: I0121 11:29:03.600812 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Jan 21 11:29:03 crc kubenswrapper[4824]: W0121 11:29:03.601781 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod55525bd3_ae2b_494d_9a33_dd7d00c576b2.slice/crio-450475ff361db096670a972124133a11e8549d19be646c798d0b7d4f12c00208 WatchSource:0}: Error finding container 450475ff361db096670a972124133a11e8549d19be646c798d0b7d4f12c00208: Status 404 returned error can't find the container with id 450475ff361db096670a972124133a11e8549d19be646c798d0b7d4f12c00208 Jan 21 11:29:03 crc kubenswrapper[4824]: I0121 11:29:03.767538 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"654a41d0-b44a-4194-a514-a3c60126d0d9","Type":"ContainerStarted","Data":"39d8671f823e852b86d992e4469bf9ba938c5889950d2d64d5b2f68626370183"} Jan 21 11:29:03 crc kubenswrapper[4824]: I0121 11:29:03.767752 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"654a41d0-b44a-4194-a514-a3c60126d0d9","Type":"ContainerStarted","Data":"c9941457bd25dfdb724c33502c3621bf31db3922f63f60e11fce2ece4cef0c7f"} Jan 21 11:29:03 crc kubenswrapper[4824]: I0121 11:29:03.767766 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"654a41d0-b44a-4194-a514-a3c60126d0d9","Type":"ContainerStarted","Data":"750307738ca25e9c7bf090aaadeeb9fbd671e51a467f1d767a28cf93a8e85a15"} Jan 21 11:29:03 crc kubenswrapper[4824]: I0121 11:29:03.777325 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"55525bd3-ae2b-494d-9a33-dd7d00c576b2","Type":"ContainerStarted","Data":"7ccf6dca71329f80f6672211c8a36a53367d5588455ddf171422d231a8c4f4b8"} Jan 21 11:29:03 crc kubenswrapper[4824]: I0121 11:29:03.777355 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"55525bd3-ae2b-494d-9a33-dd7d00c576b2","Type":"ContainerStarted","Data":"450475ff361db096670a972124133a11e8549d19be646c798d0b7d4f12c00208"} Jan 21 11:29:03 crc kubenswrapper[4824]: I0121 11:29:03.788456 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=1.788441911 podStartE2EDuration="1.788441911s" podCreationTimestamp="2026-01-21 11:29:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:29:03.782496571 +0000 UTC m=+1086.075525864" watchObservedRunningTime="2026-01-21 11:29:03.788441911 +0000 UTC m=+1086.081471204" Jan 21 11:29:04 crc kubenswrapper[4824]: I0121 11:29:04.061083 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3b9162ed-f3d2-4877-b8a3-6342a25200b7" path="/var/lib/kubelet/pods/3b9162ed-f3d2-4877-b8a3-6342a25200b7/volumes" Jan 21 11:29:04 crc kubenswrapper[4824]: I0121 11:29:04.061667 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ad597758-1f2c-4f84-8035-59fa20e68084" path="/var/lib/kubelet/pods/ad597758-1f2c-4f84-8035-59fa20e68084/volumes" Jan 21 11:29:04 crc kubenswrapper[4824]: I0121 11:29:04.794376 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"55525bd3-ae2b-494d-9a33-dd7d00c576b2","Type":"ContainerStarted","Data":"58fb9a1b8b1e252d070cfa3d6bb163d38f884129a6fe386c671924e7aca82b98"} Jan 21 11:29:04 crc kubenswrapper[4824]: I0121 11:29:04.807889 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.8078750550000002 podStartE2EDuration="2.807875055s" podCreationTimestamp="2026-01-21 11:29:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:29:04.80708257 +0000 UTC m=+1087.100111862" watchObservedRunningTime="2026-01-21 11:29:04.807875055 +0000 UTC m=+1087.100904347" Jan 21 11:29:07 crc kubenswrapper[4824]: I0121 11:29:07.102400 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Jan 21 11:29:08 crc kubenswrapper[4824]: I0121 11:29:08.165207 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Jan 21 11:29:08 crc kubenswrapper[4824]: I0121 11:29:08.165433 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Jan 21 11:29:10 crc kubenswrapper[4824]: I0121 11:29:10.928467 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Jan 21 11:29:12 crc kubenswrapper[4824]: I0121 11:29:12.102656 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Jan 21 11:29:12 crc kubenswrapper[4824]: I0121 11:29:12.122124 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Jan 21 11:29:12 crc kubenswrapper[4824]: I0121 11:29:12.871374 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Jan 21 11:29:13 crc kubenswrapper[4824]: I0121 11:29:13.158249 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Jan 21 11:29:13 crc kubenswrapper[4824]: I0121 11:29:13.158466 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Jan 21 11:29:13 crc kubenswrapper[4824]: I0121 11:29:13.165912 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Jan 21 11:29:13 crc kubenswrapper[4824]: I0121 11:29:13.165943 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Jan 21 11:29:13 crc kubenswrapper[4824]: I0121 11:29:13.720688 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Jan 21 11:29:13 crc kubenswrapper[4824]: I0121 11:29:13.720872 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="2e3a943c-ee06-4e54-a984-34ecb1320f89" containerName="kube-state-metrics" containerID="cri-o://6e0a2a51cb10bfe8bc65ec0123c4fa76489cf9873a673344ebc1fbf8065155d5" gracePeriod=30 Jan 21 11:29:13 crc kubenswrapper[4824]: I0121 11:29:13.864050 4824 generic.go:334] "Generic (PLEG): container finished" podID="2e3a943c-ee06-4e54-a984-34ecb1320f89" containerID="6e0a2a51cb10bfe8bc65ec0123c4fa76489cf9873a673344ebc1fbf8065155d5" exitCode=2 Jan 21 11:29:13 crc kubenswrapper[4824]: I0121 11:29:13.864754 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"2e3a943c-ee06-4e54-a984-34ecb1320f89","Type":"ContainerDied","Data":"6e0a2a51cb10bfe8bc65ec0123c4fa76489cf9873a673344ebc1fbf8065155d5"} Jan 21 11:29:14 crc kubenswrapper[4824]: I0121 11:29:14.122973 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Jan 21 11:29:14 crc kubenswrapper[4824]: I0121 11:29:14.168105 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="55525bd3-ae2b-494d-9a33-dd7d00c576b2" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.212:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Jan 21 11:29:14 crc kubenswrapper[4824]: I0121 11:29:14.168129 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="55525bd3-ae2b-494d-9a33-dd7d00c576b2" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.212:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Jan 21 11:29:14 crc kubenswrapper[4824]: I0121 11:29:14.182347 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="654a41d0-b44a-4194-a514-a3c60126d0d9" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.213:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Jan 21 11:29:14 crc kubenswrapper[4824]: I0121 11:29:14.182505 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="654a41d0-b44a-4194-a514-a3c60126d0d9" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.213:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Jan 21 11:29:14 crc kubenswrapper[4824]: I0121 11:29:14.199651 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n6ndv\" (UniqueName: \"kubernetes.io/projected/2e3a943c-ee06-4e54-a984-34ecb1320f89-kube-api-access-n6ndv\") pod \"2e3a943c-ee06-4e54-a984-34ecb1320f89\" (UID: \"2e3a943c-ee06-4e54-a984-34ecb1320f89\") " Jan 21 11:29:14 crc kubenswrapper[4824]: I0121 11:29:14.206180 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2e3a943c-ee06-4e54-a984-34ecb1320f89-kube-api-access-n6ndv" (OuterVolumeSpecName: "kube-api-access-n6ndv") pod "2e3a943c-ee06-4e54-a984-34ecb1320f89" (UID: "2e3a943c-ee06-4e54-a984-34ecb1320f89"). InnerVolumeSpecName "kube-api-access-n6ndv". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:29:14 crc kubenswrapper[4824]: I0121 11:29:14.302069 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n6ndv\" (UniqueName: \"kubernetes.io/projected/2e3a943c-ee06-4e54-a984-34ecb1320f89-kube-api-access-n6ndv\") on node \"crc\" DevicePath \"\"" Jan 21 11:29:14 crc kubenswrapper[4824]: I0121 11:29:14.871876 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"2e3a943c-ee06-4e54-a984-34ecb1320f89","Type":"ContainerDied","Data":"6cb5ad5332507e249a47db656294541a580560b8688929dae90a9899b846d8f8"} Jan 21 11:29:14 crc kubenswrapper[4824]: I0121 11:29:14.871916 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Jan 21 11:29:14 crc kubenswrapper[4824]: I0121 11:29:14.871924 4824 scope.go:117] "RemoveContainer" containerID="6e0a2a51cb10bfe8bc65ec0123c4fa76489cf9873a673344ebc1fbf8065155d5" Jan 21 11:29:14 crc kubenswrapper[4824]: I0121 11:29:14.896762 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Jan 21 11:29:14 crc kubenswrapper[4824]: I0121 11:29:14.940525 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Jan 21 11:29:14 crc kubenswrapper[4824]: I0121 11:29:14.950489 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Jan 21 11:29:14 crc kubenswrapper[4824]: E0121 11:29:14.950881 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e3a943c-ee06-4e54-a984-34ecb1320f89" containerName="kube-state-metrics" Jan 21 11:29:14 crc kubenswrapper[4824]: I0121 11:29:14.950898 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e3a943c-ee06-4e54-a984-34ecb1320f89" containerName="kube-state-metrics" Jan 21 11:29:14 crc kubenswrapper[4824]: I0121 11:29:14.951114 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="2e3a943c-ee06-4e54-a984-34ecb1320f89" containerName="kube-state-metrics" Jan 21 11:29:14 crc kubenswrapper[4824]: I0121 11:29:14.951705 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Jan 21 11:29:14 crc kubenswrapper[4824]: I0121 11:29:14.953385 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Jan 21 11:29:14 crc kubenswrapper[4824]: I0121 11:29:14.953592 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Jan 21 11:29:14 crc kubenswrapper[4824]: I0121 11:29:14.972010 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Jan 21 11:29:15 crc kubenswrapper[4824]: I0121 11:29:15.020450 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/d42233a9-cfd9-486d-b314-f17d6a9d3b03-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"d42233a9-cfd9-486d-b314-f17d6a9d3b03\") " pod="openstack/kube-state-metrics-0" Jan 21 11:29:15 crc kubenswrapper[4824]: I0121 11:29:15.020528 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r25fc\" (UniqueName: \"kubernetes.io/projected/d42233a9-cfd9-486d-b314-f17d6a9d3b03-kube-api-access-r25fc\") pod \"kube-state-metrics-0\" (UID: \"d42233a9-cfd9-486d-b314-f17d6a9d3b03\") " pod="openstack/kube-state-metrics-0" Jan 21 11:29:15 crc kubenswrapper[4824]: I0121 11:29:15.020558 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/d42233a9-cfd9-486d-b314-f17d6a9d3b03-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"d42233a9-cfd9-486d-b314-f17d6a9d3b03\") " pod="openstack/kube-state-metrics-0" Jan 21 11:29:15 crc kubenswrapper[4824]: I0121 11:29:15.020779 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d42233a9-cfd9-486d-b314-f17d6a9d3b03-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"d42233a9-cfd9-486d-b314-f17d6a9d3b03\") " pod="openstack/kube-state-metrics-0" Jan 21 11:29:15 crc kubenswrapper[4824]: I0121 11:29:15.122483 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d42233a9-cfd9-486d-b314-f17d6a9d3b03-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"d42233a9-cfd9-486d-b314-f17d6a9d3b03\") " pod="openstack/kube-state-metrics-0" Jan 21 11:29:15 crc kubenswrapper[4824]: I0121 11:29:15.122723 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/d42233a9-cfd9-486d-b314-f17d6a9d3b03-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"d42233a9-cfd9-486d-b314-f17d6a9d3b03\") " pod="openstack/kube-state-metrics-0" Jan 21 11:29:15 crc kubenswrapper[4824]: I0121 11:29:15.122765 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r25fc\" (UniqueName: \"kubernetes.io/projected/d42233a9-cfd9-486d-b314-f17d6a9d3b03-kube-api-access-r25fc\") pod \"kube-state-metrics-0\" (UID: \"d42233a9-cfd9-486d-b314-f17d6a9d3b03\") " pod="openstack/kube-state-metrics-0" Jan 21 11:29:15 crc kubenswrapper[4824]: I0121 11:29:15.122789 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/d42233a9-cfd9-486d-b314-f17d6a9d3b03-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"d42233a9-cfd9-486d-b314-f17d6a9d3b03\") " pod="openstack/kube-state-metrics-0" Jan 21 11:29:15 crc kubenswrapper[4824]: I0121 11:29:15.130295 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d42233a9-cfd9-486d-b314-f17d6a9d3b03-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"d42233a9-cfd9-486d-b314-f17d6a9d3b03\") " pod="openstack/kube-state-metrics-0" Jan 21 11:29:15 crc kubenswrapper[4824]: I0121 11:29:15.131313 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/d42233a9-cfd9-486d-b314-f17d6a9d3b03-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"d42233a9-cfd9-486d-b314-f17d6a9d3b03\") " pod="openstack/kube-state-metrics-0" Jan 21 11:29:15 crc kubenswrapper[4824]: I0121 11:29:15.141497 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r25fc\" (UniqueName: \"kubernetes.io/projected/d42233a9-cfd9-486d-b314-f17d6a9d3b03-kube-api-access-r25fc\") pod \"kube-state-metrics-0\" (UID: \"d42233a9-cfd9-486d-b314-f17d6a9d3b03\") " pod="openstack/kube-state-metrics-0" Jan 21 11:29:15 crc kubenswrapper[4824]: I0121 11:29:15.141517 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/d42233a9-cfd9-486d-b314-f17d6a9d3b03-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"d42233a9-cfd9-486d-b314-f17d6a9d3b03\") " pod="openstack/kube-state-metrics-0" Jan 21 11:29:15 crc kubenswrapper[4824]: I0121 11:29:15.267994 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Jan 21 11:29:15 crc kubenswrapper[4824]: I0121 11:29:15.291732 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Jan 21 11:29:15 crc kubenswrapper[4824]: I0121 11:29:15.291971 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="55916065-772b-45ee-8871-37c229777deb" containerName="ceilometer-central-agent" containerID="cri-o://f6009500618f69ecbe25f69878a4c52dce4019a903ece88ba0ef9f2b52fe0e5f" gracePeriod=30 Jan 21 11:29:15 crc kubenswrapper[4824]: I0121 11:29:15.292030 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="55916065-772b-45ee-8871-37c229777deb" containerName="sg-core" containerID="cri-o://ceacb0570b30b00232645caa9a72628c88bd5083b2072630b2e8ba9033c27f4a" gracePeriod=30 Jan 21 11:29:15 crc kubenswrapper[4824]: I0121 11:29:15.292057 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="55916065-772b-45ee-8871-37c229777deb" containerName="proxy-httpd" containerID="cri-o://0949e2b629f2a233cd433291c41e4a65cbb4c871e7cd93788379c3836b2e6e74" gracePeriod=30 Jan 21 11:29:15 crc kubenswrapper[4824]: I0121 11:29:15.292037 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="55916065-772b-45ee-8871-37c229777deb" containerName="ceilometer-notification-agent" containerID="cri-o://ee73eb7a1ab0880b735d27e1db094d3fe0d66679a0480addcb5e49e67621c253" gracePeriod=30 Jan 21 11:29:15 crc kubenswrapper[4824]: I0121 11:29:15.664557 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Jan 21 11:29:15 crc kubenswrapper[4824]: W0121 11:29:15.672113 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd42233a9_cfd9_486d_b314_f17d6a9d3b03.slice/crio-bbba9f8a9650e396c656bdb12251ced4c13ba8643f286ba15c3458435cec5ce0 WatchSource:0}: Error finding container bbba9f8a9650e396c656bdb12251ced4c13ba8643f286ba15c3458435cec5ce0: Status 404 returned error can't find the container with id bbba9f8a9650e396c656bdb12251ced4c13ba8643f286ba15c3458435cec5ce0 Jan 21 11:29:15 crc kubenswrapper[4824]: I0121 11:29:15.879089 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"d42233a9-cfd9-486d-b314-f17d6a9d3b03","Type":"ContainerStarted","Data":"bbba9f8a9650e396c656bdb12251ced4c13ba8643f286ba15c3458435cec5ce0"} Jan 21 11:29:15 crc kubenswrapper[4824]: I0121 11:29:15.882123 4824 generic.go:334] "Generic (PLEG): container finished" podID="55916065-772b-45ee-8871-37c229777deb" containerID="0949e2b629f2a233cd433291c41e4a65cbb4c871e7cd93788379c3836b2e6e74" exitCode=0 Jan 21 11:29:15 crc kubenswrapper[4824]: I0121 11:29:15.882148 4824 generic.go:334] "Generic (PLEG): container finished" podID="55916065-772b-45ee-8871-37c229777deb" containerID="ceacb0570b30b00232645caa9a72628c88bd5083b2072630b2e8ba9033c27f4a" exitCode=2 Jan 21 11:29:15 crc kubenswrapper[4824]: I0121 11:29:15.882157 4824 generic.go:334] "Generic (PLEG): container finished" podID="55916065-772b-45ee-8871-37c229777deb" containerID="f6009500618f69ecbe25f69878a4c52dce4019a903ece88ba0ef9f2b52fe0e5f" exitCode=0 Jan 21 11:29:15 crc kubenswrapper[4824]: I0121 11:29:15.882170 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"55916065-772b-45ee-8871-37c229777deb","Type":"ContainerDied","Data":"0949e2b629f2a233cd433291c41e4a65cbb4c871e7cd93788379c3836b2e6e74"} Jan 21 11:29:15 crc kubenswrapper[4824]: I0121 11:29:15.882185 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"55916065-772b-45ee-8871-37c229777deb","Type":"ContainerDied","Data":"ceacb0570b30b00232645caa9a72628c88bd5083b2072630b2e8ba9033c27f4a"} Jan 21 11:29:15 crc kubenswrapper[4824]: I0121 11:29:15.882194 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"55916065-772b-45ee-8871-37c229777deb","Type":"ContainerDied","Data":"f6009500618f69ecbe25f69878a4c52dce4019a903ece88ba0ef9f2b52fe0e5f"} Jan 21 11:29:16 crc kubenswrapper[4824]: I0121 11:29:16.057636 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2e3a943c-ee06-4e54-a984-34ecb1320f89" path="/var/lib/kubelet/pods/2e3a943c-ee06-4e54-a984-34ecb1320f89/volumes" Jan 21 11:29:16 crc kubenswrapper[4824]: I0121 11:29:16.064835 4824 patch_prober.go:28] interesting pod/machine-config-daemon-zqjg9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 21 11:29:16 crc kubenswrapper[4824]: I0121 11:29:16.064879 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 21 11:29:16 crc kubenswrapper[4824]: I0121 11:29:16.894371 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"d42233a9-cfd9-486d-b314-f17d6a9d3b03","Type":"ContainerStarted","Data":"29ffe115de3536006b5bf65e63db5d06e9c710a58c3108031b8df7712c9cdfe7"} Jan 21 11:29:16 crc kubenswrapper[4824]: I0121 11:29:16.894700 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Jan 21 11:29:16 crc kubenswrapper[4824]: I0121 11:29:16.913200 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=2.646854743 podStartE2EDuration="2.913186132s" podCreationTimestamp="2026-01-21 11:29:14 +0000 UTC" firstStartedPulling="2026-01-21 11:29:15.674439131 +0000 UTC m=+1097.967468423" lastFinishedPulling="2026-01-21 11:29:15.94077052 +0000 UTC m=+1098.233799812" observedRunningTime="2026-01-21 11:29:16.906544779 +0000 UTC m=+1099.199574071" watchObservedRunningTime="2026-01-21 11:29:16.913186132 +0000 UTC m=+1099.206215424" Jan 21 11:29:17 crc kubenswrapper[4824]: I0121 11:29:17.376196 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 21 11:29:17 crc kubenswrapper[4824]: I0121 11:29:17.464786 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/55916065-772b-45ee-8871-37c229777deb-log-httpd\") pod \"55916065-772b-45ee-8871-37c229777deb\" (UID: \"55916065-772b-45ee-8871-37c229777deb\") " Jan 21 11:29:17 crc kubenswrapper[4824]: I0121 11:29:17.464851 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/55916065-772b-45ee-8871-37c229777deb-config-data\") pod \"55916065-772b-45ee-8871-37c229777deb\" (UID: \"55916065-772b-45ee-8871-37c229777deb\") " Jan 21 11:29:17 crc kubenswrapper[4824]: I0121 11:29:17.464874 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/55916065-772b-45ee-8871-37c229777deb-scripts\") pod \"55916065-772b-45ee-8871-37c229777deb\" (UID: \"55916065-772b-45ee-8871-37c229777deb\") " Jan 21 11:29:17 crc kubenswrapper[4824]: I0121 11:29:17.464895 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/55916065-772b-45ee-8871-37c229777deb-run-httpd\") pod \"55916065-772b-45ee-8871-37c229777deb\" (UID: \"55916065-772b-45ee-8871-37c229777deb\") " Jan 21 11:29:17 crc kubenswrapper[4824]: I0121 11:29:17.464939 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/55916065-772b-45ee-8871-37c229777deb-sg-core-conf-yaml\") pod \"55916065-772b-45ee-8871-37c229777deb\" (UID: \"55916065-772b-45ee-8871-37c229777deb\") " Jan 21 11:29:17 crc kubenswrapper[4824]: I0121 11:29:17.465036 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gvljc\" (UniqueName: \"kubernetes.io/projected/55916065-772b-45ee-8871-37c229777deb-kube-api-access-gvljc\") pod \"55916065-772b-45ee-8871-37c229777deb\" (UID: \"55916065-772b-45ee-8871-37c229777deb\") " Jan 21 11:29:17 crc kubenswrapper[4824]: I0121 11:29:17.465072 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/55916065-772b-45ee-8871-37c229777deb-combined-ca-bundle\") pod \"55916065-772b-45ee-8871-37c229777deb\" (UID: \"55916065-772b-45ee-8871-37c229777deb\") " Jan 21 11:29:17 crc kubenswrapper[4824]: I0121 11:29:17.466066 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/55916065-772b-45ee-8871-37c229777deb-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "55916065-772b-45ee-8871-37c229777deb" (UID: "55916065-772b-45ee-8871-37c229777deb"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:29:17 crc kubenswrapper[4824]: I0121 11:29:17.466177 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/55916065-772b-45ee-8871-37c229777deb-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "55916065-772b-45ee-8871-37c229777deb" (UID: "55916065-772b-45ee-8871-37c229777deb"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:29:17 crc kubenswrapper[4824]: I0121 11:29:17.469700 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/55916065-772b-45ee-8871-37c229777deb-kube-api-access-gvljc" (OuterVolumeSpecName: "kube-api-access-gvljc") pod "55916065-772b-45ee-8871-37c229777deb" (UID: "55916065-772b-45ee-8871-37c229777deb"). InnerVolumeSpecName "kube-api-access-gvljc". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:29:17 crc kubenswrapper[4824]: I0121 11:29:17.473534 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/55916065-772b-45ee-8871-37c229777deb-scripts" (OuterVolumeSpecName: "scripts") pod "55916065-772b-45ee-8871-37c229777deb" (UID: "55916065-772b-45ee-8871-37c229777deb"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:29:17 crc kubenswrapper[4824]: I0121 11:29:17.486299 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/55916065-772b-45ee-8871-37c229777deb-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "55916065-772b-45ee-8871-37c229777deb" (UID: "55916065-772b-45ee-8871-37c229777deb"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:29:17 crc kubenswrapper[4824]: I0121 11:29:17.518433 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/55916065-772b-45ee-8871-37c229777deb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "55916065-772b-45ee-8871-37c229777deb" (UID: "55916065-772b-45ee-8871-37c229777deb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:29:17 crc kubenswrapper[4824]: I0121 11:29:17.536125 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/55916065-772b-45ee-8871-37c229777deb-config-data" (OuterVolumeSpecName: "config-data") pod "55916065-772b-45ee-8871-37c229777deb" (UID: "55916065-772b-45ee-8871-37c229777deb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:29:17 crc kubenswrapper[4824]: I0121 11:29:17.567461 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gvljc\" (UniqueName: \"kubernetes.io/projected/55916065-772b-45ee-8871-37c229777deb-kube-api-access-gvljc\") on node \"crc\" DevicePath \"\"" Jan 21 11:29:17 crc kubenswrapper[4824]: I0121 11:29:17.567487 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/55916065-772b-45ee-8871-37c229777deb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 21 11:29:17 crc kubenswrapper[4824]: I0121 11:29:17.567496 4824 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/55916065-772b-45ee-8871-37c229777deb-log-httpd\") on node \"crc\" DevicePath \"\"" Jan 21 11:29:17 crc kubenswrapper[4824]: I0121 11:29:17.567503 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/55916065-772b-45ee-8871-37c229777deb-config-data\") on node \"crc\" DevicePath \"\"" Jan 21 11:29:17 crc kubenswrapper[4824]: I0121 11:29:17.567510 4824 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/55916065-772b-45ee-8871-37c229777deb-scripts\") on node \"crc\" DevicePath \"\"" Jan 21 11:29:17 crc kubenswrapper[4824]: I0121 11:29:17.567517 4824 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/55916065-772b-45ee-8871-37c229777deb-run-httpd\") on node \"crc\" DevicePath \"\"" Jan 21 11:29:17 crc kubenswrapper[4824]: I0121 11:29:17.567526 4824 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/55916065-772b-45ee-8871-37c229777deb-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Jan 21 11:29:17 crc kubenswrapper[4824]: I0121 11:29:17.903116 4824 generic.go:334] "Generic (PLEG): container finished" podID="55916065-772b-45ee-8871-37c229777deb" containerID="ee73eb7a1ab0880b735d27e1db094d3fe0d66679a0480addcb5e49e67621c253" exitCode=0 Jan 21 11:29:17 crc kubenswrapper[4824]: I0121 11:29:17.903172 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 21 11:29:17 crc kubenswrapper[4824]: I0121 11:29:17.903198 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"55916065-772b-45ee-8871-37c229777deb","Type":"ContainerDied","Data":"ee73eb7a1ab0880b735d27e1db094d3fe0d66679a0480addcb5e49e67621c253"} Jan 21 11:29:17 crc kubenswrapper[4824]: I0121 11:29:17.903444 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"55916065-772b-45ee-8871-37c229777deb","Type":"ContainerDied","Data":"65e4d9ce199d66a5d4b888b55edee4667e5ba88065fc68ccd26d1ad79dbffbb0"} Jan 21 11:29:17 crc kubenswrapper[4824]: I0121 11:29:17.903467 4824 scope.go:117] "RemoveContainer" containerID="0949e2b629f2a233cd433291c41e4a65cbb4c871e7cd93788379c3836b2e6e74" Jan 21 11:29:17 crc kubenswrapper[4824]: I0121 11:29:17.918097 4824 scope.go:117] "RemoveContainer" containerID="ceacb0570b30b00232645caa9a72628c88bd5083b2072630b2e8ba9033c27f4a" Jan 21 11:29:17 crc kubenswrapper[4824]: I0121 11:29:17.929905 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Jan 21 11:29:17 crc kubenswrapper[4824]: I0121 11:29:17.934726 4824 scope.go:117] "RemoveContainer" containerID="ee73eb7a1ab0880b735d27e1db094d3fe0d66679a0480addcb5e49e67621c253" Jan 21 11:29:17 crc kubenswrapper[4824]: I0121 11:29:17.939035 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Jan 21 11:29:17 crc kubenswrapper[4824]: I0121 11:29:17.951045 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Jan 21 11:29:17 crc kubenswrapper[4824]: E0121 11:29:17.951462 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="55916065-772b-45ee-8871-37c229777deb" containerName="ceilometer-central-agent" Jan 21 11:29:17 crc kubenswrapper[4824]: I0121 11:29:17.951496 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="55916065-772b-45ee-8871-37c229777deb" containerName="ceilometer-central-agent" Jan 21 11:29:17 crc kubenswrapper[4824]: E0121 11:29:17.951514 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="55916065-772b-45ee-8871-37c229777deb" containerName="sg-core" Jan 21 11:29:17 crc kubenswrapper[4824]: I0121 11:29:17.951521 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="55916065-772b-45ee-8871-37c229777deb" containerName="sg-core" Jan 21 11:29:17 crc kubenswrapper[4824]: E0121 11:29:17.951535 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="55916065-772b-45ee-8871-37c229777deb" containerName="proxy-httpd" Jan 21 11:29:17 crc kubenswrapper[4824]: I0121 11:29:17.951540 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="55916065-772b-45ee-8871-37c229777deb" containerName="proxy-httpd" Jan 21 11:29:17 crc kubenswrapper[4824]: E0121 11:29:17.951581 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="55916065-772b-45ee-8871-37c229777deb" containerName="ceilometer-notification-agent" Jan 21 11:29:17 crc kubenswrapper[4824]: I0121 11:29:17.951588 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="55916065-772b-45ee-8871-37c229777deb" containerName="ceilometer-notification-agent" Jan 21 11:29:17 crc kubenswrapper[4824]: I0121 11:29:17.951778 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="55916065-772b-45ee-8871-37c229777deb" containerName="ceilometer-central-agent" Jan 21 11:29:17 crc kubenswrapper[4824]: I0121 11:29:17.951825 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="55916065-772b-45ee-8871-37c229777deb" containerName="proxy-httpd" Jan 21 11:29:17 crc kubenswrapper[4824]: I0121 11:29:17.951834 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="55916065-772b-45ee-8871-37c229777deb" containerName="ceilometer-notification-agent" Jan 21 11:29:17 crc kubenswrapper[4824]: I0121 11:29:17.951852 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="55916065-772b-45ee-8871-37c229777deb" containerName="sg-core" Jan 21 11:29:17 crc kubenswrapper[4824]: I0121 11:29:17.954550 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 21 11:29:17 crc kubenswrapper[4824]: I0121 11:29:17.956498 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Jan 21 11:29:17 crc kubenswrapper[4824]: I0121 11:29:17.956852 4824 scope.go:117] "RemoveContainer" containerID="f6009500618f69ecbe25f69878a4c52dce4019a903ece88ba0ef9f2b52fe0e5f" Jan 21 11:29:17 crc kubenswrapper[4824]: I0121 11:29:17.956866 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Jan 21 11:29:17 crc kubenswrapper[4824]: I0121 11:29:17.956921 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Jan 21 11:29:17 crc kubenswrapper[4824]: I0121 11:29:17.956997 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Jan 21 11:29:17 crc kubenswrapper[4824]: I0121 11:29:17.972910 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/adc98d07-2b32-4f30-bdb7-40923468389e-scripts\") pod \"ceilometer-0\" (UID: \"adc98d07-2b32-4f30-bdb7-40923468389e\") " pod="openstack/ceilometer-0" Jan 21 11:29:17 crc kubenswrapper[4824]: I0121 11:29:17.973065 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/adc98d07-2b32-4f30-bdb7-40923468389e-config-data\") pod \"ceilometer-0\" (UID: \"adc98d07-2b32-4f30-bdb7-40923468389e\") " pod="openstack/ceilometer-0" Jan 21 11:29:17 crc kubenswrapper[4824]: I0121 11:29:17.973188 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/adc98d07-2b32-4f30-bdb7-40923468389e-run-httpd\") pod \"ceilometer-0\" (UID: \"adc98d07-2b32-4f30-bdb7-40923468389e\") " pod="openstack/ceilometer-0" Jan 21 11:29:17 crc kubenswrapper[4824]: I0121 11:29:17.973293 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/adc98d07-2b32-4f30-bdb7-40923468389e-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"adc98d07-2b32-4f30-bdb7-40923468389e\") " pod="openstack/ceilometer-0" Jan 21 11:29:17 crc kubenswrapper[4824]: I0121 11:29:17.973385 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/adc98d07-2b32-4f30-bdb7-40923468389e-log-httpd\") pod \"ceilometer-0\" (UID: \"adc98d07-2b32-4f30-bdb7-40923468389e\") " pod="openstack/ceilometer-0" Jan 21 11:29:17 crc kubenswrapper[4824]: I0121 11:29:17.973470 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/adc98d07-2b32-4f30-bdb7-40923468389e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"adc98d07-2b32-4f30-bdb7-40923468389e\") " pod="openstack/ceilometer-0" Jan 21 11:29:17 crc kubenswrapper[4824]: I0121 11:29:17.973701 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qjpbm\" (UniqueName: \"kubernetes.io/projected/adc98d07-2b32-4f30-bdb7-40923468389e-kube-api-access-qjpbm\") pod \"ceilometer-0\" (UID: \"adc98d07-2b32-4f30-bdb7-40923468389e\") " pod="openstack/ceilometer-0" Jan 21 11:29:17 crc kubenswrapper[4824]: I0121 11:29:17.973864 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/adc98d07-2b32-4f30-bdb7-40923468389e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"adc98d07-2b32-4f30-bdb7-40923468389e\") " pod="openstack/ceilometer-0" Jan 21 11:29:17 crc kubenswrapper[4824]: I0121 11:29:17.980251 4824 scope.go:117] "RemoveContainer" containerID="0949e2b629f2a233cd433291c41e4a65cbb4c871e7cd93788379c3836b2e6e74" Jan 21 11:29:17 crc kubenswrapper[4824]: E0121 11:29:17.980528 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0949e2b629f2a233cd433291c41e4a65cbb4c871e7cd93788379c3836b2e6e74\": container with ID starting with 0949e2b629f2a233cd433291c41e4a65cbb4c871e7cd93788379c3836b2e6e74 not found: ID does not exist" containerID="0949e2b629f2a233cd433291c41e4a65cbb4c871e7cd93788379c3836b2e6e74" Jan 21 11:29:17 crc kubenswrapper[4824]: I0121 11:29:17.980558 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0949e2b629f2a233cd433291c41e4a65cbb4c871e7cd93788379c3836b2e6e74"} err="failed to get container status \"0949e2b629f2a233cd433291c41e4a65cbb4c871e7cd93788379c3836b2e6e74\": rpc error: code = NotFound desc = could not find container \"0949e2b629f2a233cd433291c41e4a65cbb4c871e7cd93788379c3836b2e6e74\": container with ID starting with 0949e2b629f2a233cd433291c41e4a65cbb4c871e7cd93788379c3836b2e6e74 not found: ID does not exist" Jan 21 11:29:17 crc kubenswrapper[4824]: I0121 11:29:17.980575 4824 scope.go:117] "RemoveContainer" containerID="ceacb0570b30b00232645caa9a72628c88bd5083b2072630b2e8ba9033c27f4a" Jan 21 11:29:17 crc kubenswrapper[4824]: E0121 11:29:17.980918 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ceacb0570b30b00232645caa9a72628c88bd5083b2072630b2e8ba9033c27f4a\": container with ID starting with ceacb0570b30b00232645caa9a72628c88bd5083b2072630b2e8ba9033c27f4a not found: ID does not exist" containerID="ceacb0570b30b00232645caa9a72628c88bd5083b2072630b2e8ba9033c27f4a" Jan 21 11:29:17 crc kubenswrapper[4824]: I0121 11:29:17.980940 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ceacb0570b30b00232645caa9a72628c88bd5083b2072630b2e8ba9033c27f4a"} err="failed to get container status \"ceacb0570b30b00232645caa9a72628c88bd5083b2072630b2e8ba9033c27f4a\": rpc error: code = NotFound desc = could not find container \"ceacb0570b30b00232645caa9a72628c88bd5083b2072630b2e8ba9033c27f4a\": container with ID starting with ceacb0570b30b00232645caa9a72628c88bd5083b2072630b2e8ba9033c27f4a not found: ID does not exist" Jan 21 11:29:17 crc kubenswrapper[4824]: I0121 11:29:17.980967 4824 scope.go:117] "RemoveContainer" containerID="ee73eb7a1ab0880b735d27e1db094d3fe0d66679a0480addcb5e49e67621c253" Jan 21 11:29:17 crc kubenswrapper[4824]: E0121 11:29:17.981265 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ee73eb7a1ab0880b735d27e1db094d3fe0d66679a0480addcb5e49e67621c253\": container with ID starting with ee73eb7a1ab0880b735d27e1db094d3fe0d66679a0480addcb5e49e67621c253 not found: ID does not exist" containerID="ee73eb7a1ab0880b735d27e1db094d3fe0d66679a0480addcb5e49e67621c253" Jan 21 11:29:17 crc kubenswrapper[4824]: I0121 11:29:17.981285 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ee73eb7a1ab0880b735d27e1db094d3fe0d66679a0480addcb5e49e67621c253"} err="failed to get container status \"ee73eb7a1ab0880b735d27e1db094d3fe0d66679a0480addcb5e49e67621c253\": rpc error: code = NotFound desc = could not find container \"ee73eb7a1ab0880b735d27e1db094d3fe0d66679a0480addcb5e49e67621c253\": container with ID starting with ee73eb7a1ab0880b735d27e1db094d3fe0d66679a0480addcb5e49e67621c253 not found: ID does not exist" Jan 21 11:29:17 crc kubenswrapper[4824]: I0121 11:29:17.981297 4824 scope.go:117] "RemoveContainer" containerID="f6009500618f69ecbe25f69878a4c52dce4019a903ece88ba0ef9f2b52fe0e5f" Jan 21 11:29:17 crc kubenswrapper[4824]: E0121 11:29:17.981463 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f6009500618f69ecbe25f69878a4c52dce4019a903ece88ba0ef9f2b52fe0e5f\": container with ID starting with f6009500618f69ecbe25f69878a4c52dce4019a903ece88ba0ef9f2b52fe0e5f not found: ID does not exist" containerID="f6009500618f69ecbe25f69878a4c52dce4019a903ece88ba0ef9f2b52fe0e5f" Jan 21 11:29:17 crc kubenswrapper[4824]: I0121 11:29:17.981481 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f6009500618f69ecbe25f69878a4c52dce4019a903ece88ba0ef9f2b52fe0e5f"} err="failed to get container status \"f6009500618f69ecbe25f69878a4c52dce4019a903ece88ba0ef9f2b52fe0e5f\": rpc error: code = NotFound desc = could not find container \"f6009500618f69ecbe25f69878a4c52dce4019a903ece88ba0ef9f2b52fe0e5f\": container with ID starting with f6009500618f69ecbe25f69878a4c52dce4019a903ece88ba0ef9f2b52fe0e5f not found: ID does not exist" Jan 21 11:29:18 crc kubenswrapper[4824]: I0121 11:29:18.063261 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="55916065-772b-45ee-8871-37c229777deb" path="/var/lib/kubelet/pods/55916065-772b-45ee-8871-37c229777deb/volumes" Jan 21 11:29:18 crc kubenswrapper[4824]: I0121 11:29:18.074824 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/adc98d07-2b32-4f30-bdb7-40923468389e-run-httpd\") pod \"ceilometer-0\" (UID: \"adc98d07-2b32-4f30-bdb7-40923468389e\") " pod="openstack/ceilometer-0" Jan 21 11:29:18 crc kubenswrapper[4824]: I0121 11:29:18.074884 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/adc98d07-2b32-4f30-bdb7-40923468389e-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"adc98d07-2b32-4f30-bdb7-40923468389e\") " pod="openstack/ceilometer-0" Jan 21 11:29:18 crc kubenswrapper[4824]: I0121 11:29:18.074932 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/adc98d07-2b32-4f30-bdb7-40923468389e-log-httpd\") pod \"ceilometer-0\" (UID: \"adc98d07-2b32-4f30-bdb7-40923468389e\") " pod="openstack/ceilometer-0" Jan 21 11:29:18 crc kubenswrapper[4824]: I0121 11:29:18.074980 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/adc98d07-2b32-4f30-bdb7-40923468389e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"adc98d07-2b32-4f30-bdb7-40923468389e\") " pod="openstack/ceilometer-0" Jan 21 11:29:18 crc kubenswrapper[4824]: I0121 11:29:18.075044 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qjpbm\" (UniqueName: \"kubernetes.io/projected/adc98d07-2b32-4f30-bdb7-40923468389e-kube-api-access-qjpbm\") pod \"ceilometer-0\" (UID: \"adc98d07-2b32-4f30-bdb7-40923468389e\") " pod="openstack/ceilometer-0" Jan 21 11:29:18 crc kubenswrapper[4824]: I0121 11:29:18.075104 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/adc98d07-2b32-4f30-bdb7-40923468389e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"adc98d07-2b32-4f30-bdb7-40923468389e\") " pod="openstack/ceilometer-0" Jan 21 11:29:18 crc kubenswrapper[4824]: I0121 11:29:18.075124 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/adc98d07-2b32-4f30-bdb7-40923468389e-scripts\") pod \"ceilometer-0\" (UID: \"adc98d07-2b32-4f30-bdb7-40923468389e\") " pod="openstack/ceilometer-0" Jan 21 11:29:18 crc kubenswrapper[4824]: I0121 11:29:18.075138 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/adc98d07-2b32-4f30-bdb7-40923468389e-config-data\") pod \"ceilometer-0\" (UID: \"adc98d07-2b32-4f30-bdb7-40923468389e\") " pod="openstack/ceilometer-0" Jan 21 11:29:18 crc kubenswrapper[4824]: I0121 11:29:18.076209 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/adc98d07-2b32-4f30-bdb7-40923468389e-log-httpd\") pod \"ceilometer-0\" (UID: \"adc98d07-2b32-4f30-bdb7-40923468389e\") " pod="openstack/ceilometer-0" Jan 21 11:29:18 crc kubenswrapper[4824]: I0121 11:29:18.076641 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/adc98d07-2b32-4f30-bdb7-40923468389e-run-httpd\") pod \"ceilometer-0\" (UID: \"adc98d07-2b32-4f30-bdb7-40923468389e\") " pod="openstack/ceilometer-0" Jan 21 11:29:18 crc kubenswrapper[4824]: I0121 11:29:18.078484 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/adc98d07-2b32-4f30-bdb7-40923468389e-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"adc98d07-2b32-4f30-bdb7-40923468389e\") " pod="openstack/ceilometer-0" Jan 21 11:29:18 crc kubenswrapper[4824]: I0121 11:29:18.078633 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/adc98d07-2b32-4f30-bdb7-40923468389e-config-data\") pod \"ceilometer-0\" (UID: \"adc98d07-2b32-4f30-bdb7-40923468389e\") " pod="openstack/ceilometer-0" Jan 21 11:29:18 crc kubenswrapper[4824]: I0121 11:29:18.078855 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/adc98d07-2b32-4f30-bdb7-40923468389e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"adc98d07-2b32-4f30-bdb7-40923468389e\") " pod="openstack/ceilometer-0" Jan 21 11:29:18 crc kubenswrapper[4824]: I0121 11:29:18.079401 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/adc98d07-2b32-4f30-bdb7-40923468389e-scripts\") pod \"ceilometer-0\" (UID: \"adc98d07-2b32-4f30-bdb7-40923468389e\") " pod="openstack/ceilometer-0" Jan 21 11:29:18 crc kubenswrapper[4824]: I0121 11:29:18.080033 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/adc98d07-2b32-4f30-bdb7-40923468389e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"adc98d07-2b32-4f30-bdb7-40923468389e\") " pod="openstack/ceilometer-0" Jan 21 11:29:18 crc kubenswrapper[4824]: I0121 11:29:18.087645 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qjpbm\" (UniqueName: \"kubernetes.io/projected/adc98d07-2b32-4f30-bdb7-40923468389e-kube-api-access-qjpbm\") pod \"ceilometer-0\" (UID: \"adc98d07-2b32-4f30-bdb7-40923468389e\") " pod="openstack/ceilometer-0" Jan 21 11:29:18 crc kubenswrapper[4824]: I0121 11:29:18.277654 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 21 11:29:18 crc kubenswrapper[4824]: I0121 11:29:18.646710 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Jan 21 11:29:18 crc kubenswrapper[4824]: W0121 11:29:18.654249 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podadc98d07_2b32_4f30_bdb7_40923468389e.slice/crio-2916545b5864c587e6e609faa9526231fc3891c256061f4fb6e3de6ffb82750f WatchSource:0}: Error finding container 2916545b5864c587e6e609faa9526231fc3891c256061f4fb6e3de6ffb82750f: Status 404 returned error can't find the container with id 2916545b5864c587e6e609faa9526231fc3891c256061f4fb6e3de6ffb82750f Jan 21 11:29:18 crc kubenswrapper[4824]: I0121 11:29:18.911800 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"adc98d07-2b32-4f30-bdb7-40923468389e","Type":"ContainerStarted","Data":"2916545b5864c587e6e609faa9526231fc3891c256061f4fb6e3de6ffb82750f"} Jan 21 11:29:19 crc kubenswrapper[4824]: I0121 11:29:19.918978 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"adc98d07-2b32-4f30-bdb7-40923468389e","Type":"ContainerStarted","Data":"a90e401f791539feac129f45795db65909d69b5d726e41e25c99fb59745fa42e"} Jan 21 11:29:20 crc kubenswrapper[4824]: I0121 11:29:20.929678 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"adc98d07-2b32-4f30-bdb7-40923468389e","Type":"ContainerStarted","Data":"90a1264a92ff9f804e52949d65ec8a022c71f5ff93ed7e2a3db6674d2d1d512c"} Jan 21 11:29:20 crc kubenswrapper[4824]: I0121 11:29:20.930152 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"adc98d07-2b32-4f30-bdb7-40923468389e","Type":"ContainerStarted","Data":"b03e582376e52022beb2bcf9dc6dfe2b49f9da87daff5d1c7851a1edc48f84cb"} Jan 21 11:29:22 crc kubenswrapper[4824]: I0121 11:29:22.945177 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"adc98d07-2b32-4f30-bdb7-40923468389e","Type":"ContainerStarted","Data":"3d1045aabdaf7cb011d4cf495207121285049e47b5e1472749fd1020f5ee11e9"} Jan 21 11:29:22 crc kubenswrapper[4824]: I0121 11:29:22.945575 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Jan 21 11:29:22 crc kubenswrapper[4824]: I0121 11:29:22.966065 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.75376542 podStartE2EDuration="5.966050794s" podCreationTimestamp="2026-01-21 11:29:17 +0000 UTC" firstStartedPulling="2026-01-21 11:29:18.656472743 +0000 UTC m=+1100.949502035" lastFinishedPulling="2026-01-21 11:29:21.868758117 +0000 UTC m=+1104.161787409" observedRunningTime="2026-01-21 11:29:22.95960423 +0000 UTC m=+1105.252633532" watchObservedRunningTime="2026-01-21 11:29:22.966050794 +0000 UTC m=+1105.259080087" Jan 21 11:29:23 crc kubenswrapper[4824]: I0121 11:29:23.163089 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Jan 21 11:29:23 crc kubenswrapper[4824]: I0121 11:29:23.163404 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Jan 21 11:29:23 crc kubenswrapper[4824]: I0121 11:29:23.163555 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Jan 21 11:29:23 crc kubenswrapper[4824]: I0121 11:29:23.166889 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Jan 21 11:29:23 crc kubenswrapper[4824]: I0121 11:29:23.173208 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Jan 21 11:29:23 crc kubenswrapper[4824]: I0121 11:29:23.178794 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Jan 21 11:29:23 crc kubenswrapper[4824]: I0121 11:29:23.180554 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Jan 21 11:29:23 crc kubenswrapper[4824]: I0121 11:29:23.953816 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Jan 21 11:29:23 crc kubenswrapper[4824]: I0121 11:29:23.956628 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Jan 21 11:29:23 crc kubenswrapper[4824]: I0121 11:29:23.958397 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Jan 21 11:29:25 crc kubenswrapper[4824]: I0121 11:29:25.277569 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Jan 21 11:29:46 crc kubenswrapper[4824]: I0121 11:29:46.065175 4824 patch_prober.go:28] interesting pod/machine-config-daemon-zqjg9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 21 11:29:46 crc kubenswrapper[4824]: I0121 11:29:46.065487 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 21 11:29:48 crc kubenswrapper[4824]: I0121 11:29:48.285195 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Jan 21 11:29:55 crc kubenswrapper[4824]: I0121 11:29:55.955165 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Jan 21 11:29:56 crc kubenswrapper[4824]: I0121 11:29:56.679703 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Jan 21 11:29:59 crc kubenswrapper[4824]: I0121 11:29:59.347089 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="84e74d7b-18a0-4a3f-8680-6246ac538a6e" containerName="rabbitmq" containerID="cri-o://407422f3cadd164c7680cc8b0bf1b21f0ebb5d9d1b6255a64420fbd1143666e7" gracePeriod=604797 Jan 21 11:29:59 crc kubenswrapper[4824]: I0121 11:29:59.851710 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="156fec50-b486-4e84-a7bf-b40491a863c7" containerName="rabbitmq" containerID="cri-o://902d1a3e9a8d54f58be516ba97809a95404a4ebe849ca1cb1fe5805c23862e32" gracePeriod=604797 Jan 21 11:30:00 crc kubenswrapper[4824]: I0121 11:30:00.146395 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29483250-mrvrz"] Jan 21 11:30:00 crc kubenswrapper[4824]: I0121 11:30:00.147698 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29483250-mrvrz" Jan 21 11:30:00 crc kubenswrapper[4824]: I0121 11:30:00.149672 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Jan 21 11:30:00 crc kubenswrapper[4824]: I0121 11:30:00.150616 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Jan 21 11:30:00 crc kubenswrapper[4824]: I0121 11:30:00.153098 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29483250-mrvrz"] Jan 21 11:30:00 crc kubenswrapper[4824]: I0121 11:30:00.282466 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4c02bf47-f856-44f3-af2c-cc96c45471c1-config-volume\") pod \"collect-profiles-29483250-mrvrz\" (UID: \"4c02bf47-f856-44f3-af2c-cc96c45471c1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29483250-mrvrz" Jan 21 11:30:00 crc kubenswrapper[4824]: I0121 11:30:00.282942 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h6hhk\" (UniqueName: \"kubernetes.io/projected/4c02bf47-f856-44f3-af2c-cc96c45471c1-kube-api-access-h6hhk\") pod \"collect-profiles-29483250-mrvrz\" (UID: \"4c02bf47-f856-44f3-af2c-cc96c45471c1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29483250-mrvrz" Jan 21 11:30:00 crc kubenswrapper[4824]: I0121 11:30:00.283201 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4c02bf47-f856-44f3-af2c-cc96c45471c1-secret-volume\") pod \"collect-profiles-29483250-mrvrz\" (UID: \"4c02bf47-f856-44f3-af2c-cc96c45471c1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29483250-mrvrz" Jan 21 11:30:00 crc kubenswrapper[4824]: I0121 11:30:00.384677 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h6hhk\" (UniqueName: \"kubernetes.io/projected/4c02bf47-f856-44f3-af2c-cc96c45471c1-kube-api-access-h6hhk\") pod \"collect-profiles-29483250-mrvrz\" (UID: \"4c02bf47-f856-44f3-af2c-cc96c45471c1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29483250-mrvrz" Jan 21 11:30:00 crc kubenswrapper[4824]: I0121 11:30:00.385382 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4c02bf47-f856-44f3-af2c-cc96c45471c1-secret-volume\") pod \"collect-profiles-29483250-mrvrz\" (UID: \"4c02bf47-f856-44f3-af2c-cc96c45471c1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29483250-mrvrz" Jan 21 11:30:00 crc kubenswrapper[4824]: I0121 11:30:00.385528 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4c02bf47-f856-44f3-af2c-cc96c45471c1-config-volume\") pod \"collect-profiles-29483250-mrvrz\" (UID: \"4c02bf47-f856-44f3-af2c-cc96c45471c1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29483250-mrvrz" Jan 21 11:30:00 crc kubenswrapper[4824]: I0121 11:30:00.386290 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4c02bf47-f856-44f3-af2c-cc96c45471c1-config-volume\") pod \"collect-profiles-29483250-mrvrz\" (UID: \"4c02bf47-f856-44f3-af2c-cc96c45471c1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29483250-mrvrz" Jan 21 11:30:00 crc kubenswrapper[4824]: I0121 11:30:00.390373 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4c02bf47-f856-44f3-af2c-cc96c45471c1-secret-volume\") pod \"collect-profiles-29483250-mrvrz\" (UID: \"4c02bf47-f856-44f3-af2c-cc96c45471c1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29483250-mrvrz" Jan 21 11:30:00 crc kubenswrapper[4824]: I0121 11:30:00.398786 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h6hhk\" (UniqueName: \"kubernetes.io/projected/4c02bf47-f856-44f3-af2c-cc96c45471c1-kube-api-access-h6hhk\") pod \"collect-profiles-29483250-mrvrz\" (UID: \"4c02bf47-f856-44f3-af2c-cc96c45471c1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29483250-mrvrz" Jan 21 11:30:00 crc kubenswrapper[4824]: I0121 11:30:00.473908 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29483250-mrvrz" Jan 21 11:30:00 crc kubenswrapper[4824]: I0121 11:30:00.854975 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29483250-mrvrz"] Jan 21 11:30:01 crc kubenswrapper[4824]: I0121 11:30:01.196294 4824 generic.go:334] "Generic (PLEG): container finished" podID="4c02bf47-f856-44f3-af2c-cc96c45471c1" containerID="dd264d8d4b02d26df3192da84e03b87f54958338166fb00a9d9562f0fbc98220" exitCode=0 Jan 21 11:30:01 crc kubenswrapper[4824]: I0121 11:30:01.196511 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29483250-mrvrz" event={"ID":"4c02bf47-f856-44f3-af2c-cc96c45471c1","Type":"ContainerDied","Data":"dd264d8d4b02d26df3192da84e03b87f54958338166fb00a9d9562f0fbc98220"} Jan 21 11:30:01 crc kubenswrapper[4824]: I0121 11:30:01.196535 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29483250-mrvrz" event={"ID":"4c02bf47-f856-44f3-af2c-cc96c45471c1","Type":"ContainerStarted","Data":"5404443892f777ddba4e33e0fe95bb8c01a0c937d229122cc30c76b5b89517d6"} Jan 21 11:30:01 crc kubenswrapper[4824]: I0121 11:30:01.373268 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="84e74d7b-18a0-4a3f-8680-6246ac538a6e" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.94:5671: connect: connection refused" Jan 21 11:30:01 crc kubenswrapper[4824]: I0121 11:30:01.631637 4824 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="156fec50-b486-4e84-a7bf-b40491a863c7" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.95:5671: connect: connection refused" Jan 21 11:30:02 crc kubenswrapper[4824]: I0121 11:30:02.454599 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29483250-mrvrz" Jan 21 11:30:02 crc kubenswrapper[4824]: I0121 11:30:02.520845 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4c02bf47-f856-44f3-af2c-cc96c45471c1-config-volume\") pod \"4c02bf47-f856-44f3-af2c-cc96c45471c1\" (UID: \"4c02bf47-f856-44f3-af2c-cc96c45471c1\") " Jan 21 11:30:02 crc kubenswrapper[4824]: I0121 11:30:02.520969 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h6hhk\" (UniqueName: \"kubernetes.io/projected/4c02bf47-f856-44f3-af2c-cc96c45471c1-kube-api-access-h6hhk\") pod \"4c02bf47-f856-44f3-af2c-cc96c45471c1\" (UID: \"4c02bf47-f856-44f3-af2c-cc96c45471c1\") " Jan 21 11:30:02 crc kubenswrapper[4824]: I0121 11:30:02.521015 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4c02bf47-f856-44f3-af2c-cc96c45471c1-secret-volume\") pod \"4c02bf47-f856-44f3-af2c-cc96c45471c1\" (UID: \"4c02bf47-f856-44f3-af2c-cc96c45471c1\") " Jan 21 11:30:02 crc kubenswrapper[4824]: I0121 11:30:02.521507 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4c02bf47-f856-44f3-af2c-cc96c45471c1-config-volume" (OuterVolumeSpecName: "config-volume") pod "4c02bf47-f856-44f3-af2c-cc96c45471c1" (UID: "4c02bf47-f856-44f3-af2c-cc96c45471c1"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:30:02 crc kubenswrapper[4824]: I0121 11:30:02.525468 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c02bf47-f856-44f3-af2c-cc96c45471c1-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "4c02bf47-f856-44f3-af2c-cc96c45471c1" (UID: "4c02bf47-f856-44f3-af2c-cc96c45471c1"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:30:02 crc kubenswrapper[4824]: I0121 11:30:02.525517 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4c02bf47-f856-44f3-af2c-cc96c45471c1-kube-api-access-h6hhk" (OuterVolumeSpecName: "kube-api-access-h6hhk") pod "4c02bf47-f856-44f3-af2c-cc96c45471c1" (UID: "4c02bf47-f856-44f3-af2c-cc96c45471c1"). InnerVolumeSpecName "kube-api-access-h6hhk". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:30:02 crc kubenswrapper[4824]: I0121 11:30:02.623177 4824 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4c02bf47-f856-44f3-af2c-cc96c45471c1-config-volume\") on node \"crc\" DevicePath \"\"" Jan 21 11:30:02 crc kubenswrapper[4824]: I0121 11:30:02.623207 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h6hhk\" (UniqueName: \"kubernetes.io/projected/4c02bf47-f856-44f3-af2c-cc96c45471c1-kube-api-access-h6hhk\") on node \"crc\" DevicePath \"\"" Jan 21 11:30:02 crc kubenswrapper[4824]: I0121 11:30:02.623220 4824 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4c02bf47-f856-44f3-af2c-cc96c45471c1-secret-volume\") on node \"crc\" DevicePath \"\"" Jan 21 11:30:03 crc kubenswrapper[4824]: I0121 11:30:03.209866 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29483250-mrvrz" event={"ID":"4c02bf47-f856-44f3-af2c-cc96c45471c1","Type":"ContainerDied","Data":"5404443892f777ddba4e33e0fe95bb8c01a0c937d229122cc30c76b5b89517d6"} Jan 21 11:30:03 crc kubenswrapper[4824]: I0121 11:30:03.209900 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29483250-mrvrz" Jan 21 11:30:03 crc kubenswrapper[4824]: I0121 11:30:03.209907 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5404443892f777ddba4e33e0fe95bb8c01a0c937d229122cc30c76b5b89517d6" Jan 21 11:30:05 crc kubenswrapper[4824]: I0121 11:30:05.713433 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Jan 21 11:30:05 crc kubenswrapper[4824]: I0121 11:30:05.876529 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/84e74d7b-18a0-4a3f-8680-6246ac538a6e-config-data\") pod \"84e74d7b-18a0-4a3f-8680-6246ac538a6e\" (UID: \"84e74d7b-18a0-4a3f-8680-6246ac538a6e\") " Jan 21 11:30:05 crc kubenswrapper[4824]: I0121 11:30:05.876569 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/84e74d7b-18a0-4a3f-8680-6246ac538a6e-pod-info\") pod \"84e74d7b-18a0-4a3f-8680-6246ac538a6e\" (UID: \"84e74d7b-18a0-4a3f-8680-6246ac538a6e\") " Jan 21 11:30:05 crc kubenswrapper[4824]: I0121 11:30:05.876648 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/84e74d7b-18a0-4a3f-8680-6246ac538a6e-rabbitmq-tls\") pod \"84e74d7b-18a0-4a3f-8680-6246ac538a6e\" (UID: \"84e74d7b-18a0-4a3f-8680-6246ac538a6e\") " Jan 21 11:30:05 crc kubenswrapper[4824]: I0121 11:30:05.876715 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/84e74d7b-18a0-4a3f-8680-6246ac538a6e-rabbitmq-confd\") pod \"84e74d7b-18a0-4a3f-8680-6246ac538a6e\" (UID: \"84e74d7b-18a0-4a3f-8680-6246ac538a6e\") " Jan 21 11:30:05 crc kubenswrapper[4824]: I0121 11:30:05.876744 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/84e74d7b-18a0-4a3f-8680-6246ac538a6e-rabbitmq-erlang-cookie\") pod \"84e74d7b-18a0-4a3f-8680-6246ac538a6e\" (UID: \"84e74d7b-18a0-4a3f-8680-6246ac538a6e\") " Jan 21 11:30:05 crc kubenswrapper[4824]: I0121 11:30:05.876806 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/84e74d7b-18a0-4a3f-8680-6246ac538a6e-plugins-conf\") pod \"84e74d7b-18a0-4a3f-8680-6246ac538a6e\" (UID: \"84e74d7b-18a0-4a3f-8680-6246ac538a6e\") " Jan 21 11:30:05 crc kubenswrapper[4824]: I0121 11:30:05.876841 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/84e74d7b-18a0-4a3f-8680-6246ac538a6e-rabbitmq-plugins\") pod \"84e74d7b-18a0-4a3f-8680-6246ac538a6e\" (UID: \"84e74d7b-18a0-4a3f-8680-6246ac538a6e\") " Jan 21 11:30:05 crc kubenswrapper[4824]: I0121 11:30:05.876862 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"84e74d7b-18a0-4a3f-8680-6246ac538a6e\" (UID: \"84e74d7b-18a0-4a3f-8680-6246ac538a6e\") " Jan 21 11:30:05 crc kubenswrapper[4824]: I0121 11:30:05.876909 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/84e74d7b-18a0-4a3f-8680-6246ac538a6e-server-conf\") pod \"84e74d7b-18a0-4a3f-8680-6246ac538a6e\" (UID: \"84e74d7b-18a0-4a3f-8680-6246ac538a6e\") " Jan 21 11:30:05 crc kubenswrapper[4824]: I0121 11:30:05.876935 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/84e74d7b-18a0-4a3f-8680-6246ac538a6e-erlang-cookie-secret\") pod \"84e74d7b-18a0-4a3f-8680-6246ac538a6e\" (UID: \"84e74d7b-18a0-4a3f-8680-6246ac538a6e\") " Jan 21 11:30:05 crc kubenswrapper[4824]: I0121 11:30:05.876949 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkwkl\" (UniqueName: \"kubernetes.io/projected/84e74d7b-18a0-4a3f-8680-6246ac538a6e-kube-api-access-zkwkl\") pod \"84e74d7b-18a0-4a3f-8680-6246ac538a6e\" (UID: \"84e74d7b-18a0-4a3f-8680-6246ac538a6e\") " Jan 21 11:30:05 crc kubenswrapper[4824]: I0121 11:30:05.877371 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/84e74d7b-18a0-4a3f-8680-6246ac538a6e-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "84e74d7b-18a0-4a3f-8680-6246ac538a6e" (UID: "84e74d7b-18a0-4a3f-8680-6246ac538a6e"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:30:05 crc kubenswrapper[4824]: I0121 11:30:05.877606 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/84e74d7b-18a0-4a3f-8680-6246ac538a6e-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "84e74d7b-18a0-4a3f-8680-6246ac538a6e" (UID: "84e74d7b-18a0-4a3f-8680-6246ac538a6e"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:30:05 crc kubenswrapper[4824]: I0121 11:30:05.877738 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/84e74d7b-18a0-4a3f-8680-6246ac538a6e-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "84e74d7b-18a0-4a3f-8680-6246ac538a6e" (UID: "84e74d7b-18a0-4a3f-8680-6246ac538a6e"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:30:05 crc kubenswrapper[4824]: I0121 11:30:05.878139 4824 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/84e74d7b-18a0-4a3f-8680-6246ac538a6e-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Jan 21 11:30:05 crc kubenswrapper[4824]: I0121 11:30:05.878163 4824 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/84e74d7b-18a0-4a3f-8680-6246ac538a6e-plugins-conf\") on node \"crc\" DevicePath \"\"" Jan 21 11:30:05 crc kubenswrapper[4824]: I0121 11:30:05.878174 4824 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/84e74d7b-18a0-4a3f-8680-6246ac538a6e-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Jan 21 11:30:05 crc kubenswrapper[4824]: I0121 11:30:05.881065 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/84e74d7b-18a0-4a3f-8680-6246ac538a6e-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "84e74d7b-18a0-4a3f-8680-6246ac538a6e" (UID: "84e74d7b-18a0-4a3f-8680-6246ac538a6e"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:30:05 crc kubenswrapper[4824]: I0121 11:30:05.881563 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/84e74d7b-18a0-4a3f-8680-6246ac538a6e-pod-info" (OuterVolumeSpecName: "pod-info") pod "84e74d7b-18a0-4a3f-8680-6246ac538a6e" (UID: "84e74d7b-18a0-4a3f-8680-6246ac538a6e"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Jan 21 11:30:05 crc kubenswrapper[4824]: I0121 11:30:05.881572 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/84e74d7b-18a0-4a3f-8680-6246ac538a6e-kube-api-access-zkwkl" (OuterVolumeSpecName: "kube-api-access-zkwkl") pod "84e74d7b-18a0-4a3f-8680-6246ac538a6e" (UID: "84e74d7b-18a0-4a3f-8680-6246ac538a6e"). InnerVolumeSpecName "kube-api-access-zkwkl". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:30:05 crc kubenswrapper[4824]: I0121 11:30:05.881634 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage09-crc" (OuterVolumeSpecName: "persistence") pod "84e74d7b-18a0-4a3f-8680-6246ac538a6e" (UID: "84e74d7b-18a0-4a3f-8680-6246ac538a6e"). InnerVolumeSpecName "local-storage09-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Jan 21 11:30:05 crc kubenswrapper[4824]: I0121 11:30:05.884074 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/84e74d7b-18a0-4a3f-8680-6246ac538a6e-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "84e74d7b-18a0-4a3f-8680-6246ac538a6e" (UID: "84e74d7b-18a0-4a3f-8680-6246ac538a6e"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:30:05 crc kubenswrapper[4824]: I0121 11:30:05.914340 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/84e74d7b-18a0-4a3f-8680-6246ac538a6e-config-data" (OuterVolumeSpecName: "config-data") pod "84e74d7b-18a0-4a3f-8680-6246ac538a6e" (UID: "84e74d7b-18a0-4a3f-8680-6246ac538a6e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:30:05 crc kubenswrapper[4824]: I0121 11:30:05.928004 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/84e74d7b-18a0-4a3f-8680-6246ac538a6e-server-conf" (OuterVolumeSpecName: "server-conf") pod "84e74d7b-18a0-4a3f-8680-6246ac538a6e" (UID: "84e74d7b-18a0-4a3f-8680-6246ac538a6e"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:30:05 crc kubenswrapper[4824]: I0121 11:30:05.972927 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/84e74d7b-18a0-4a3f-8680-6246ac538a6e-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "84e74d7b-18a0-4a3f-8680-6246ac538a6e" (UID: "84e74d7b-18a0-4a3f-8680-6246ac538a6e"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:30:05 crc kubenswrapper[4824]: I0121 11:30:05.979802 4824 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" " Jan 21 11:30:05 crc kubenswrapper[4824]: I0121 11:30:05.979844 4824 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/84e74d7b-18a0-4a3f-8680-6246ac538a6e-server-conf\") on node \"crc\" DevicePath \"\"" Jan 21 11:30:05 crc kubenswrapper[4824]: I0121 11:30:05.979855 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkwkl\" (UniqueName: \"kubernetes.io/projected/84e74d7b-18a0-4a3f-8680-6246ac538a6e-kube-api-access-zkwkl\") on node \"crc\" DevicePath \"\"" Jan 21 11:30:05 crc kubenswrapper[4824]: I0121 11:30:05.979864 4824 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/84e74d7b-18a0-4a3f-8680-6246ac538a6e-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Jan 21 11:30:05 crc kubenswrapper[4824]: I0121 11:30:05.979872 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/84e74d7b-18a0-4a3f-8680-6246ac538a6e-config-data\") on node \"crc\" DevicePath \"\"" Jan 21 11:30:05 crc kubenswrapper[4824]: I0121 11:30:05.979878 4824 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/84e74d7b-18a0-4a3f-8680-6246ac538a6e-pod-info\") on node \"crc\" DevicePath \"\"" Jan 21 11:30:05 crc kubenswrapper[4824]: I0121 11:30:05.979885 4824 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/84e74d7b-18a0-4a3f-8680-6246ac538a6e-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Jan 21 11:30:05 crc kubenswrapper[4824]: I0121 11:30:05.979893 4824 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/84e74d7b-18a0-4a3f-8680-6246ac538a6e-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:05.999774 4824 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage09-crc" (UniqueName: "kubernetes.io/local-volume/local-storage09-crc") on node "crc" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.082520 4824 reconciler_common.go:293] "Volume detached for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" DevicePath \"\"" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.178906 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.234201 4824 generic.go:334] "Generic (PLEG): container finished" podID="156fec50-b486-4e84-a7bf-b40491a863c7" containerID="902d1a3e9a8d54f58be516ba97809a95404a4ebe849ca1cb1fe5805c23862e32" exitCode=0 Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.234246 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.234272 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"156fec50-b486-4e84-a7bf-b40491a863c7","Type":"ContainerDied","Data":"902d1a3e9a8d54f58be516ba97809a95404a4ebe849ca1cb1fe5805c23862e32"} Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.234298 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"156fec50-b486-4e84-a7bf-b40491a863c7","Type":"ContainerDied","Data":"37d36a29724a133bc6f138979451aa8f9c0eeec9cad658937121a853dcbd295e"} Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.234315 4824 scope.go:117] "RemoveContainer" containerID="902d1a3e9a8d54f58be516ba97809a95404a4ebe849ca1cb1fe5805c23862e32" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.236427 4824 generic.go:334] "Generic (PLEG): container finished" podID="84e74d7b-18a0-4a3f-8680-6246ac538a6e" containerID="407422f3cadd164c7680cc8b0bf1b21f0ebb5d9d1b6255a64420fbd1143666e7" exitCode=0 Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.236458 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"84e74d7b-18a0-4a3f-8680-6246ac538a6e","Type":"ContainerDied","Data":"407422f3cadd164c7680cc8b0bf1b21f0ebb5d9d1b6255a64420fbd1143666e7"} Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.236477 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"84e74d7b-18a0-4a3f-8680-6246ac538a6e","Type":"ContainerDied","Data":"8ee895c662c7814313d66afa49854eb939a2dc25462867e35d0362794391ee25"} Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.236524 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.251928 4824 scope.go:117] "RemoveContainer" containerID="57e1badc440cdd38a60828271554d42b57947d40b89d5290cb4d30f5fcb54487" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.265868 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.277023 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.282107 4824 scope.go:117] "RemoveContainer" containerID="902d1a3e9a8d54f58be516ba97809a95404a4ebe849ca1cb1fe5805c23862e32" Jan 21 11:30:06 crc kubenswrapper[4824]: E0121 11:30:06.284928 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"902d1a3e9a8d54f58be516ba97809a95404a4ebe849ca1cb1fe5805c23862e32\": container with ID starting with 902d1a3e9a8d54f58be516ba97809a95404a4ebe849ca1cb1fe5805c23862e32 not found: ID does not exist" containerID="902d1a3e9a8d54f58be516ba97809a95404a4ebe849ca1cb1fe5805c23862e32" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.284973 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j9nc5\" (UniqueName: \"kubernetes.io/projected/156fec50-b486-4e84-a7bf-b40491a863c7-kube-api-access-j9nc5\") pod \"156fec50-b486-4e84-a7bf-b40491a863c7\" (UID: \"156fec50-b486-4e84-a7bf-b40491a863c7\") " Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.284984 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"902d1a3e9a8d54f58be516ba97809a95404a4ebe849ca1cb1fe5805c23862e32"} err="failed to get container status \"902d1a3e9a8d54f58be516ba97809a95404a4ebe849ca1cb1fe5805c23862e32\": rpc error: code = NotFound desc = could not find container \"902d1a3e9a8d54f58be516ba97809a95404a4ebe849ca1cb1fe5805c23862e32\": container with ID starting with 902d1a3e9a8d54f58be516ba97809a95404a4ebe849ca1cb1fe5805c23862e32 not found: ID does not exist" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.285010 4824 scope.go:117] "RemoveContainer" containerID="57e1badc440cdd38a60828271554d42b57947d40b89d5290cb4d30f5fcb54487" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.285021 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/156fec50-b486-4e84-a7bf-b40491a863c7-rabbitmq-plugins\") pod \"156fec50-b486-4e84-a7bf-b40491a863c7\" (UID: \"156fec50-b486-4e84-a7bf-b40491a863c7\") " Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.285082 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/156fec50-b486-4e84-a7bf-b40491a863c7-pod-info\") pod \"156fec50-b486-4e84-a7bf-b40491a863c7\" (UID: \"156fec50-b486-4e84-a7bf-b40491a863c7\") " Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.285167 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/156fec50-b486-4e84-a7bf-b40491a863c7-rabbitmq-erlang-cookie\") pod \"156fec50-b486-4e84-a7bf-b40491a863c7\" (UID: \"156fec50-b486-4e84-a7bf-b40491a863c7\") " Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.285210 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"156fec50-b486-4e84-a7bf-b40491a863c7\" (UID: \"156fec50-b486-4e84-a7bf-b40491a863c7\") " Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.285227 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/156fec50-b486-4e84-a7bf-b40491a863c7-rabbitmq-tls\") pod \"156fec50-b486-4e84-a7bf-b40491a863c7\" (UID: \"156fec50-b486-4e84-a7bf-b40491a863c7\") " Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.285267 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/156fec50-b486-4e84-a7bf-b40491a863c7-plugins-conf\") pod \"156fec50-b486-4e84-a7bf-b40491a863c7\" (UID: \"156fec50-b486-4e84-a7bf-b40491a863c7\") " Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.285292 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/156fec50-b486-4e84-a7bf-b40491a863c7-server-conf\") pod \"156fec50-b486-4e84-a7bf-b40491a863c7\" (UID: \"156fec50-b486-4e84-a7bf-b40491a863c7\") " Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.285314 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/156fec50-b486-4e84-a7bf-b40491a863c7-rabbitmq-confd\") pod \"156fec50-b486-4e84-a7bf-b40491a863c7\" (UID: \"156fec50-b486-4e84-a7bf-b40491a863c7\") " Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.285354 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/156fec50-b486-4e84-a7bf-b40491a863c7-erlang-cookie-secret\") pod \"156fec50-b486-4e84-a7bf-b40491a863c7\" (UID: \"156fec50-b486-4e84-a7bf-b40491a863c7\") " Jan 21 11:30:06 crc kubenswrapper[4824]: E0121 11:30:06.285360 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"57e1badc440cdd38a60828271554d42b57947d40b89d5290cb4d30f5fcb54487\": container with ID starting with 57e1badc440cdd38a60828271554d42b57947d40b89d5290cb4d30f5fcb54487 not found: ID does not exist" containerID="57e1badc440cdd38a60828271554d42b57947d40b89d5290cb4d30f5fcb54487" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.285393 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"57e1badc440cdd38a60828271554d42b57947d40b89d5290cb4d30f5fcb54487"} err="failed to get container status \"57e1badc440cdd38a60828271554d42b57947d40b89d5290cb4d30f5fcb54487\": rpc error: code = NotFound desc = could not find container \"57e1badc440cdd38a60828271554d42b57947d40b89d5290cb4d30f5fcb54487\": container with ID starting with 57e1badc440cdd38a60828271554d42b57947d40b89d5290cb4d30f5fcb54487 not found: ID does not exist" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.285416 4824 scope.go:117] "RemoveContainer" containerID="407422f3cadd164c7680cc8b0bf1b21f0ebb5d9d1b6255a64420fbd1143666e7" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.285377 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/156fec50-b486-4e84-a7bf-b40491a863c7-config-data\") pod \"156fec50-b486-4e84-a7bf-b40491a863c7\" (UID: \"156fec50-b486-4e84-a7bf-b40491a863c7\") " Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.285606 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/156fec50-b486-4e84-a7bf-b40491a863c7-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "156fec50-b486-4e84-a7bf-b40491a863c7" (UID: "156fec50-b486-4e84-a7bf-b40491a863c7"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.285891 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/156fec50-b486-4e84-a7bf-b40491a863c7-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "156fec50-b486-4e84-a7bf-b40491a863c7" (UID: "156fec50-b486-4e84-a7bf-b40491a863c7"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.286150 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/156fec50-b486-4e84-a7bf-b40491a863c7-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "156fec50-b486-4e84-a7bf-b40491a863c7" (UID: "156fec50-b486-4e84-a7bf-b40491a863c7"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.286975 4824 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/156fec50-b486-4e84-a7bf-b40491a863c7-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.286993 4824 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/156fec50-b486-4e84-a7bf-b40491a863c7-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.287002 4824 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/156fec50-b486-4e84-a7bf-b40491a863c7-plugins-conf\") on node \"crc\" DevicePath \"\"" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.293875 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage04-crc" (OuterVolumeSpecName: "persistence") pod "156fec50-b486-4e84-a7bf-b40491a863c7" (UID: "156fec50-b486-4e84-a7bf-b40491a863c7"). InnerVolumeSpecName "local-storage04-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.294396 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/156fec50-b486-4e84-a7bf-b40491a863c7-kube-api-access-j9nc5" (OuterVolumeSpecName: "kube-api-access-j9nc5") pod "156fec50-b486-4e84-a7bf-b40491a863c7" (UID: "156fec50-b486-4e84-a7bf-b40491a863c7"). InnerVolumeSpecName "kube-api-access-j9nc5". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.294440 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/156fec50-b486-4e84-a7bf-b40491a863c7-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "156fec50-b486-4e84-a7bf-b40491a863c7" (UID: "156fec50-b486-4e84-a7bf-b40491a863c7"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.294645 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/156fec50-b486-4e84-a7bf-b40491a863c7-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "156fec50-b486-4e84-a7bf-b40491a863c7" (UID: "156fec50-b486-4e84-a7bf-b40491a863c7"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.294760 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Jan 21 11:30:06 crc kubenswrapper[4824]: E0121 11:30:06.295107 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="156fec50-b486-4e84-a7bf-b40491a863c7" containerName="setup-container" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.295124 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="156fec50-b486-4e84-a7bf-b40491a863c7" containerName="setup-container" Jan 21 11:30:06 crc kubenswrapper[4824]: E0121 11:30:06.295140 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="156fec50-b486-4e84-a7bf-b40491a863c7" containerName="rabbitmq" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.295147 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="156fec50-b486-4e84-a7bf-b40491a863c7" containerName="rabbitmq" Jan 21 11:30:06 crc kubenswrapper[4824]: E0121 11:30:06.295159 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c02bf47-f856-44f3-af2c-cc96c45471c1" containerName="collect-profiles" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.295165 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c02bf47-f856-44f3-af2c-cc96c45471c1" containerName="collect-profiles" Jan 21 11:30:06 crc kubenswrapper[4824]: E0121 11:30:06.295176 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="84e74d7b-18a0-4a3f-8680-6246ac538a6e" containerName="rabbitmq" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.295181 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="84e74d7b-18a0-4a3f-8680-6246ac538a6e" containerName="rabbitmq" Jan 21 11:30:06 crc kubenswrapper[4824]: E0121 11:30:06.295188 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="84e74d7b-18a0-4a3f-8680-6246ac538a6e" containerName="setup-container" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.295193 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="84e74d7b-18a0-4a3f-8680-6246ac538a6e" containerName="setup-container" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.295369 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="4c02bf47-f856-44f3-af2c-cc96c45471c1" containerName="collect-profiles" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.295384 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="84e74d7b-18a0-4a3f-8680-6246ac538a6e" containerName="rabbitmq" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.295398 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="156fec50-b486-4e84-a7bf-b40491a863c7" containerName="rabbitmq" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.296750 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.301405 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.301447 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.301597 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.301612 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.301692 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.301758 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-22jwj" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.301806 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.303174 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/156fec50-b486-4e84-a7bf-b40491a863c7-pod-info" (OuterVolumeSpecName: "pod-info") pod "156fec50-b486-4e84-a7bf-b40491a863c7" (UID: "156fec50-b486-4e84-a7bf-b40491a863c7"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.307657 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.308591 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/156fec50-b486-4e84-a7bf-b40491a863c7-config-data" (OuterVolumeSpecName: "config-data") pod "156fec50-b486-4e84-a7bf-b40491a863c7" (UID: "156fec50-b486-4e84-a7bf-b40491a863c7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.323787 4824 scope.go:117] "RemoveContainer" containerID="12b90320ffa88a78c84d62cf8b65c324fb6c4b884b2714e0a27b33b29cca2e17" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.343511 4824 scope.go:117] "RemoveContainer" containerID="407422f3cadd164c7680cc8b0bf1b21f0ebb5d9d1b6255a64420fbd1143666e7" Jan 21 11:30:06 crc kubenswrapper[4824]: E0121 11:30:06.343968 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"407422f3cadd164c7680cc8b0bf1b21f0ebb5d9d1b6255a64420fbd1143666e7\": container with ID starting with 407422f3cadd164c7680cc8b0bf1b21f0ebb5d9d1b6255a64420fbd1143666e7 not found: ID does not exist" containerID="407422f3cadd164c7680cc8b0bf1b21f0ebb5d9d1b6255a64420fbd1143666e7" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.343996 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"407422f3cadd164c7680cc8b0bf1b21f0ebb5d9d1b6255a64420fbd1143666e7"} err="failed to get container status \"407422f3cadd164c7680cc8b0bf1b21f0ebb5d9d1b6255a64420fbd1143666e7\": rpc error: code = NotFound desc = could not find container \"407422f3cadd164c7680cc8b0bf1b21f0ebb5d9d1b6255a64420fbd1143666e7\": container with ID starting with 407422f3cadd164c7680cc8b0bf1b21f0ebb5d9d1b6255a64420fbd1143666e7 not found: ID does not exist" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.344039 4824 scope.go:117] "RemoveContainer" containerID="12b90320ffa88a78c84d62cf8b65c324fb6c4b884b2714e0a27b33b29cca2e17" Jan 21 11:30:06 crc kubenswrapper[4824]: E0121 11:30:06.344501 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"12b90320ffa88a78c84d62cf8b65c324fb6c4b884b2714e0a27b33b29cca2e17\": container with ID starting with 12b90320ffa88a78c84d62cf8b65c324fb6c4b884b2714e0a27b33b29cca2e17 not found: ID does not exist" containerID="12b90320ffa88a78c84d62cf8b65c324fb6c4b884b2714e0a27b33b29cca2e17" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.344542 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"12b90320ffa88a78c84d62cf8b65c324fb6c4b884b2714e0a27b33b29cca2e17"} err="failed to get container status \"12b90320ffa88a78c84d62cf8b65c324fb6c4b884b2714e0a27b33b29cca2e17\": rpc error: code = NotFound desc = could not find container \"12b90320ffa88a78c84d62cf8b65c324fb6c4b884b2714e0a27b33b29cca2e17\": container with ID starting with 12b90320ffa88a78c84d62cf8b65c324fb6c4b884b2714e0a27b33b29cca2e17 not found: ID does not exist" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.357111 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/156fec50-b486-4e84-a7bf-b40491a863c7-server-conf" (OuterVolumeSpecName: "server-conf") pod "156fec50-b486-4e84-a7bf-b40491a863c7" (UID: "156fec50-b486-4e84-a7bf-b40491a863c7"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.387918 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e4a4d725-1a49-4292-af74-6ab09ab15cdb-config-data\") pod \"rabbitmq-server-0\" (UID: \"e4a4d725-1a49-4292-af74-6ab09ab15cdb\") " pod="openstack/rabbitmq-server-0" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.387981 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/e4a4d725-1a49-4292-af74-6ab09ab15cdb-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"e4a4d725-1a49-4292-af74-6ab09ab15cdb\") " pod="openstack/rabbitmq-server-0" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.388010 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/e4a4d725-1a49-4292-af74-6ab09ab15cdb-pod-info\") pod \"rabbitmq-server-0\" (UID: \"e4a4d725-1a49-4292-af74-6ab09ab15cdb\") " pod="openstack/rabbitmq-server-0" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.388079 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/e4a4d725-1a49-4292-af74-6ab09ab15cdb-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"e4a4d725-1a49-4292-af74-6ab09ab15cdb\") " pod="openstack/rabbitmq-server-0" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.388120 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/e4a4d725-1a49-4292-af74-6ab09ab15cdb-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"e4a4d725-1a49-4292-af74-6ab09ab15cdb\") " pod="openstack/rabbitmq-server-0" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.388142 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/e4a4d725-1a49-4292-af74-6ab09ab15cdb-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"e4a4d725-1a49-4292-af74-6ab09ab15cdb\") " pod="openstack/rabbitmq-server-0" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.388203 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/e4a4d725-1a49-4292-af74-6ab09ab15cdb-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"e4a4d725-1a49-4292-af74-6ab09ab15cdb\") " pod="openstack/rabbitmq-server-0" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.388220 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/e4a4d725-1a49-4292-af74-6ab09ab15cdb-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"e4a4d725-1a49-4292-af74-6ab09ab15cdb\") " pod="openstack/rabbitmq-server-0" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.388300 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/e4a4d725-1a49-4292-af74-6ab09ab15cdb-server-conf\") pod \"rabbitmq-server-0\" (UID: \"e4a4d725-1a49-4292-af74-6ab09ab15cdb\") " pod="openstack/rabbitmq-server-0" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.388316 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-server-0\" (UID: \"e4a4d725-1a49-4292-af74-6ab09ab15cdb\") " pod="openstack/rabbitmq-server-0" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.388440 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f8m7b\" (UniqueName: \"kubernetes.io/projected/e4a4d725-1a49-4292-af74-6ab09ab15cdb-kube-api-access-f8m7b\") pod \"rabbitmq-server-0\" (UID: \"e4a4d725-1a49-4292-af74-6ab09ab15cdb\") " pod="openstack/rabbitmq-server-0" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.388599 4824 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/156fec50-b486-4e84-a7bf-b40491a863c7-pod-info\") on node \"crc\" DevicePath \"\"" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.388619 4824 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" " Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.388629 4824 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/156fec50-b486-4e84-a7bf-b40491a863c7-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.388639 4824 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/156fec50-b486-4e84-a7bf-b40491a863c7-server-conf\") on node \"crc\" DevicePath \"\"" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.388647 4824 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/156fec50-b486-4e84-a7bf-b40491a863c7-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.388674 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/156fec50-b486-4e84-a7bf-b40491a863c7-config-data\") on node \"crc\" DevicePath \"\"" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.388683 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j9nc5\" (UniqueName: \"kubernetes.io/projected/156fec50-b486-4e84-a7bf-b40491a863c7-kube-api-access-j9nc5\") on node \"crc\" DevicePath \"\"" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.392606 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/156fec50-b486-4e84-a7bf-b40491a863c7-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "156fec50-b486-4e84-a7bf-b40491a863c7" (UID: "156fec50-b486-4e84-a7bf-b40491a863c7"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.404820 4824 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage04-crc" (UniqueName: "kubernetes.io/local-volume/local-storage04-crc") on node "crc" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.490207 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e4a4d725-1a49-4292-af74-6ab09ab15cdb-config-data\") pod \"rabbitmq-server-0\" (UID: \"e4a4d725-1a49-4292-af74-6ab09ab15cdb\") " pod="openstack/rabbitmq-server-0" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.490360 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/e4a4d725-1a49-4292-af74-6ab09ab15cdb-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"e4a4d725-1a49-4292-af74-6ab09ab15cdb\") " pod="openstack/rabbitmq-server-0" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.490451 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/e4a4d725-1a49-4292-af74-6ab09ab15cdb-pod-info\") pod \"rabbitmq-server-0\" (UID: \"e4a4d725-1a49-4292-af74-6ab09ab15cdb\") " pod="openstack/rabbitmq-server-0" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.490588 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/e4a4d725-1a49-4292-af74-6ab09ab15cdb-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"e4a4d725-1a49-4292-af74-6ab09ab15cdb\") " pod="openstack/rabbitmq-server-0" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.490695 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/e4a4d725-1a49-4292-af74-6ab09ab15cdb-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"e4a4d725-1a49-4292-af74-6ab09ab15cdb\") " pod="openstack/rabbitmq-server-0" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.491241 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/e4a4d725-1a49-4292-af74-6ab09ab15cdb-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"e4a4d725-1a49-4292-af74-6ab09ab15cdb\") " pod="openstack/rabbitmq-server-0" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.491871 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/e4a4d725-1a49-4292-af74-6ab09ab15cdb-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"e4a4d725-1a49-4292-af74-6ab09ab15cdb\") " pod="openstack/rabbitmq-server-0" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.490991 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e4a4d725-1a49-4292-af74-6ab09ab15cdb-config-data\") pod \"rabbitmq-server-0\" (UID: \"e4a4d725-1a49-4292-af74-6ab09ab15cdb\") " pod="openstack/rabbitmq-server-0" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.491067 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/e4a4d725-1a49-4292-af74-6ab09ab15cdb-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"e4a4d725-1a49-4292-af74-6ab09ab15cdb\") " pod="openstack/rabbitmq-server-0" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.491880 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/e4a4d725-1a49-4292-af74-6ab09ab15cdb-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"e4a4d725-1a49-4292-af74-6ab09ab15cdb\") " pod="openstack/rabbitmq-server-0" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.492224 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/e4a4d725-1a49-4292-af74-6ab09ab15cdb-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"e4a4d725-1a49-4292-af74-6ab09ab15cdb\") " pod="openstack/rabbitmq-server-0" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.492360 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/e4a4d725-1a49-4292-af74-6ab09ab15cdb-server-conf\") pod \"rabbitmq-server-0\" (UID: \"e4a4d725-1a49-4292-af74-6ab09ab15cdb\") " pod="openstack/rabbitmq-server-0" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.492440 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-server-0\" (UID: \"e4a4d725-1a49-4292-af74-6ab09ab15cdb\") " pod="openstack/rabbitmq-server-0" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.492563 4824 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-server-0\" (UID: \"e4a4d725-1a49-4292-af74-6ab09ab15cdb\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/rabbitmq-server-0" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.492574 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/e4a4d725-1a49-4292-af74-6ab09ab15cdb-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"e4a4d725-1a49-4292-af74-6ab09ab15cdb\") " pod="openstack/rabbitmq-server-0" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.492849 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f8m7b\" (UniqueName: \"kubernetes.io/projected/e4a4d725-1a49-4292-af74-6ab09ab15cdb-kube-api-access-f8m7b\") pod \"rabbitmq-server-0\" (UID: \"e4a4d725-1a49-4292-af74-6ab09ab15cdb\") " pod="openstack/rabbitmq-server-0" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.493303 4824 reconciler_common.go:293] "Volume detached for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" DevicePath \"\"" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.493355 4824 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/156fec50-b486-4e84-a7bf-b40491a863c7-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.493362 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/e4a4d725-1a49-4292-af74-6ab09ab15cdb-server-conf\") pod \"rabbitmq-server-0\" (UID: \"e4a4d725-1a49-4292-af74-6ab09ab15cdb\") " pod="openstack/rabbitmq-server-0" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.493510 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/e4a4d725-1a49-4292-af74-6ab09ab15cdb-pod-info\") pod \"rabbitmq-server-0\" (UID: \"e4a4d725-1a49-4292-af74-6ab09ab15cdb\") " pod="openstack/rabbitmq-server-0" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.493904 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/e4a4d725-1a49-4292-af74-6ab09ab15cdb-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"e4a4d725-1a49-4292-af74-6ab09ab15cdb\") " pod="openstack/rabbitmq-server-0" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.494040 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/e4a4d725-1a49-4292-af74-6ab09ab15cdb-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"e4a4d725-1a49-4292-af74-6ab09ab15cdb\") " pod="openstack/rabbitmq-server-0" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.494585 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/e4a4d725-1a49-4292-af74-6ab09ab15cdb-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"e4a4d725-1a49-4292-af74-6ab09ab15cdb\") " pod="openstack/rabbitmq-server-0" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.507925 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f8m7b\" (UniqueName: \"kubernetes.io/projected/e4a4d725-1a49-4292-af74-6ab09ab15cdb-kube-api-access-f8m7b\") pod \"rabbitmq-server-0\" (UID: \"e4a4d725-1a49-4292-af74-6ab09ab15cdb\") " pod="openstack/rabbitmq-server-0" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.519175 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-server-0\" (UID: \"e4a4d725-1a49-4292-af74-6ab09ab15cdb\") " pod="openstack/rabbitmq-server-0" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.615442 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.619632 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.626142 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.639520 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.641482 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.644479 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.644743 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.644970 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.645201 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-zjdwv" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.645429 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.645597 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.645842 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.680811 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.799098 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/5d8e256a-aa1b-4ccb-8e8c-c86be1e31762-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"5d8e256a-aa1b-4ccb-8e8c-c86be1e31762\") " pod="openstack/rabbitmq-cell1-server-0" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.799293 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7vbvn\" (UniqueName: \"kubernetes.io/projected/5d8e256a-aa1b-4ccb-8e8c-c86be1e31762-kube-api-access-7vbvn\") pod \"rabbitmq-cell1-server-0\" (UID: \"5d8e256a-aa1b-4ccb-8e8c-c86be1e31762\") " pod="openstack/rabbitmq-cell1-server-0" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.799372 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/5d8e256a-aa1b-4ccb-8e8c-c86be1e31762-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"5d8e256a-aa1b-4ccb-8e8c-c86be1e31762\") " pod="openstack/rabbitmq-cell1-server-0" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.799398 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"5d8e256a-aa1b-4ccb-8e8c-c86be1e31762\") " pod="openstack/rabbitmq-cell1-server-0" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.799450 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/5d8e256a-aa1b-4ccb-8e8c-c86be1e31762-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"5d8e256a-aa1b-4ccb-8e8c-c86be1e31762\") " pod="openstack/rabbitmq-cell1-server-0" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.799539 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/5d8e256a-aa1b-4ccb-8e8c-c86be1e31762-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"5d8e256a-aa1b-4ccb-8e8c-c86be1e31762\") " pod="openstack/rabbitmq-cell1-server-0" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.799684 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/5d8e256a-aa1b-4ccb-8e8c-c86be1e31762-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"5d8e256a-aa1b-4ccb-8e8c-c86be1e31762\") " pod="openstack/rabbitmq-cell1-server-0" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.799711 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/5d8e256a-aa1b-4ccb-8e8c-c86be1e31762-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"5d8e256a-aa1b-4ccb-8e8c-c86be1e31762\") " pod="openstack/rabbitmq-cell1-server-0" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.799916 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5d8e256a-aa1b-4ccb-8e8c-c86be1e31762-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"5d8e256a-aa1b-4ccb-8e8c-c86be1e31762\") " pod="openstack/rabbitmq-cell1-server-0" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.800032 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/5d8e256a-aa1b-4ccb-8e8c-c86be1e31762-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"5d8e256a-aa1b-4ccb-8e8c-c86be1e31762\") " pod="openstack/rabbitmq-cell1-server-0" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.800058 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/5d8e256a-aa1b-4ccb-8e8c-c86be1e31762-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"5d8e256a-aa1b-4ccb-8e8c-c86be1e31762\") " pod="openstack/rabbitmq-cell1-server-0" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.901788 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/5d8e256a-aa1b-4ccb-8e8c-c86be1e31762-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"5d8e256a-aa1b-4ccb-8e8c-c86be1e31762\") " pod="openstack/rabbitmq-cell1-server-0" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.901878 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/5d8e256a-aa1b-4ccb-8e8c-c86be1e31762-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"5d8e256a-aa1b-4ccb-8e8c-c86be1e31762\") " pod="openstack/rabbitmq-cell1-server-0" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.901984 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/5d8e256a-aa1b-4ccb-8e8c-c86be1e31762-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"5d8e256a-aa1b-4ccb-8e8c-c86be1e31762\") " pod="openstack/rabbitmq-cell1-server-0" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.902007 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/5d8e256a-aa1b-4ccb-8e8c-c86be1e31762-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"5d8e256a-aa1b-4ccb-8e8c-c86be1e31762\") " pod="openstack/rabbitmq-cell1-server-0" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.902052 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5d8e256a-aa1b-4ccb-8e8c-c86be1e31762-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"5d8e256a-aa1b-4ccb-8e8c-c86be1e31762\") " pod="openstack/rabbitmq-cell1-server-0" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.902102 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/5d8e256a-aa1b-4ccb-8e8c-c86be1e31762-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"5d8e256a-aa1b-4ccb-8e8c-c86be1e31762\") " pod="openstack/rabbitmq-cell1-server-0" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.902118 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/5d8e256a-aa1b-4ccb-8e8c-c86be1e31762-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"5d8e256a-aa1b-4ccb-8e8c-c86be1e31762\") " pod="openstack/rabbitmq-cell1-server-0" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.902137 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/5d8e256a-aa1b-4ccb-8e8c-c86be1e31762-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"5d8e256a-aa1b-4ccb-8e8c-c86be1e31762\") " pod="openstack/rabbitmq-cell1-server-0" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.902204 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7vbvn\" (UniqueName: \"kubernetes.io/projected/5d8e256a-aa1b-4ccb-8e8c-c86be1e31762-kube-api-access-7vbvn\") pod \"rabbitmq-cell1-server-0\" (UID: \"5d8e256a-aa1b-4ccb-8e8c-c86be1e31762\") " pod="openstack/rabbitmq-cell1-server-0" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.902232 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/5d8e256a-aa1b-4ccb-8e8c-c86be1e31762-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"5d8e256a-aa1b-4ccb-8e8c-c86be1e31762\") " pod="openstack/rabbitmq-cell1-server-0" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.902250 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"5d8e256a-aa1b-4ccb-8e8c-c86be1e31762\") " pod="openstack/rabbitmq-cell1-server-0" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.902543 4824 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"5d8e256a-aa1b-4ccb-8e8c-c86be1e31762\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/rabbitmq-cell1-server-0" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.903327 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/5d8e256a-aa1b-4ccb-8e8c-c86be1e31762-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"5d8e256a-aa1b-4ccb-8e8c-c86be1e31762\") " pod="openstack/rabbitmq-cell1-server-0" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.903728 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/5d8e256a-aa1b-4ccb-8e8c-c86be1e31762-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"5d8e256a-aa1b-4ccb-8e8c-c86be1e31762\") " pod="openstack/rabbitmq-cell1-server-0" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.903807 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/5d8e256a-aa1b-4ccb-8e8c-c86be1e31762-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"5d8e256a-aa1b-4ccb-8e8c-c86be1e31762\") " pod="openstack/rabbitmq-cell1-server-0" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.903951 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/5d8e256a-aa1b-4ccb-8e8c-c86be1e31762-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"5d8e256a-aa1b-4ccb-8e8c-c86be1e31762\") " pod="openstack/rabbitmq-cell1-server-0" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.904228 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5d8e256a-aa1b-4ccb-8e8c-c86be1e31762-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"5d8e256a-aa1b-4ccb-8e8c-c86be1e31762\") " pod="openstack/rabbitmq-cell1-server-0" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.908986 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/5d8e256a-aa1b-4ccb-8e8c-c86be1e31762-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"5d8e256a-aa1b-4ccb-8e8c-c86be1e31762\") " pod="openstack/rabbitmq-cell1-server-0" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.909171 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/5d8e256a-aa1b-4ccb-8e8c-c86be1e31762-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"5d8e256a-aa1b-4ccb-8e8c-c86be1e31762\") " pod="openstack/rabbitmq-cell1-server-0" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.909289 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/5d8e256a-aa1b-4ccb-8e8c-c86be1e31762-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"5d8e256a-aa1b-4ccb-8e8c-c86be1e31762\") " pod="openstack/rabbitmq-cell1-server-0" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.911534 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/5d8e256a-aa1b-4ccb-8e8c-c86be1e31762-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"5d8e256a-aa1b-4ccb-8e8c-c86be1e31762\") " pod="openstack/rabbitmq-cell1-server-0" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.918126 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7vbvn\" (UniqueName: \"kubernetes.io/projected/5d8e256a-aa1b-4ccb-8e8c-c86be1e31762-kube-api-access-7vbvn\") pod \"rabbitmq-cell1-server-0\" (UID: \"5d8e256a-aa1b-4ccb-8e8c-c86be1e31762\") " pod="openstack/rabbitmq-cell1-server-0" Jan 21 11:30:06 crc kubenswrapper[4824]: I0121 11:30:06.926582 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"5d8e256a-aa1b-4ccb-8e8c-c86be1e31762\") " pod="openstack/rabbitmq-cell1-server-0" Jan 21 11:30:07 crc kubenswrapper[4824]: I0121 11:30:07.011187 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Jan 21 11:30:07 crc kubenswrapper[4824]: I0121 11:30:07.032331 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Jan 21 11:30:07 crc kubenswrapper[4824]: I0121 11:30:07.179265 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7d84b4d45c-l88p2"] Jan 21 11:30:07 crc kubenswrapper[4824]: I0121 11:30:07.185046 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7d84b4d45c-l88p2" Jan 21 11:30:07 crc kubenswrapper[4824]: I0121 11:30:07.188742 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7d84b4d45c-l88p2"] Jan 21 11:30:07 crc kubenswrapper[4824]: I0121 11:30:07.189157 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-edpm-ipam" Jan 21 11:30:07 crc kubenswrapper[4824]: I0121 11:30:07.245461 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"e4a4d725-1a49-4292-af74-6ab09ab15cdb","Type":"ContainerStarted","Data":"8b84f8fc3aaf1e628cbd83d5b25c754b6f48c97cb5b78b8c4f181d0a7a03affa"} Jan 21 11:30:07 crc kubenswrapper[4824]: I0121 11:30:07.309513 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/aed87a15-74e4-47e3-8f90-f98672d62792-openstack-edpm-ipam\") pod \"dnsmasq-dns-7d84b4d45c-l88p2\" (UID: \"aed87a15-74e4-47e3-8f90-f98672d62792\") " pod="openstack/dnsmasq-dns-7d84b4d45c-l88p2" Jan 21 11:30:07 crc kubenswrapper[4824]: I0121 11:30:07.309605 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/aed87a15-74e4-47e3-8f90-f98672d62792-ovsdbserver-sb\") pod \"dnsmasq-dns-7d84b4d45c-l88p2\" (UID: \"aed87a15-74e4-47e3-8f90-f98672d62792\") " pod="openstack/dnsmasq-dns-7d84b4d45c-l88p2" Jan 21 11:30:07 crc kubenswrapper[4824]: I0121 11:30:07.309629 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/aed87a15-74e4-47e3-8f90-f98672d62792-dns-svc\") pod \"dnsmasq-dns-7d84b4d45c-l88p2\" (UID: \"aed87a15-74e4-47e3-8f90-f98672d62792\") " pod="openstack/dnsmasq-dns-7d84b4d45c-l88p2" Jan 21 11:30:07 crc kubenswrapper[4824]: I0121 11:30:07.309679 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/aed87a15-74e4-47e3-8f90-f98672d62792-dns-swift-storage-0\") pod \"dnsmasq-dns-7d84b4d45c-l88p2\" (UID: \"aed87a15-74e4-47e3-8f90-f98672d62792\") " pod="openstack/dnsmasq-dns-7d84b4d45c-l88p2" Jan 21 11:30:07 crc kubenswrapper[4824]: I0121 11:30:07.309712 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/aed87a15-74e4-47e3-8f90-f98672d62792-ovsdbserver-nb\") pod \"dnsmasq-dns-7d84b4d45c-l88p2\" (UID: \"aed87a15-74e4-47e3-8f90-f98672d62792\") " pod="openstack/dnsmasq-dns-7d84b4d45c-l88p2" Jan 21 11:30:07 crc kubenswrapper[4824]: I0121 11:30:07.309774 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-txbf7\" (UniqueName: \"kubernetes.io/projected/aed87a15-74e4-47e3-8f90-f98672d62792-kube-api-access-txbf7\") pod \"dnsmasq-dns-7d84b4d45c-l88p2\" (UID: \"aed87a15-74e4-47e3-8f90-f98672d62792\") " pod="openstack/dnsmasq-dns-7d84b4d45c-l88p2" Jan 21 11:30:07 crc kubenswrapper[4824]: I0121 11:30:07.310004 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aed87a15-74e4-47e3-8f90-f98672d62792-config\") pod \"dnsmasq-dns-7d84b4d45c-l88p2\" (UID: \"aed87a15-74e4-47e3-8f90-f98672d62792\") " pod="openstack/dnsmasq-dns-7d84b4d45c-l88p2" Jan 21 11:30:07 crc kubenswrapper[4824]: I0121 11:30:07.411692 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aed87a15-74e4-47e3-8f90-f98672d62792-config\") pod \"dnsmasq-dns-7d84b4d45c-l88p2\" (UID: \"aed87a15-74e4-47e3-8f90-f98672d62792\") " pod="openstack/dnsmasq-dns-7d84b4d45c-l88p2" Jan 21 11:30:07 crc kubenswrapper[4824]: I0121 11:30:07.411787 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/aed87a15-74e4-47e3-8f90-f98672d62792-openstack-edpm-ipam\") pod \"dnsmasq-dns-7d84b4d45c-l88p2\" (UID: \"aed87a15-74e4-47e3-8f90-f98672d62792\") " pod="openstack/dnsmasq-dns-7d84b4d45c-l88p2" Jan 21 11:30:07 crc kubenswrapper[4824]: I0121 11:30:07.411889 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/aed87a15-74e4-47e3-8f90-f98672d62792-ovsdbserver-sb\") pod \"dnsmasq-dns-7d84b4d45c-l88p2\" (UID: \"aed87a15-74e4-47e3-8f90-f98672d62792\") " pod="openstack/dnsmasq-dns-7d84b4d45c-l88p2" Jan 21 11:30:07 crc kubenswrapper[4824]: I0121 11:30:07.411909 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/aed87a15-74e4-47e3-8f90-f98672d62792-dns-svc\") pod \"dnsmasq-dns-7d84b4d45c-l88p2\" (UID: \"aed87a15-74e4-47e3-8f90-f98672d62792\") " pod="openstack/dnsmasq-dns-7d84b4d45c-l88p2" Jan 21 11:30:07 crc kubenswrapper[4824]: I0121 11:30:07.411932 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/aed87a15-74e4-47e3-8f90-f98672d62792-dns-swift-storage-0\") pod \"dnsmasq-dns-7d84b4d45c-l88p2\" (UID: \"aed87a15-74e4-47e3-8f90-f98672d62792\") " pod="openstack/dnsmasq-dns-7d84b4d45c-l88p2" Jan 21 11:30:07 crc kubenswrapper[4824]: I0121 11:30:07.411971 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/aed87a15-74e4-47e3-8f90-f98672d62792-ovsdbserver-nb\") pod \"dnsmasq-dns-7d84b4d45c-l88p2\" (UID: \"aed87a15-74e4-47e3-8f90-f98672d62792\") " pod="openstack/dnsmasq-dns-7d84b4d45c-l88p2" Jan 21 11:30:07 crc kubenswrapper[4824]: I0121 11:30:07.412035 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-txbf7\" (UniqueName: \"kubernetes.io/projected/aed87a15-74e4-47e3-8f90-f98672d62792-kube-api-access-txbf7\") pod \"dnsmasq-dns-7d84b4d45c-l88p2\" (UID: \"aed87a15-74e4-47e3-8f90-f98672d62792\") " pod="openstack/dnsmasq-dns-7d84b4d45c-l88p2" Jan 21 11:30:07 crc kubenswrapper[4824]: I0121 11:30:07.412644 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aed87a15-74e4-47e3-8f90-f98672d62792-config\") pod \"dnsmasq-dns-7d84b4d45c-l88p2\" (UID: \"aed87a15-74e4-47e3-8f90-f98672d62792\") " pod="openstack/dnsmasq-dns-7d84b4d45c-l88p2" Jan 21 11:30:07 crc kubenswrapper[4824]: I0121 11:30:07.412659 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/aed87a15-74e4-47e3-8f90-f98672d62792-ovsdbserver-sb\") pod \"dnsmasq-dns-7d84b4d45c-l88p2\" (UID: \"aed87a15-74e4-47e3-8f90-f98672d62792\") " pod="openstack/dnsmasq-dns-7d84b4d45c-l88p2" Jan 21 11:30:07 crc kubenswrapper[4824]: I0121 11:30:07.412695 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/aed87a15-74e4-47e3-8f90-f98672d62792-openstack-edpm-ipam\") pod \"dnsmasq-dns-7d84b4d45c-l88p2\" (UID: \"aed87a15-74e4-47e3-8f90-f98672d62792\") " pod="openstack/dnsmasq-dns-7d84b4d45c-l88p2" Jan 21 11:30:07 crc kubenswrapper[4824]: I0121 11:30:07.412735 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/aed87a15-74e4-47e3-8f90-f98672d62792-ovsdbserver-nb\") pod \"dnsmasq-dns-7d84b4d45c-l88p2\" (UID: \"aed87a15-74e4-47e3-8f90-f98672d62792\") " pod="openstack/dnsmasq-dns-7d84b4d45c-l88p2" Jan 21 11:30:07 crc kubenswrapper[4824]: I0121 11:30:07.412838 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/aed87a15-74e4-47e3-8f90-f98672d62792-dns-swift-storage-0\") pod \"dnsmasq-dns-7d84b4d45c-l88p2\" (UID: \"aed87a15-74e4-47e3-8f90-f98672d62792\") " pod="openstack/dnsmasq-dns-7d84b4d45c-l88p2" Jan 21 11:30:07 crc kubenswrapper[4824]: I0121 11:30:07.413222 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/aed87a15-74e4-47e3-8f90-f98672d62792-dns-svc\") pod \"dnsmasq-dns-7d84b4d45c-l88p2\" (UID: \"aed87a15-74e4-47e3-8f90-f98672d62792\") " pod="openstack/dnsmasq-dns-7d84b4d45c-l88p2" Jan 21 11:30:07 crc kubenswrapper[4824]: I0121 11:30:07.426112 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-txbf7\" (UniqueName: \"kubernetes.io/projected/aed87a15-74e4-47e3-8f90-f98672d62792-kube-api-access-txbf7\") pod \"dnsmasq-dns-7d84b4d45c-l88p2\" (UID: \"aed87a15-74e4-47e3-8f90-f98672d62792\") " pod="openstack/dnsmasq-dns-7d84b4d45c-l88p2" Jan 21 11:30:07 crc kubenswrapper[4824]: I0121 11:30:07.499556 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7d84b4d45c-l88p2" Jan 21 11:30:07 crc kubenswrapper[4824]: I0121 11:30:07.902511 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Jan 21 11:30:07 crc kubenswrapper[4824]: I0121 11:30:07.910525 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7d84b4d45c-l88p2"] Jan 21 11:30:08 crc kubenswrapper[4824]: I0121 11:30:08.057225 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="156fec50-b486-4e84-a7bf-b40491a863c7" path="/var/lib/kubelet/pods/156fec50-b486-4e84-a7bf-b40491a863c7/volumes" Jan 21 11:30:08 crc kubenswrapper[4824]: I0121 11:30:08.059178 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="84e74d7b-18a0-4a3f-8680-6246ac538a6e" path="/var/lib/kubelet/pods/84e74d7b-18a0-4a3f-8680-6246ac538a6e/volumes" Jan 21 11:30:08 crc kubenswrapper[4824]: I0121 11:30:08.281107 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"e4a4d725-1a49-4292-af74-6ab09ab15cdb","Type":"ContainerStarted","Data":"8915f56cc0770d64ee6a0d4ce2d2834b7ba05053d7b8c9baa030f6a50a88dee1"} Jan 21 11:30:08 crc kubenswrapper[4824]: I0121 11:30:08.292103 4824 generic.go:334] "Generic (PLEG): container finished" podID="aed87a15-74e4-47e3-8f90-f98672d62792" containerID="2b3debae743097c64a642d1edc62ad587f2e44aabf5b01018238eb55be99a0de" exitCode=0 Jan 21 11:30:08 crc kubenswrapper[4824]: I0121 11:30:08.292153 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d84b4d45c-l88p2" event={"ID":"aed87a15-74e4-47e3-8f90-f98672d62792","Type":"ContainerDied","Data":"2b3debae743097c64a642d1edc62ad587f2e44aabf5b01018238eb55be99a0de"} Jan 21 11:30:08 crc kubenswrapper[4824]: I0121 11:30:08.292172 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d84b4d45c-l88p2" event={"ID":"aed87a15-74e4-47e3-8f90-f98672d62792","Type":"ContainerStarted","Data":"d3d7d6a8cb817cd1fffab43e3dc0bcf2917c60a0925857f2eded916d57a51538"} Jan 21 11:30:08 crc kubenswrapper[4824]: I0121 11:30:08.301111 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"5d8e256a-aa1b-4ccb-8e8c-c86be1e31762","Type":"ContainerStarted","Data":"cb0628a064449b01379611bee44c81428d76b736554608f5c0a5f9e831d04f4b"} Jan 21 11:30:09 crc kubenswrapper[4824]: I0121 11:30:09.309076 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d84b4d45c-l88p2" event={"ID":"aed87a15-74e4-47e3-8f90-f98672d62792","Type":"ContainerStarted","Data":"3d9bc7267f56122685da59d95824994c3ba510410d650c54f20f513e800a799f"} Jan 21 11:30:09 crc kubenswrapper[4824]: I0121 11:30:09.309429 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7d84b4d45c-l88p2" Jan 21 11:30:09 crc kubenswrapper[4824]: I0121 11:30:09.310095 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"5d8e256a-aa1b-4ccb-8e8c-c86be1e31762","Type":"ContainerStarted","Data":"1ffd78c66a880f9e4de61c175a16fa8e25a34ed2dd520861ea49b68a2f4172fe"} Jan 21 11:30:09 crc kubenswrapper[4824]: I0121 11:30:09.324673 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7d84b4d45c-l88p2" podStartSLOduration=2.324658385 podStartE2EDuration="2.324658385s" podCreationTimestamp="2026-01-21 11:30:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:30:09.321753348 +0000 UTC m=+1151.614782690" watchObservedRunningTime="2026-01-21 11:30:09.324658385 +0000 UTC m=+1151.617687677" Jan 21 11:30:16 crc kubenswrapper[4824]: I0121 11:30:16.065321 4824 patch_prober.go:28] interesting pod/machine-config-daemon-zqjg9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 21 11:30:16 crc kubenswrapper[4824]: I0121 11:30:16.065722 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 21 11:30:16 crc kubenswrapper[4824]: I0121 11:30:16.065762 4824 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" Jan 21 11:30:16 crc kubenswrapper[4824]: I0121 11:30:16.066405 4824 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"74ad8e1889db466232cc3528a7e51d12f4be65b09faed3c15716a73518537690"} pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 21 11:30:16 crc kubenswrapper[4824]: I0121 11:30:16.066456 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" containerName="machine-config-daemon" containerID="cri-o://74ad8e1889db466232cc3528a7e51d12f4be65b09faed3c15716a73518537690" gracePeriod=600 Jan 21 11:30:16 crc kubenswrapper[4824]: I0121 11:30:16.358715 4824 generic.go:334] "Generic (PLEG): container finished" podID="33f3d922-4ffe-409b-a49a-d88c85898260" containerID="74ad8e1889db466232cc3528a7e51d12f4be65b09faed3c15716a73518537690" exitCode=0 Jan 21 11:30:16 crc kubenswrapper[4824]: I0121 11:30:16.358867 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" event={"ID":"33f3d922-4ffe-409b-a49a-d88c85898260","Type":"ContainerDied","Data":"74ad8e1889db466232cc3528a7e51d12f4be65b09faed3c15716a73518537690"} Jan 21 11:30:16 crc kubenswrapper[4824]: I0121 11:30:16.358993 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" event={"ID":"33f3d922-4ffe-409b-a49a-d88c85898260","Type":"ContainerStarted","Data":"f5273806c339b109c1d902705cf82c97535d3adc4a021ab1d42339f22a956af4"} Jan 21 11:30:16 crc kubenswrapper[4824]: I0121 11:30:16.359016 4824 scope.go:117] "RemoveContainer" containerID="365550d435931ba3a56d32c848ae2c97f8d7721cf74faadf03f5fc5d15e5b5a5" Jan 21 11:30:17 crc kubenswrapper[4824]: I0121 11:30:17.501094 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7d84b4d45c-l88p2" Jan 21 11:30:17 crc kubenswrapper[4824]: I0121 11:30:17.537450 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6b7bbf7cf9-tmvm9"] Jan 21 11:30:17 crc kubenswrapper[4824]: I0121 11:30:17.537669 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6b7bbf7cf9-tmvm9" podUID="8ad8dd9e-6fe4-462b-98a9-d0192072589d" containerName="dnsmasq-dns" containerID="cri-o://8687725a582d2593d699bc27c414dbcbeca025f725a9dfdc0c21cbe095d4f575" gracePeriod=10 Jan 21 11:30:17 crc kubenswrapper[4824]: I0121 11:30:17.650026 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6f6df4f56c-26d8b"] Jan 21 11:30:17 crc kubenswrapper[4824]: I0121 11:30:17.651595 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6f6df4f56c-26d8b" Jan 21 11:30:17 crc kubenswrapper[4824]: I0121 11:30:17.677591 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6f6df4f56c-26d8b"] Jan 21 11:30:17 crc kubenswrapper[4824]: I0121 11:30:17.786862 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/976a6446-e79c-42db-be45-43147767f1da-dns-swift-storage-0\") pod \"dnsmasq-dns-6f6df4f56c-26d8b\" (UID: \"976a6446-e79c-42db-be45-43147767f1da\") " pod="openstack/dnsmasq-dns-6f6df4f56c-26d8b" Jan 21 11:30:17 crc kubenswrapper[4824]: I0121 11:30:17.787458 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pvbf7\" (UniqueName: \"kubernetes.io/projected/976a6446-e79c-42db-be45-43147767f1da-kube-api-access-pvbf7\") pod \"dnsmasq-dns-6f6df4f56c-26d8b\" (UID: \"976a6446-e79c-42db-be45-43147767f1da\") " pod="openstack/dnsmasq-dns-6f6df4f56c-26d8b" Jan 21 11:30:17 crc kubenswrapper[4824]: I0121 11:30:17.787503 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/976a6446-e79c-42db-be45-43147767f1da-config\") pod \"dnsmasq-dns-6f6df4f56c-26d8b\" (UID: \"976a6446-e79c-42db-be45-43147767f1da\") " pod="openstack/dnsmasq-dns-6f6df4f56c-26d8b" Jan 21 11:30:17 crc kubenswrapper[4824]: I0121 11:30:17.787540 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/976a6446-e79c-42db-be45-43147767f1da-openstack-edpm-ipam\") pod \"dnsmasq-dns-6f6df4f56c-26d8b\" (UID: \"976a6446-e79c-42db-be45-43147767f1da\") " pod="openstack/dnsmasq-dns-6f6df4f56c-26d8b" Jan 21 11:30:17 crc kubenswrapper[4824]: I0121 11:30:17.787557 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/976a6446-e79c-42db-be45-43147767f1da-ovsdbserver-sb\") pod \"dnsmasq-dns-6f6df4f56c-26d8b\" (UID: \"976a6446-e79c-42db-be45-43147767f1da\") " pod="openstack/dnsmasq-dns-6f6df4f56c-26d8b" Jan 21 11:30:17 crc kubenswrapper[4824]: I0121 11:30:17.787654 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/976a6446-e79c-42db-be45-43147767f1da-dns-svc\") pod \"dnsmasq-dns-6f6df4f56c-26d8b\" (UID: \"976a6446-e79c-42db-be45-43147767f1da\") " pod="openstack/dnsmasq-dns-6f6df4f56c-26d8b" Jan 21 11:30:17 crc kubenswrapper[4824]: I0121 11:30:17.787779 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/976a6446-e79c-42db-be45-43147767f1da-ovsdbserver-nb\") pod \"dnsmasq-dns-6f6df4f56c-26d8b\" (UID: \"976a6446-e79c-42db-be45-43147767f1da\") " pod="openstack/dnsmasq-dns-6f6df4f56c-26d8b" Jan 21 11:30:17 crc kubenswrapper[4824]: I0121 11:30:17.889684 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/976a6446-e79c-42db-be45-43147767f1da-dns-swift-storage-0\") pod \"dnsmasq-dns-6f6df4f56c-26d8b\" (UID: \"976a6446-e79c-42db-be45-43147767f1da\") " pod="openstack/dnsmasq-dns-6f6df4f56c-26d8b" Jan 21 11:30:17 crc kubenswrapper[4824]: I0121 11:30:17.889777 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pvbf7\" (UniqueName: \"kubernetes.io/projected/976a6446-e79c-42db-be45-43147767f1da-kube-api-access-pvbf7\") pod \"dnsmasq-dns-6f6df4f56c-26d8b\" (UID: \"976a6446-e79c-42db-be45-43147767f1da\") " pod="openstack/dnsmasq-dns-6f6df4f56c-26d8b" Jan 21 11:30:17 crc kubenswrapper[4824]: I0121 11:30:17.889830 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/976a6446-e79c-42db-be45-43147767f1da-config\") pod \"dnsmasq-dns-6f6df4f56c-26d8b\" (UID: \"976a6446-e79c-42db-be45-43147767f1da\") " pod="openstack/dnsmasq-dns-6f6df4f56c-26d8b" Jan 21 11:30:17 crc kubenswrapper[4824]: I0121 11:30:17.889863 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/976a6446-e79c-42db-be45-43147767f1da-openstack-edpm-ipam\") pod \"dnsmasq-dns-6f6df4f56c-26d8b\" (UID: \"976a6446-e79c-42db-be45-43147767f1da\") " pod="openstack/dnsmasq-dns-6f6df4f56c-26d8b" Jan 21 11:30:17 crc kubenswrapper[4824]: I0121 11:30:17.889892 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/976a6446-e79c-42db-be45-43147767f1da-ovsdbserver-sb\") pod \"dnsmasq-dns-6f6df4f56c-26d8b\" (UID: \"976a6446-e79c-42db-be45-43147767f1da\") " pod="openstack/dnsmasq-dns-6f6df4f56c-26d8b" Jan 21 11:30:17 crc kubenswrapper[4824]: I0121 11:30:17.889952 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/976a6446-e79c-42db-be45-43147767f1da-dns-svc\") pod \"dnsmasq-dns-6f6df4f56c-26d8b\" (UID: \"976a6446-e79c-42db-be45-43147767f1da\") " pod="openstack/dnsmasq-dns-6f6df4f56c-26d8b" Jan 21 11:30:17 crc kubenswrapper[4824]: I0121 11:30:17.890057 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/976a6446-e79c-42db-be45-43147767f1da-ovsdbserver-nb\") pod \"dnsmasq-dns-6f6df4f56c-26d8b\" (UID: \"976a6446-e79c-42db-be45-43147767f1da\") " pod="openstack/dnsmasq-dns-6f6df4f56c-26d8b" Jan 21 11:30:17 crc kubenswrapper[4824]: I0121 11:30:17.891186 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/976a6446-e79c-42db-be45-43147767f1da-ovsdbserver-sb\") pod \"dnsmasq-dns-6f6df4f56c-26d8b\" (UID: \"976a6446-e79c-42db-be45-43147767f1da\") " pod="openstack/dnsmasq-dns-6f6df4f56c-26d8b" Jan 21 11:30:17 crc kubenswrapper[4824]: I0121 11:30:17.891189 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/976a6446-e79c-42db-be45-43147767f1da-dns-swift-storage-0\") pod \"dnsmasq-dns-6f6df4f56c-26d8b\" (UID: \"976a6446-e79c-42db-be45-43147767f1da\") " pod="openstack/dnsmasq-dns-6f6df4f56c-26d8b" Jan 21 11:30:17 crc kubenswrapper[4824]: I0121 11:30:17.891192 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/976a6446-e79c-42db-be45-43147767f1da-config\") pod \"dnsmasq-dns-6f6df4f56c-26d8b\" (UID: \"976a6446-e79c-42db-be45-43147767f1da\") " pod="openstack/dnsmasq-dns-6f6df4f56c-26d8b" Jan 21 11:30:17 crc kubenswrapper[4824]: I0121 11:30:17.891283 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/976a6446-e79c-42db-be45-43147767f1da-ovsdbserver-nb\") pod \"dnsmasq-dns-6f6df4f56c-26d8b\" (UID: \"976a6446-e79c-42db-be45-43147767f1da\") " pod="openstack/dnsmasq-dns-6f6df4f56c-26d8b" Jan 21 11:30:17 crc kubenswrapper[4824]: I0121 11:30:17.891734 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/976a6446-e79c-42db-be45-43147767f1da-openstack-edpm-ipam\") pod \"dnsmasq-dns-6f6df4f56c-26d8b\" (UID: \"976a6446-e79c-42db-be45-43147767f1da\") " pod="openstack/dnsmasq-dns-6f6df4f56c-26d8b" Jan 21 11:30:17 crc kubenswrapper[4824]: I0121 11:30:17.892138 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/976a6446-e79c-42db-be45-43147767f1da-dns-svc\") pod \"dnsmasq-dns-6f6df4f56c-26d8b\" (UID: \"976a6446-e79c-42db-be45-43147767f1da\") " pod="openstack/dnsmasq-dns-6f6df4f56c-26d8b" Jan 21 11:30:17 crc kubenswrapper[4824]: I0121 11:30:17.907445 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pvbf7\" (UniqueName: \"kubernetes.io/projected/976a6446-e79c-42db-be45-43147767f1da-kube-api-access-pvbf7\") pod \"dnsmasq-dns-6f6df4f56c-26d8b\" (UID: \"976a6446-e79c-42db-be45-43147767f1da\") " pod="openstack/dnsmasq-dns-6f6df4f56c-26d8b" Jan 21 11:30:17 crc kubenswrapper[4824]: I0121 11:30:17.996738 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b7bbf7cf9-tmvm9" Jan 21 11:30:18 crc kubenswrapper[4824]: I0121 11:30:18.002434 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6f6df4f56c-26d8b" Jan 21 11:30:18 crc kubenswrapper[4824]: I0121 11:30:18.092504 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8ad8dd9e-6fe4-462b-98a9-d0192072589d-dns-svc\") pod \"8ad8dd9e-6fe4-462b-98a9-d0192072589d\" (UID: \"8ad8dd9e-6fe4-462b-98a9-d0192072589d\") " Jan 21 11:30:18 crc kubenswrapper[4824]: I0121 11:30:18.093059 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8ad8dd9e-6fe4-462b-98a9-d0192072589d-ovsdbserver-nb\") pod \"8ad8dd9e-6fe4-462b-98a9-d0192072589d\" (UID: \"8ad8dd9e-6fe4-462b-98a9-d0192072589d\") " Jan 21 11:30:18 crc kubenswrapper[4824]: I0121 11:30:18.093100 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bkm2s\" (UniqueName: \"kubernetes.io/projected/8ad8dd9e-6fe4-462b-98a9-d0192072589d-kube-api-access-bkm2s\") pod \"8ad8dd9e-6fe4-462b-98a9-d0192072589d\" (UID: \"8ad8dd9e-6fe4-462b-98a9-d0192072589d\") " Jan 21 11:30:18 crc kubenswrapper[4824]: I0121 11:30:18.093133 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8ad8dd9e-6fe4-462b-98a9-d0192072589d-config\") pod \"8ad8dd9e-6fe4-462b-98a9-d0192072589d\" (UID: \"8ad8dd9e-6fe4-462b-98a9-d0192072589d\") " Jan 21 11:30:18 crc kubenswrapper[4824]: I0121 11:30:18.093175 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8ad8dd9e-6fe4-462b-98a9-d0192072589d-dns-swift-storage-0\") pod \"8ad8dd9e-6fe4-462b-98a9-d0192072589d\" (UID: \"8ad8dd9e-6fe4-462b-98a9-d0192072589d\") " Jan 21 11:30:18 crc kubenswrapper[4824]: I0121 11:30:18.093227 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8ad8dd9e-6fe4-462b-98a9-d0192072589d-ovsdbserver-sb\") pod \"8ad8dd9e-6fe4-462b-98a9-d0192072589d\" (UID: \"8ad8dd9e-6fe4-462b-98a9-d0192072589d\") " Jan 21 11:30:18 crc kubenswrapper[4824]: I0121 11:30:18.097353 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8ad8dd9e-6fe4-462b-98a9-d0192072589d-kube-api-access-bkm2s" (OuterVolumeSpecName: "kube-api-access-bkm2s") pod "8ad8dd9e-6fe4-462b-98a9-d0192072589d" (UID: "8ad8dd9e-6fe4-462b-98a9-d0192072589d"). InnerVolumeSpecName "kube-api-access-bkm2s". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:30:18 crc kubenswrapper[4824]: I0121 11:30:18.137281 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8ad8dd9e-6fe4-462b-98a9-d0192072589d-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "8ad8dd9e-6fe4-462b-98a9-d0192072589d" (UID: "8ad8dd9e-6fe4-462b-98a9-d0192072589d"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:30:18 crc kubenswrapper[4824]: I0121 11:30:18.137333 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8ad8dd9e-6fe4-462b-98a9-d0192072589d-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "8ad8dd9e-6fe4-462b-98a9-d0192072589d" (UID: "8ad8dd9e-6fe4-462b-98a9-d0192072589d"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:30:18 crc kubenswrapper[4824]: I0121 11:30:18.137304 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8ad8dd9e-6fe4-462b-98a9-d0192072589d-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "8ad8dd9e-6fe4-462b-98a9-d0192072589d" (UID: "8ad8dd9e-6fe4-462b-98a9-d0192072589d"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:30:18 crc kubenswrapper[4824]: I0121 11:30:18.140631 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8ad8dd9e-6fe4-462b-98a9-d0192072589d-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "8ad8dd9e-6fe4-462b-98a9-d0192072589d" (UID: "8ad8dd9e-6fe4-462b-98a9-d0192072589d"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:30:18 crc kubenswrapper[4824]: I0121 11:30:18.146269 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8ad8dd9e-6fe4-462b-98a9-d0192072589d-config" (OuterVolumeSpecName: "config") pod "8ad8dd9e-6fe4-462b-98a9-d0192072589d" (UID: "8ad8dd9e-6fe4-462b-98a9-d0192072589d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:30:18 crc kubenswrapper[4824]: I0121 11:30:18.196975 4824 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8ad8dd9e-6fe4-462b-98a9-d0192072589d-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Jan 21 11:30:18 crc kubenswrapper[4824]: I0121 11:30:18.197005 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bkm2s\" (UniqueName: \"kubernetes.io/projected/8ad8dd9e-6fe4-462b-98a9-d0192072589d-kube-api-access-bkm2s\") on node \"crc\" DevicePath \"\"" Jan 21 11:30:18 crc kubenswrapper[4824]: I0121 11:30:18.197017 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8ad8dd9e-6fe4-462b-98a9-d0192072589d-config\") on node \"crc\" DevicePath \"\"" Jan 21 11:30:18 crc kubenswrapper[4824]: I0121 11:30:18.197027 4824 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8ad8dd9e-6fe4-462b-98a9-d0192072589d-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Jan 21 11:30:18 crc kubenswrapper[4824]: I0121 11:30:18.197036 4824 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8ad8dd9e-6fe4-462b-98a9-d0192072589d-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Jan 21 11:30:18 crc kubenswrapper[4824]: I0121 11:30:18.197044 4824 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8ad8dd9e-6fe4-462b-98a9-d0192072589d-dns-svc\") on node \"crc\" DevicePath \"\"" Jan 21 11:30:18 crc kubenswrapper[4824]: I0121 11:30:18.375418 4824 generic.go:334] "Generic (PLEG): container finished" podID="8ad8dd9e-6fe4-462b-98a9-d0192072589d" containerID="8687725a582d2593d699bc27c414dbcbeca025f725a9dfdc0c21cbe095d4f575" exitCode=0 Jan 21 11:30:18 crc kubenswrapper[4824]: I0121 11:30:18.375460 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7bbf7cf9-tmvm9" event={"ID":"8ad8dd9e-6fe4-462b-98a9-d0192072589d","Type":"ContainerDied","Data":"8687725a582d2593d699bc27c414dbcbeca025f725a9dfdc0c21cbe095d4f575"} Jan 21 11:30:18 crc kubenswrapper[4824]: I0121 11:30:18.375483 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7bbf7cf9-tmvm9" event={"ID":"8ad8dd9e-6fe4-462b-98a9-d0192072589d","Type":"ContainerDied","Data":"6738b1680c3c03cab4e1e7a4ccdc6941fbdb963c0035872cd1e033904215c51b"} Jan 21 11:30:18 crc kubenswrapper[4824]: I0121 11:30:18.375499 4824 scope.go:117] "RemoveContainer" containerID="8687725a582d2593d699bc27c414dbcbeca025f725a9dfdc0c21cbe095d4f575" Jan 21 11:30:18 crc kubenswrapper[4824]: I0121 11:30:18.375592 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b7bbf7cf9-tmvm9" Jan 21 11:30:18 crc kubenswrapper[4824]: I0121 11:30:18.393511 4824 scope.go:117] "RemoveContainer" containerID="ed4051f2d4931880d9659716eb3f4ebdfc337e399698e29d497ad42bb33cd14f" Jan 21 11:30:18 crc kubenswrapper[4824]: I0121 11:30:18.402292 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6b7bbf7cf9-tmvm9"] Jan 21 11:30:18 crc kubenswrapper[4824]: I0121 11:30:18.409051 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6b7bbf7cf9-tmvm9"] Jan 21 11:30:18 crc kubenswrapper[4824]: I0121 11:30:18.416111 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6f6df4f56c-26d8b"] Jan 21 11:30:18 crc kubenswrapper[4824]: W0121 11:30:18.419115 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod976a6446_e79c_42db_be45_43147767f1da.slice/crio-c4cb9cd169639fc7ee58ef770aad0eba3e444be63bdd9e108aa65f4d2ec941f5 WatchSource:0}: Error finding container c4cb9cd169639fc7ee58ef770aad0eba3e444be63bdd9e108aa65f4d2ec941f5: Status 404 returned error can't find the container with id c4cb9cd169639fc7ee58ef770aad0eba3e444be63bdd9e108aa65f4d2ec941f5 Jan 21 11:30:18 crc kubenswrapper[4824]: I0121 11:30:18.426784 4824 scope.go:117] "RemoveContainer" containerID="8687725a582d2593d699bc27c414dbcbeca025f725a9dfdc0c21cbe095d4f575" Jan 21 11:30:18 crc kubenswrapper[4824]: E0121 11:30:18.427099 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8687725a582d2593d699bc27c414dbcbeca025f725a9dfdc0c21cbe095d4f575\": container with ID starting with 8687725a582d2593d699bc27c414dbcbeca025f725a9dfdc0c21cbe095d4f575 not found: ID does not exist" containerID="8687725a582d2593d699bc27c414dbcbeca025f725a9dfdc0c21cbe095d4f575" Jan 21 11:30:18 crc kubenswrapper[4824]: I0121 11:30:18.427130 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8687725a582d2593d699bc27c414dbcbeca025f725a9dfdc0c21cbe095d4f575"} err="failed to get container status \"8687725a582d2593d699bc27c414dbcbeca025f725a9dfdc0c21cbe095d4f575\": rpc error: code = NotFound desc = could not find container \"8687725a582d2593d699bc27c414dbcbeca025f725a9dfdc0c21cbe095d4f575\": container with ID starting with 8687725a582d2593d699bc27c414dbcbeca025f725a9dfdc0c21cbe095d4f575 not found: ID does not exist" Jan 21 11:30:18 crc kubenswrapper[4824]: I0121 11:30:18.427150 4824 scope.go:117] "RemoveContainer" containerID="ed4051f2d4931880d9659716eb3f4ebdfc337e399698e29d497ad42bb33cd14f" Jan 21 11:30:18 crc kubenswrapper[4824]: E0121 11:30:18.427325 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ed4051f2d4931880d9659716eb3f4ebdfc337e399698e29d497ad42bb33cd14f\": container with ID starting with ed4051f2d4931880d9659716eb3f4ebdfc337e399698e29d497ad42bb33cd14f not found: ID does not exist" containerID="ed4051f2d4931880d9659716eb3f4ebdfc337e399698e29d497ad42bb33cd14f" Jan 21 11:30:18 crc kubenswrapper[4824]: I0121 11:30:18.427358 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ed4051f2d4931880d9659716eb3f4ebdfc337e399698e29d497ad42bb33cd14f"} err="failed to get container status \"ed4051f2d4931880d9659716eb3f4ebdfc337e399698e29d497ad42bb33cd14f\": rpc error: code = NotFound desc = could not find container \"ed4051f2d4931880d9659716eb3f4ebdfc337e399698e29d497ad42bb33cd14f\": container with ID starting with ed4051f2d4931880d9659716eb3f4ebdfc337e399698e29d497ad42bb33cd14f not found: ID does not exist" Jan 21 11:30:19 crc kubenswrapper[4824]: I0121 11:30:19.386202 4824 generic.go:334] "Generic (PLEG): container finished" podID="976a6446-e79c-42db-be45-43147767f1da" containerID="5ea28abe017300018ccb30238cf19011b8c323951c42f861b0e10547b86eb301" exitCode=0 Jan 21 11:30:19 crc kubenswrapper[4824]: I0121 11:30:19.386394 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6f6df4f56c-26d8b" event={"ID":"976a6446-e79c-42db-be45-43147767f1da","Type":"ContainerDied","Data":"5ea28abe017300018ccb30238cf19011b8c323951c42f861b0e10547b86eb301"} Jan 21 11:30:19 crc kubenswrapper[4824]: I0121 11:30:19.386702 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6f6df4f56c-26d8b" event={"ID":"976a6446-e79c-42db-be45-43147767f1da","Type":"ContainerStarted","Data":"c4cb9cd169639fc7ee58ef770aad0eba3e444be63bdd9e108aa65f4d2ec941f5"} Jan 21 11:30:20 crc kubenswrapper[4824]: I0121 11:30:20.067783 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8ad8dd9e-6fe4-462b-98a9-d0192072589d" path="/var/lib/kubelet/pods/8ad8dd9e-6fe4-462b-98a9-d0192072589d/volumes" Jan 21 11:30:20 crc kubenswrapper[4824]: I0121 11:30:20.394586 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6f6df4f56c-26d8b" event={"ID":"976a6446-e79c-42db-be45-43147767f1da","Type":"ContainerStarted","Data":"8e7c6b4f7a98a2f392bc8913c5a09cf641991cd342a6a59f4cbd9a66fec5cf89"} Jan 21 11:30:20 crc kubenswrapper[4824]: I0121 11:30:20.394985 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6f6df4f56c-26d8b" Jan 21 11:30:20 crc kubenswrapper[4824]: I0121 11:30:20.414562 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6f6df4f56c-26d8b" podStartSLOduration=3.414548463 podStartE2EDuration="3.414548463s" podCreationTimestamp="2026-01-21 11:30:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:30:20.408675849 +0000 UTC m=+1162.701705141" watchObservedRunningTime="2026-01-21 11:30:20.414548463 +0000 UTC m=+1162.707577754" Jan 21 11:30:28 crc kubenswrapper[4824]: I0121 11:30:28.003656 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6f6df4f56c-26d8b" Jan 21 11:30:28 crc kubenswrapper[4824]: I0121 11:30:28.043150 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7d84b4d45c-l88p2"] Jan 21 11:30:28 crc kubenswrapper[4824]: I0121 11:30:28.043358 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7d84b4d45c-l88p2" podUID="aed87a15-74e4-47e3-8f90-f98672d62792" containerName="dnsmasq-dns" containerID="cri-o://3d9bc7267f56122685da59d95824994c3ba510410d650c54f20f513e800a799f" gracePeriod=10 Jan 21 11:30:28 crc kubenswrapper[4824]: I0121 11:30:28.409058 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7d84b4d45c-l88p2" Jan 21 11:30:28 crc kubenswrapper[4824]: I0121 11:30:28.445346 4824 generic.go:334] "Generic (PLEG): container finished" podID="aed87a15-74e4-47e3-8f90-f98672d62792" containerID="3d9bc7267f56122685da59d95824994c3ba510410d650c54f20f513e800a799f" exitCode=0 Jan 21 11:30:28 crc kubenswrapper[4824]: I0121 11:30:28.445386 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7d84b4d45c-l88p2" Jan 21 11:30:28 crc kubenswrapper[4824]: I0121 11:30:28.445383 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d84b4d45c-l88p2" event={"ID":"aed87a15-74e4-47e3-8f90-f98672d62792","Type":"ContainerDied","Data":"3d9bc7267f56122685da59d95824994c3ba510410d650c54f20f513e800a799f"} Jan 21 11:30:28 crc kubenswrapper[4824]: I0121 11:30:28.445657 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d84b4d45c-l88p2" event={"ID":"aed87a15-74e4-47e3-8f90-f98672d62792","Type":"ContainerDied","Data":"d3d7d6a8cb817cd1fffab43e3dc0bcf2917c60a0925857f2eded916d57a51538"} Jan 21 11:30:28 crc kubenswrapper[4824]: I0121 11:30:28.445682 4824 scope.go:117] "RemoveContainer" containerID="3d9bc7267f56122685da59d95824994c3ba510410d650c54f20f513e800a799f" Jan 21 11:30:28 crc kubenswrapper[4824]: I0121 11:30:28.464291 4824 scope.go:117] "RemoveContainer" containerID="2b3debae743097c64a642d1edc62ad587f2e44aabf5b01018238eb55be99a0de" Jan 21 11:30:28 crc kubenswrapper[4824]: I0121 11:30:28.480332 4824 scope.go:117] "RemoveContainer" containerID="3d9bc7267f56122685da59d95824994c3ba510410d650c54f20f513e800a799f" Jan 21 11:30:28 crc kubenswrapper[4824]: E0121 11:30:28.480619 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3d9bc7267f56122685da59d95824994c3ba510410d650c54f20f513e800a799f\": container with ID starting with 3d9bc7267f56122685da59d95824994c3ba510410d650c54f20f513e800a799f not found: ID does not exist" containerID="3d9bc7267f56122685da59d95824994c3ba510410d650c54f20f513e800a799f" Jan 21 11:30:28 crc kubenswrapper[4824]: I0121 11:30:28.480650 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3d9bc7267f56122685da59d95824994c3ba510410d650c54f20f513e800a799f"} err="failed to get container status \"3d9bc7267f56122685da59d95824994c3ba510410d650c54f20f513e800a799f\": rpc error: code = NotFound desc = could not find container \"3d9bc7267f56122685da59d95824994c3ba510410d650c54f20f513e800a799f\": container with ID starting with 3d9bc7267f56122685da59d95824994c3ba510410d650c54f20f513e800a799f not found: ID does not exist" Jan 21 11:30:28 crc kubenswrapper[4824]: I0121 11:30:28.480670 4824 scope.go:117] "RemoveContainer" containerID="2b3debae743097c64a642d1edc62ad587f2e44aabf5b01018238eb55be99a0de" Jan 21 11:30:28 crc kubenswrapper[4824]: E0121 11:30:28.480920 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2b3debae743097c64a642d1edc62ad587f2e44aabf5b01018238eb55be99a0de\": container with ID starting with 2b3debae743097c64a642d1edc62ad587f2e44aabf5b01018238eb55be99a0de not found: ID does not exist" containerID="2b3debae743097c64a642d1edc62ad587f2e44aabf5b01018238eb55be99a0de" Jan 21 11:30:28 crc kubenswrapper[4824]: I0121 11:30:28.480947 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2b3debae743097c64a642d1edc62ad587f2e44aabf5b01018238eb55be99a0de"} err="failed to get container status \"2b3debae743097c64a642d1edc62ad587f2e44aabf5b01018238eb55be99a0de\": rpc error: code = NotFound desc = could not find container \"2b3debae743097c64a642d1edc62ad587f2e44aabf5b01018238eb55be99a0de\": container with ID starting with 2b3debae743097c64a642d1edc62ad587f2e44aabf5b01018238eb55be99a0de not found: ID does not exist" Jan 21 11:30:28 crc kubenswrapper[4824]: I0121 11:30:28.560649 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aed87a15-74e4-47e3-8f90-f98672d62792-config\") pod \"aed87a15-74e4-47e3-8f90-f98672d62792\" (UID: \"aed87a15-74e4-47e3-8f90-f98672d62792\") " Jan 21 11:30:28 crc kubenswrapper[4824]: I0121 11:30:28.560728 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/aed87a15-74e4-47e3-8f90-f98672d62792-dns-svc\") pod \"aed87a15-74e4-47e3-8f90-f98672d62792\" (UID: \"aed87a15-74e4-47e3-8f90-f98672d62792\") " Jan 21 11:30:28 crc kubenswrapper[4824]: I0121 11:30:28.560784 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/aed87a15-74e4-47e3-8f90-f98672d62792-openstack-edpm-ipam\") pod \"aed87a15-74e4-47e3-8f90-f98672d62792\" (UID: \"aed87a15-74e4-47e3-8f90-f98672d62792\") " Jan 21 11:30:28 crc kubenswrapper[4824]: I0121 11:30:28.560843 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/aed87a15-74e4-47e3-8f90-f98672d62792-dns-swift-storage-0\") pod \"aed87a15-74e4-47e3-8f90-f98672d62792\" (UID: \"aed87a15-74e4-47e3-8f90-f98672d62792\") " Jan 21 11:30:28 crc kubenswrapper[4824]: I0121 11:30:28.560902 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/aed87a15-74e4-47e3-8f90-f98672d62792-ovsdbserver-nb\") pod \"aed87a15-74e4-47e3-8f90-f98672d62792\" (UID: \"aed87a15-74e4-47e3-8f90-f98672d62792\") " Jan 21 11:30:28 crc kubenswrapper[4824]: I0121 11:30:28.560987 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-txbf7\" (UniqueName: \"kubernetes.io/projected/aed87a15-74e4-47e3-8f90-f98672d62792-kube-api-access-txbf7\") pod \"aed87a15-74e4-47e3-8f90-f98672d62792\" (UID: \"aed87a15-74e4-47e3-8f90-f98672d62792\") " Jan 21 11:30:28 crc kubenswrapper[4824]: I0121 11:30:28.561069 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/aed87a15-74e4-47e3-8f90-f98672d62792-ovsdbserver-sb\") pod \"aed87a15-74e4-47e3-8f90-f98672d62792\" (UID: \"aed87a15-74e4-47e3-8f90-f98672d62792\") " Jan 21 11:30:28 crc kubenswrapper[4824]: I0121 11:30:28.564900 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aed87a15-74e4-47e3-8f90-f98672d62792-kube-api-access-txbf7" (OuterVolumeSpecName: "kube-api-access-txbf7") pod "aed87a15-74e4-47e3-8f90-f98672d62792" (UID: "aed87a15-74e4-47e3-8f90-f98672d62792"). InnerVolumeSpecName "kube-api-access-txbf7". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:30:28 crc kubenswrapper[4824]: I0121 11:30:28.602551 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aed87a15-74e4-47e3-8f90-f98672d62792-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "aed87a15-74e4-47e3-8f90-f98672d62792" (UID: "aed87a15-74e4-47e3-8f90-f98672d62792"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:30:28 crc kubenswrapper[4824]: I0121 11:30:28.615555 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aed87a15-74e4-47e3-8f90-f98672d62792-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "aed87a15-74e4-47e3-8f90-f98672d62792" (UID: "aed87a15-74e4-47e3-8f90-f98672d62792"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:30:28 crc kubenswrapper[4824]: I0121 11:30:28.615566 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aed87a15-74e4-47e3-8f90-f98672d62792-config" (OuterVolumeSpecName: "config") pod "aed87a15-74e4-47e3-8f90-f98672d62792" (UID: "aed87a15-74e4-47e3-8f90-f98672d62792"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:30:28 crc kubenswrapper[4824]: I0121 11:30:28.616142 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aed87a15-74e4-47e3-8f90-f98672d62792-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "aed87a15-74e4-47e3-8f90-f98672d62792" (UID: "aed87a15-74e4-47e3-8f90-f98672d62792"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:30:28 crc kubenswrapper[4824]: I0121 11:30:28.616158 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aed87a15-74e4-47e3-8f90-f98672d62792-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "aed87a15-74e4-47e3-8f90-f98672d62792" (UID: "aed87a15-74e4-47e3-8f90-f98672d62792"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:30:28 crc kubenswrapper[4824]: I0121 11:30:28.618461 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aed87a15-74e4-47e3-8f90-f98672d62792-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "aed87a15-74e4-47e3-8f90-f98672d62792" (UID: "aed87a15-74e4-47e3-8f90-f98672d62792"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:30:28 crc kubenswrapper[4824]: I0121 11:30:28.663521 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-txbf7\" (UniqueName: \"kubernetes.io/projected/aed87a15-74e4-47e3-8f90-f98672d62792-kube-api-access-txbf7\") on node \"crc\" DevicePath \"\"" Jan 21 11:30:28 crc kubenswrapper[4824]: I0121 11:30:28.663549 4824 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/aed87a15-74e4-47e3-8f90-f98672d62792-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Jan 21 11:30:28 crc kubenswrapper[4824]: I0121 11:30:28.663560 4824 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aed87a15-74e4-47e3-8f90-f98672d62792-config\") on node \"crc\" DevicePath \"\"" Jan 21 11:30:28 crc kubenswrapper[4824]: I0121 11:30:28.663591 4824 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/aed87a15-74e4-47e3-8f90-f98672d62792-dns-svc\") on node \"crc\" DevicePath \"\"" Jan 21 11:30:28 crc kubenswrapper[4824]: I0121 11:30:28.663600 4824 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/aed87a15-74e4-47e3-8f90-f98672d62792-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Jan 21 11:30:28 crc kubenswrapper[4824]: I0121 11:30:28.663609 4824 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/aed87a15-74e4-47e3-8f90-f98672d62792-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Jan 21 11:30:28 crc kubenswrapper[4824]: I0121 11:30:28.663616 4824 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/aed87a15-74e4-47e3-8f90-f98672d62792-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Jan 21 11:30:28 crc kubenswrapper[4824]: I0121 11:30:28.769325 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7d84b4d45c-l88p2"] Jan 21 11:30:28 crc kubenswrapper[4824]: I0121 11:30:28.776265 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7d84b4d45c-l88p2"] Jan 21 11:30:30 crc kubenswrapper[4824]: I0121 11:30:30.057375 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aed87a15-74e4-47e3-8f90-f98672d62792" path="/var/lib/kubelet/pods/aed87a15-74e4-47e3-8f90-f98672d62792/volumes" Jan 21 11:30:39 crc kubenswrapper[4824]: I0121 11:30:39.524967 4824 generic.go:334] "Generic (PLEG): container finished" podID="e4a4d725-1a49-4292-af74-6ab09ab15cdb" containerID="8915f56cc0770d64ee6a0d4ce2d2834b7ba05053d7b8c9baa030f6a50a88dee1" exitCode=0 Jan 21 11:30:39 crc kubenswrapper[4824]: I0121 11:30:39.524986 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"e4a4d725-1a49-4292-af74-6ab09ab15cdb","Type":"ContainerDied","Data":"8915f56cc0770d64ee6a0d4ce2d2834b7ba05053d7b8c9baa030f6a50a88dee1"} Jan 21 11:30:40 crc kubenswrapper[4824]: I0121 11:30:40.533785 4824 generic.go:334] "Generic (PLEG): container finished" podID="5d8e256a-aa1b-4ccb-8e8c-c86be1e31762" containerID="1ffd78c66a880f9e4de61c175a16fa8e25a34ed2dd520861ea49b68a2f4172fe" exitCode=0 Jan 21 11:30:40 crc kubenswrapper[4824]: I0121 11:30:40.533872 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"5d8e256a-aa1b-4ccb-8e8c-c86be1e31762","Type":"ContainerDied","Data":"1ffd78c66a880f9e4de61c175a16fa8e25a34ed2dd520861ea49b68a2f4172fe"} Jan 21 11:30:40 crc kubenswrapper[4824]: I0121 11:30:40.536832 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"e4a4d725-1a49-4292-af74-6ab09ab15cdb","Type":"ContainerStarted","Data":"43a3bc01f1993863308a904ac6733e32020664a934b3e525c85fa49c95638ce6"} Jan 21 11:30:40 crc kubenswrapper[4824]: I0121 11:30:40.537005 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Jan 21 11:30:40 crc kubenswrapper[4824]: I0121 11:30:40.576291 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=34.576277187 podStartE2EDuration="34.576277187s" podCreationTimestamp="2026-01-21 11:30:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:30:40.567911366 +0000 UTC m=+1182.860940658" watchObservedRunningTime="2026-01-21 11:30:40.576277187 +0000 UTC m=+1182.869306479" Jan 21 11:30:41 crc kubenswrapper[4824]: I0121 11:30:41.154100 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-2mbcf"] Jan 21 11:30:41 crc kubenswrapper[4824]: E0121 11:30:41.154662 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ad8dd9e-6fe4-462b-98a9-d0192072589d" containerName="init" Jan 21 11:30:41 crc kubenswrapper[4824]: I0121 11:30:41.154681 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ad8dd9e-6fe4-462b-98a9-d0192072589d" containerName="init" Jan 21 11:30:41 crc kubenswrapper[4824]: E0121 11:30:41.154696 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aed87a15-74e4-47e3-8f90-f98672d62792" containerName="dnsmasq-dns" Jan 21 11:30:41 crc kubenswrapper[4824]: I0121 11:30:41.154702 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="aed87a15-74e4-47e3-8f90-f98672d62792" containerName="dnsmasq-dns" Jan 21 11:30:41 crc kubenswrapper[4824]: E0121 11:30:41.154714 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aed87a15-74e4-47e3-8f90-f98672d62792" containerName="init" Jan 21 11:30:41 crc kubenswrapper[4824]: I0121 11:30:41.154719 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="aed87a15-74e4-47e3-8f90-f98672d62792" containerName="init" Jan 21 11:30:41 crc kubenswrapper[4824]: E0121 11:30:41.154749 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ad8dd9e-6fe4-462b-98a9-d0192072589d" containerName="dnsmasq-dns" Jan 21 11:30:41 crc kubenswrapper[4824]: I0121 11:30:41.154754 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ad8dd9e-6fe4-462b-98a9-d0192072589d" containerName="dnsmasq-dns" Jan 21 11:30:41 crc kubenswrapper[4824]: I0121 11:30:41.154930 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="aed87a15-74e4-47e3-8f90-f98672d62792" containerName="dnsmasq-dns" Jan 21 11:30:41 crc kubenswrapper[4824]: I0121 11:30:41.154950 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="8ad8dd9e-6fe4-462b-98a9-d0192072589d" containerName="dnsmasq-dns" Jan 21 11:30:41 crc kubenswrapper[4824]: I0121 11:30:41.155488 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-2mbcf" Jan 21 11:30:41 crc kubenswrapper[4824]: I0121 11:30:41.156843 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Jan 21 11:30:41 crc kubenswrapper[4824]: I0121 11:30:41.157396 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Jan 21 11:30:41 crc kubenswrapper[4824]: I0121 11:30:41.160795 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-tcqk6" Jan 21 11:30:41 crc kubenswrapper[4824]: I0121 11:30:41.160836 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Jan 21 11:30:41 crc kubenswrapper[4824]: I0121 11:30:41.160792 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-2mbcf"] Jan 21 11:30:41 crc kubenswrapper[4824]: I0121 11:30:41.267033 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4hwm7\" (UniqueName: \"kubernetes.io/projected/e3664d03-d53d-4a49-8cab-c4266f2d1426-kube-api-access-4hwm7\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-2mbcf\" (UID: \"e3664d03-d53d-4a49-8cab-c4266f2d1426\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-2mbcf" Jan 21 11:30:41 crc kubenswrapper[4824]: I0121 11:30:41.267259 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3664d03-d53d-4a49-8cab-c4266f2d1426-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-2mbcf\" (UID: \"e3664d03-d53d-4a49-8cab-c4266f2d1426\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-2mbcf" Jan 21 11:30:41 crc kubenswrapper[4824]: I0121 11:30:41.267422 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e3664d03-d53d-4a49-8cab-c4266f2d1426-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-2mbcf\" (UID: \"e3664d03-d53d-4a49-8cab-c4266f2d1426\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-2mbcf" Jan 21 11:30:41 crc kubenswrapper[4824]: I0121 11:30:41.267500 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/e3664d03-d53d-4a49-8cab-c4266f2d1426-ssh-key-openstack-edpm-ipam\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-2mbcf\" (UID: \"e3664d03-d53d-4a49-8cab-c4266f2d1426\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-2mbcf" Jan 21 11:30:41 crc kubenswrapper[4824]: I0121 11:30:41.368740 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e3664d03-d53d-4a49-8cab-c4266f2d1426-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-2mbcf\" (UID: \"e3664d03-d53d-4a49-8cab-c4266f2d1426\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-2mbcf" Jan 21 11:30:41 crc kubenswrapper[4824]: I0121 11:30:41.368933 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/e3664d03-d53d-4a49-8cab-c4266f2d1426-ssh-key-openstack-edpm-ipam\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-2mbcf\" (UID: \"e3664d03-d53d-4a49-8cab-c4266f2d1426\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-2mbcf" Jan 21 11:30:41 crc kubenswrapper[4824]: I0121 11:30:41.369207 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4hwm7\" (UniqueName: \"kubernetes.io/projected/e3664d03-d53d-4a49-8cab-c4266f2d1426-kube-api-access-4hwm7\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-2mbcf\" (UID: \"e3664d03-d53d-4a49-8cab-c4266f2d1426\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-2mbcf" Jan 21 11:30:41 crc kubenswrapper[4824]: I0121 11:30:41.369305 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3664d03-d53d-4a49-8cab-c4266f2d1426-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-2mbcf\" (UID: \"e3664d03-d53d-4a49-8cab-c4266f2d1426\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-2mbcf" Jan 21 11:30:41 crc kubenswrapper[4824]: I0121 11:30:41.372614 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/e3664d03-d53d-4a49-8cab-c4266f2d1426-ssh-key-openstack-edpm-ipam\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-2mbcf\" (UID: \"e3664d03-d53d-4a49-8cab-c4266f2d1426\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-2mbcf" Jan 21 11:30:41 crc kubenswrapper[4824]: I0121 11:30:41.373216 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e3664d03-d53d-4a49-8cab-c4266f2d1426-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-2mbcf\" (UID: \"e3664d03-d53d-4a49-8cab-c4266f2d1426\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-2mbcf" Jan 21 11:30:41 crc kubenswrapper[4824]: I0121 11:30:41.373490 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3664d03-d53d-4a49-8cab-c4266f2d1426-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-2mbcf\" (UID: \"e3664d03-d53d-4a49-8cab-c4266f2d1426\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-2mbcf" Jan 21 11:30:41 crc kubenswrapper[4824]: I0121 11:30:41.388764 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4hwm7\" (UniqueName: \"kubernetes.io/projected/e3664d03-d53d-4a49-8cab-c4266f2d1426-kube-api-access-4hwm7\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-2mbcf\" (UID: \"e3664d03-d53d-4a49-8cab-c4266f2d1426\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-2mbcf" Jan 21 11:30:41 crc kubenswrapper[4824]: I0121 11:30:41.470463 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-2mbcf" Jan 21 11:30:41 crc kubenswrapper[4824]: I0121 11:30:41.571285 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"5d8e256a-aa1b-4ccb-8e8c-c86be1e31762","Type":"ContainerStarted","Data":"bb6a9a152262299cc82912966110b0d79c9feaa705bae5ac5405357e09be0b2c"} Jan 21 11:30:41 crc kubenswrapper[4824]: I0121 11:30:41.571578 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Jan 21 11:30:41 crc kubenswrapper[4824]: I0121 11:30:41.589267 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=35.589252015 podStartE2EDuration="35.589252015s" podCreationTimestamp="2026-01-21 11:30:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 11:30:41.589046087 +0000 UTC m=+1183.882075379" watchObservedRunningTime="2026-01-21 11:30:41.589252015 +0000 UTC m=+1183.882281307" Jan 21 11:30:41 crc kubenswrapper[4824]: W0121 11:30:41.908095 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode3664d03_d53d_4a49_8cab_c4266f2d1426.slice/crio-a98155bffce9953a801b789919e7cde9fb3c2ba97aac7d2e8350794cd89e6960 WatchSource:0}: Error finding container a98155bffce9953a801b789919e7cde9fb3c2ba97aac7d2e8350794cd89e6960: Status 404 returned error can't find the container with id a98155bffce9953a801b789919e7cde9fb3c2ba97aac7d2e8350794cd89e6960 Jan 21 11:30:41 crc kubenswrapper[4824]: I0121 11:30:41.909586 4824 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 21 11:30:41 crc kubenswrapper[4824]: I0121 11:30:41.910333 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-2mbcf"] Jan 21 11:30:42 crc kubenswrapper[4824]: I0121 11:30:42.578213 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-2mbcf" event={"ID":"e3664d03-d53d-4a49-8cab-c4266f2d1426","Type":"ContainerStarted","Data":"a98155bffce9953a801b789919e7cde9fb3c2ba97aac7d2e8350794cd89e6960"} Jan 21 11:30:49 crc kubenswrapper[4824]: I0121 11:30:49.628798 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-2mbcf" event={"ID":"e3664d03-d53d-4a49-8cab-c4266f2d1426","Type":"ContainerStarted","Data":"26eb66e7a517666b02ef2ea7752b23c05bcd7e5afdc97bb9b0069008f7ff7fc5"} Jan 21 11:30:49 crc kubenswrapper[4824]: I0121 11:30:49.643265 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-2mbcf" podStartSLOduration=1.337568133 podStartE2EDuration="8.64325259s" podCreationTimestamp="2026-01-21 11:30:41 +0000 UTC" firstStartedPulling="2026-01-21 11:30:41.909384791 +0000 UTC m=+1184.202414083" lastFinishedPulling="2026-01-21 11:30:49.215069258 +0000 UTC m=+1191.508098540" observedRunningTime="2026-01-21 11:30:49.638516724 +0000 UTC m=+1191.931546015" watchObservedRunningTime="2026-01-21 11:30:49.64325259 +0000 UTC m=+1191.936281883" Jan 21 11:30:56 crc kubenswrapper[4824]: I0121 11:30:56.618093 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Jan 21 11:30:57 crc kubenswrapper[4824]: I0121 11:30:57.014144 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Jan 21 11:31:00 crc kubenswrapper[4824]: I0121 11:31:00.708514 4824 generic.go:334] "Generic (PLEG): container finished" podID="e3664d03-d53d-4a49-8cab-c4266f2d1426" containerID="26eb66e7a517666b02ef2ea7752b23c05bcd7e5afdc97bb9b0069008f7ff7fc5" exitCode=0 Jan 21 11:31:00 crc kubenswrapper[4824]: I0121 11:31:00.708594 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-2mbcf" event={"ID":"e3664d03-d53d-4a49-8cab-c4266f2d1426","Type":"ContainerDied","Data":"26eb66e7a517666b02ef2ea7752b23c05bcd7e5afdc97bb9b0069008f7ff7fc5"} Jan 21 11:31:02 crc kubenswrapper[4824]: I0121 11:31:02.018593 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-2mbcf" Jan 21 11:31:02 crc kubenswrapper[4824]: I0121 11:31:02.161847 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4hwm7\" (UniqueName: \"kubernetes.io/projected/e3664d03-d53d-4a49-8cab-c4266f2d1426-kube-api-access-4hwm7\") pod \"e3664d03-d53d-4a49-8cab-c4266f2d1426\" (UID: \"e3664d03-d53d-4a49-8cab-c4266f2d1426\") " Jan 21 11:31:02 crc kubenswrapper[4824]: I0121 11:31:02.161932 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3664d03-d53d-4a49-8cab-c4266f2d1426-repo-setup-combined-ca-bundle\") pod \"e3664d03-d53d-4a49-8cab-c4266f2d1426\" (UID: \"e3664d03-d53d-4a49-8cab-c4266f2d1426\") " Jan 21 11:31:02 crc kubenswrapper[4824]: I0121 11:31:02.162039 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/e3664d03-d53d-4a49-8cab-c4266f2d1426-ssh-key-openstack-edpm-ipam\") pod \"e3664d03-d53d-4a49-8cab-c4266f2d1426\" (UID: \"e3664d03-d53d-4a49-8cab-c4266f2d1426\") " Jan 21 11:31:02 crc kubenswrapper[4824]: I0121 11:31:02.162076 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e3664d03-d53d-4a49-8cab-c4266f2d1426-inventory\") pod \"e3664d03-d53d-4a49-8cab-c4266f2d1426\" (UID: \"e3664d03-d53d-4a49-8cab-c4266f2d1426\") " Jan 21 11:31:02 crc kubenswrapper[4824]: I0121 11:31:02.166405 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e3664d03-d53d-4a49-8cab-c4266f2d1426-kube-api-access-4hwm7" (OuterVolumeSpecName: "kube-api-access-4hwm7") pod "e3664d03-d53d-4a49-8cab-c4266f2d1426" (UID: "e3664d03-d53d-4a49-8cab-c4266f2d1426"). InnerVolumeSpecName "kube-api-access-4hwm7". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:31:02 crc kubenswrapper[4824]: I0121 11:31:02.166692 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e3664d03-d53d-4a49-8cab-c4266f2d1426-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "e3664d03-d53d-4a49-8cab-c4266f2d1426" (UID: "e3664d03-d53d-4a49-8cab-c4266f2d1426"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:31:02 crc kubenswrapper[4824]: I0121 11:31:02.183276 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e3664d03-d53d-4a49-8cab-c4266f2d1426-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "e3664d03-d53d-4a49-8cab-c4266f2d1426" (UID: "e3664d03-d53d-4a49-8cab-c4266f2d1426"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:31:02 crc kubenswrapper[4824]: I0121 11:31:02.183904 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e3664d03-d53d-4a49-8cab-c4266f2d1426-inventory" (OuterVolumeSpecName: "inventory") pod "e3664d03-d53d-4a49-8cab-c4266f2d1426" (UID: "e3664d03-d53d-4a49-8cab-c4266f2d1426"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:31:02 crc kubenswrapper[4824]: I0121 11:31:02.264462 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4hwm7\" (UniqueName: \"kubernetes.io/projected/e3664d03-d53d-4a49-8cab-c4266f2d1426-kube-api-access-4hwm7\") on node \"crc\" DevicePath \"\"" Jan 21 11:31:02 crc kubenswrapper[4824]: I0121 11:31:02.264646 4824 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3664d03-d53d-4a49-8cab-c4266f2d1426-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 21 11:31:02 crc kubenswrapper[4824]: I0121 11:31:02.264657 4824 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/e3664d03-d53d-4a49-8cab-c4266f2d1426-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Jan 21 11:31:02 crc kubenswrapper[4824]: I0121 11:31:02.264685 4824 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e3664d03-d53d-4a49-8cab-c4266f2d1426-inventory\") on node \"crc\" DevicePath \"\"" Jan 21 11:31:02 crc kubenswrapper[4824]: I0121 11:31:02.723162 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-2mbcf" event={"ID":"e3664d03-d53d-4a49-8cab-c4266f2d1426","Type":"ContainerDied","Data":"a98155bffce9953a801b789919e7cde9fb3c2ba97aac7d2e8350794cd89e6960"} Jan 21 11:31:02 crc kubenswrapper[4824]: I0121 11:31:02.723207 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a98155bffce9953a801b789919e7cde9fb3c2ba97aac7d2e8350794cd89e6960" Jan 21 11:31:02 crc kubenswrapper[4824]: I0121 11:31:02.723287 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-2mbcf" Jan 21 11:31:02 crc kubenswrapper[4824]: I0121 11:31:02.774923 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-2flzn"] Jan 21 11:31:02 crc kubenswrapper[4824]: E0121 11:31:02.775237 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3664d03-d53d-4a49-8cab-c4266f2d1426" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Jan 21 11:31:02 crc kubenswrapper[4824]: I0121 11:31:02.775254 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3664d03-d53d-4a49-8cab-c4266f2d1426" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Jan 21 11:31:02 crc kubenswrapper[4824]: I0121 11:31:02.775417 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="e3664d03-d53d-4a49-8cab-c4266f2d1426" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Jan 21 11:31:02 crc kubenswrapper[4824]: I0121 11:31:02.775879 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-2flzn" Jan 21 11:31:02 crc kubenswrapper[4824]: I0121 11:31:02.785111 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Jan 21 11:31:02 crc kubenswrapper[4824]: I0121 11:31:02.785233 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-tcqk6" Jan 21 11:31:02 crc kubenswrapper[4824]: I0121 11:31:02.785305 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Jan 21 11:31:02 crc kubenswrapper[4824]: I0121 11:31:02.785434 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Jan 21 11:31:02 crc kubenswrapper[4824]: I0121 11:31:02.793504 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-2flzn"] Jan 21 11:31:02 crc kubenswrapper[4824]: I0121 11:31:02.973905 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/34c68ecf-cf29-44cb-9880-a3d45f625454-ssh-key-openstack-edpm-ipam\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-2flzn\" (UID: \"34c68ecf-cf29-44cb-9880-a3d45f625454\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-2flzn" Jan 21 11:31:02 crc kubenswrapper[4824]: I0121 11:31:02.974041 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/34c68ecf-cf29-44cb-9880-a3d45f625454-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-2flzn\" (UID: \"34c68ecf-cf29-44cb-9880-a3d45f625454\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-2flzn" Jan 21 11:31:02 crc kubenswrapper[4824]: I0121 11:31:02.974266 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2n7ls\" (UniqueName: \"kubernetes.io/projected/34c68ecf-cf29-44cb-9880-a3d45f625454-kube-api-access-2n7ls\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-2flzn\" (UID: \"34c68ecf-cf29-44cb-9880-a3d45f625454\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-2flzn" Jan 21 11:31:03 crc kubenswrapper[4824]: I0121 11:31:03.075217 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2n7ls\" (UniqueName: \"kubernetes.io/projected/34c68ecf-cf29-44cb-9880-a3d45f625454-kube-api-access-2n7ls\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-2flzn\" (UID: \"34c68ecf-cf29-44cb-9880-a3d45f625454\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-2flzn" Jan 21 11:31:03 crc kubenswrapper[4824]: I0121 11:31:03.075901 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/34c68ecf-cf29-44cb-9880-a3d45f625454-ssh-key-openstack-edpm-ipam\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-2flzn\" (UID: \"34c68ecf-cf29-44cb-9880-a3d45f625454\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-2flzn" Jan 21 11:31:03 crc kubenswrapper[4824]: I0121 11:31:03.076003 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/34c68ecf-cf29-44cb-9880-a3d45f625454-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-2flzn\" (UID: \"34c68ecf-cf29-44cb-9880-a3d45f625454\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-2flzn" Jan 21 11:31:03 crc kubenswrapper[4824]: I0121 11:31:03.079586 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/34c68ecf-cf29-44cb-9880-a3d45f625454-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-2flzn\" (UID: \"34c68ecf-cf29-44cb-9880-a3d45f625454\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-2flzn" Jan 21 11:31:03 crc kubenswrapper[4824]: I0121 11:31:03.079602 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/34c68ecf-cf29-44cb-9880-a3d45f625454-ssh-key-openstack-edpm-ipam\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-2flzn\" (UID: \"34c68ecf-cf29-44cb-9880-a3d45f625454\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-2flzn" Jan 21 11:31:03 crc kubenswrapper[4824]: I0121 11:31:03.088457 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2n7ls\" (UniqueName: \"kubernetes.io/projected/34c68ecf-cf29-44cb-9880-a3d45f625454-kube-api-access-2n7ls\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-2flzn\" (UID: \"34c68ecf-cf29-44cb-9880-a3d45f625454\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-2flzn" Jan 21 11:31:03 crc kubenswrapper[4824]: I0121 11:31:03.096376 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-2flzn" Jan 21 11:31:03 crc kubenswrapper[4824]: I0121 11:31:03.523746 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-2flzn"] Jan 21 11:31:03 crc kubenswrapper[4824]: W0121 11:31:03.525472 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod34c68ecf_cf29_44cb_9880_a3d45f625454.slice/crio-e9f9e50d04ca524befd4d31f5b8e591287db6afe37e3990aba6d3abeb81603fc WatchSource:0}: Error finding container e9f9e50d04ca524befd4d31f5b8e591287db6afe37e3990aba6d3abeb81603fc: Status 404 returned error can't find the container with id e9f9e50d04ca524befd4d31f5b8e591287db6afe37e3990aba6d3abeb81603fc Jan 21 11:31:03 crc kubenswrapper[4824]: I0121 11:31:03.730103 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-2flzn" event={"ID":"34c68ecf-cf29-44cb-9880-a3d45f625454","Type":"ContainerStarted","Data":"e9f9e50d04ca524befd4d31f5b8e591287db6afe37e3990aba6d3abeb81603fc"} Jan 21 11:31:04 crc kubenswrapper[4824]: I0121 11:31:04.737700 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-2flzn" event={"ID":"34c68ecf-cf29-44cb-9880-a3d45f625454","Type":"ContainerStarted","Data":"2a39f90f288e86ab7f527b5bc67bfc6b30d16751b1be1f809ef23d416cfc707c"} Jan 21 11:31:04 crc kubenswrapper[4824]: I0121 11:31:04.752320 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-2flzn" podStartSLOduration=2.231292759 podStartE2EDuration="2.752305525s" podCreationTimestamp="2026-01-21 11:31:02 +0000 UTC" firstStartedPulling="2026-01-21 11:31:03.527609308 +0000 UTC m=+1205.820638600" lastFinishedPulling="2026-01-21 11:31:04.048622075 +0000 UTC m=+1206.341651366" observedRunningTime="2026-01-21 11:31:04.748108754 +0000 UTC m=+1207.041138046" watchObservedRunningTime="2026-01-21 11:31:04.752305525 +0000 UTC m=+1207.045334817" Jan 21 11:31:06 crc kubenswrapper[4824]: I0121 11:31:06.752212 4824 generic.go:334] "Generic (PLEG): container finished" podID="34c68ecf-cf29-44cb-9880-a3d45f625454" containerID="2a39f90f288e86ab7f527b5bc67bfc6b30d16751b1be1f809ef23d416cfc707c" exitCode=0 Jan 21 11:31:06 crc kubenswrapper[4824]: I0121 11:31:06.752302 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-2flzn" event={"ID":"34c68ecf-cf29-44cb-9880-a3d45f625454","Type":"ContainerDied","Data":"2a39f90f288e86ab7f527b5bc67bfc6b30d16751b1be1f809ef23d416cfc707c"} Jan 21 11:31:08 crc kubenswrapper[4824]: I0121 11:31:08.063258 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-2flzn" Jan 21 11:31:08 crc kubenswrapper[4824]: I0121 11:31:08.257276 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/34c68ecf-cf29-44cb-9880-a3d45f625454-inventory\") pod \"34c68ecf-cf29-44cb-9880-a3d45f625454\" (UID: \"34c68ecf-cf29-44cb-9880-a3d45f625454\") " Jan 21 11:31:08 crc kubenswrapper[4824]: I0121 11:31:08.257433 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/34c68ecf-cf29-44cb-9880-a3d45f625454-ssh-key-openstack-edpm-ipam\") pod \"34c68ecf-cf29-44cb-9880-a3d45f625454\" (UID: \"34c68ecf-cf29-44cb-9880-a3d45f625454\") " Jan 21 11:31:08 crc kubenswrapper[4824]: I0121 11:31:08.257503 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2n7ls\" (UniqueName: \"kubernetes.io/projected/34c68ecf-cf29-44cb-9880-a3d45f625454-kube-api-access-2n7ls\") pod \"34c68ecf-cf29-44cb-9880-a3d45f625454\" (UID: \"34c68ecf-cf29-44cb-9880-a3d45f625454\") " Jan 21 11:31:08 crc kubenswrapper[4824]: I0121 11:31:08.261552 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/34c68ecf-cf29-44cb-9880-a3d45f625454-kube-api-access-2n7ls" (OuterVolumeSpecName: "kube-api-access-2n7ls") pod "34c68ecf-cf29-44cb-9880-a3d45f625454" (UID: "34c68ecf-cf29-44cb-9880-a3d45f625454"). InnerVolumeSpecName "kube-api-access-2n7ls". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:31:08 crc kubenswrapper[4824]: I0121 11:31:08.277424 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/34c68ecf-cf29-44cb-9880-a3d45f625454-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "34c68ecf-cf29-44cb-9880-a3d45f625454" (UID: "34c68ecf-cf29-44cb-9880-a3d45f625454"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:31:08 crc kubenswrapper[4824]: I0121 11:31:08.280832 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/34c68ecf-cf29-44cb-9880-a3d45f625454-inventory" (OuterVolumeSpecName: "inventory") pod "34c68ecf-cf29-44cb-9880-a3d45f625454" (UID: "34c68ecf-cf29-44cb-9880-a3d45f625454"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:31:08 crc kubenswrapper[4824]: I0121 11:31:08.359619 4824 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/34c68ecf-cf29-44cb-9880-a3d45f625454-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Jan 21 11:31:08 crc kubenswrapper[4824]: I0121 11:31:08.359649 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2n7ls\" (UniqueName: \"kubernetes.io/projected/34c68ecf-cf29-44cb-9880-a3d45f625454-kube-api-access-2n7ls\") on node \"crc\" DevicePath \"\"" Jan 21 11:31:08 crc kubenswrapper[4824]: I0121 11:31:08.359660 4824 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/34c68ecf-cf29-44cb-9880-a3d45f625454-inventory\") on node \"crc\" DevicePath \"\"" Jan 21 11:31:08 crc kubenswrapper[4824]: I0121 11:31:08.768542 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-2flzn" event={"ID":"34c68ecf-cf29-44cb-9880-a3d45f625454","Type":"ContainerDied","Data":"e9f9e50d04ca524befd4d31f5b8e591287db6afe37e3990aba6d3abeb81603fc"} Jan 21 11:31:08 crc kubenswrapper[4824]: I0121 11:31:08.768745 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e9f9e50d04ca524befd4d31f5b8e591287db6afe37e3990aba6d3abeb81603fc" Jan 21 11:31:08 crc kubenswrapper[4824]: I0121 11:31:08.768596 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-2flzn" Jan 21 11:31:08 crc kubenswrapper[4824]: I0121 11:31:08.808417 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-9z7z6"] Jan 21 11:31:08 crc kubenswrapper[4824]: E0121 11:31:08.808788 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34c68ecf-cf29-44cb-9880-a3d45f625454" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Jan 21 11:31:08 crc kubenswrapper[4824]: I0121 11:31:08.808806 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="34c68ecf-cf29-44cb-9880-a3d45f625454" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Jan 21 11:31:08 crc kubenswrapper[4824]: I0121 11:31:08.809011 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="34c68ecf-cf29-44cb-9880-a3d45f625454" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Jan 21 11:31:08 crc kubenswrapper[4824]: I0121 11:31:08.809605 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-9z7z6" Jan 21 11:31:08 crc kubenswrapper[4824]: I0121 11:31:08.815614 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Jan 21 11:31:08 crc kubenswrapper[4824]: I0121 11:31:08.815669 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-tcqk6" Jan 21 11:31:08 crc kubenswrapper[4824]: I0121 11:31:08.815827 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Jan 21 11:31:08 crc kubenswrapper[4824]: I0121 11:31:08.815852 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Jan 21 11:31:08 crc kubenswrapper[4824]: I0121 11:31:08.816182 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-9z7z6"] Jan 21 11:31:08 crc kubenswrapper[4824]: I0121 11:31:08.866665 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/300a037e-ac25-4a9e-9339-6cc707969f7c-ssh-key-openstack-edpm-ipam\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-9z7z6\" (UID: \"300a037e-ac25-4a9e-9339-6cc707969f7c\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-9z7z6" Jan 21 11:31:08 crc kubenswrapper[4824]: I0121 11:31:08.866854 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/300a037e-ac25-4a9e-9339-6cc707969f7c-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-9z7z6\" (UID: \"300a037e-ac25-4a9e-9339-6cc707969f7c\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-9z7z6" Jan 21 11:31:08 crc kubenswrapper[4824]: I0121 11:31:08.867021 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-srdjg\" (UniqueName: \"kubernetes.io/projected/300a037e-ac25-4a9e-9339-6cc707969f7c-kube-api-access-srdjg\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-9z7z6\" (UID: \"300a037e-ac25-4a9e-9339-6cc707969f7c\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-9z7z6" Jan 21 11:31:08 crc kubenswrapper[4824]: I0121 11:31:08.867063 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/300a037e-ac25-4a9e-9339-6cc707969f7c-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-9z7z6\" (UID: \"300a037e-ac25-4a9e-9339-6cc707969f7c\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-9z7z6" Jan 21 11:31:08 crc kubenswrapper[4824]: I0121 11:31:08.969532 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-srdjg\" (UniqueName: \"kubernetes.io/projected/300a037e-ac25-4a9e-9339-6cc707969f7c-kube-api-access-srdjg\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-9z7z6\" (UID: \"300a037e-ac25-4a9e-9339-6cc707969f7c\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-9z7z6" Jan 21 11:31:08 crc kubenswrapper[4824]: I0121 11:31:08.969594 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/300a037e-ac25-4a9e-9339-6cc707969f7c-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-9z7z6\" (UID: \"300a037e-ac25-4a9e-9339-6cc707969f7c\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-9z7z6" Jan 21 11:31:08 crc kubenswrapper[4824]: I0121 11:31:08.969731 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/300a037e-ac25-4a9e-9339-6cc707969f7c-ssh-key-openstack-edpm-ipam\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-9z7z6\" (UID: \"300a037e-ac25-4a9e-9339-6cc707969f7c\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-9z7z6" Jan 21 11:31:08 crc kubenswrapper[4824]: I0121 11:31:08.969889 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/300a037e-ac25-4a9e-9339-6cc707969f7c-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-9z7z6\" (UID: \"300a037e-ac25-4a9e-9339-6cc707969f7c\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-9z7z6" Jan 21 11:31:08 crc kubenswrapper[4824]: I0121 11:31:08.974025 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/300a037e-ac25-4a9e-9339-6cc707969f7c-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-9z7z6\" (UID: \"300a037e-ac25-4a9e-9339-6cc707969f7c\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-9z7z6" Jan 21 11:31:08 crc kubenswrapper[4824]: I0121 11:31:08.979756 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/300a037e-ac25-4a9e-9339-6cc707969f7c-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-9z7z6\" (UID: \"300a037e-ac25-4a9e-9339-6cc707969f7c\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-9z7z6" Jan 21 11:31:08 crc kubenswrapper[4824]: I0121 11:31:08.982876 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/300a037e-ac25-4a9e-9339-6cc707969f7c-ssh-key-openstack-edpm-ipam\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-9z7z6\" (UID: \"300a037e-ac25-4a9e-9339-6cc707969f7c\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-9z7z6" Jan 21 11:31:08 crc kubenswrapper[4824]: I0121 11:31:08.987061 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-srdjg\" (UniqueName: \"kubernetes.io/projected/300a037e-ac25-4a9e-9339-6cc707969f7c-kube-api-access-srdjg\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-9z7z6\" (UID: \"300a037e-ac25-4a9e-9339-6cc707969f7c\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-9z7z6" Jan 21 11:31:09 crc kubenswrapper[4824]: I0121 11:31:09.123908 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-9z7z6" Jan 21 11:31:09 crc kubenswrapper[4824]: I0121 11:31:09.619332 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-9z7z6"] Jan 21 11:31:09 crc kubenswrapper[4824]: I0121 11:31:09.776695 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-9z7z6" event={"ID":"300a037e-ac25-4a9e-9339-6cc707969f7c","Type":"ContainerStarted","Data":"3606afc5a84332a98e06b75db8f8ca12eba78a06f6dfd19a1acd5dec3cb653f4"} Jan 21 11:31:10 crc kubenswrapper[4824]: I0121 11:31:10.783922 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-9z7z6" event={"ID":"300a037e-ac25-4a9e-9339-6cc707969f7c","Type":"ContainerStarted","Data":"36e8b1c82d028e5d8507608d6152420eb1389504eb7eb596bd23d52fdf2184e2"} Jan 21 11:31:10 crc kubenswrapper[4824]: I0121 11:31:10.796287 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-9z7z6" podStartSLOduration=2.170215529 podStartE2EDuration="2.796273247s" podCreationTimestamp="2026-01-21 11:31:08 +0000 UTC" firstStartedPulling="2026-01-21 11:31:09.621684068 +0000 UTC m=+1211.914713360" lastFinishedPulling="2026-01-21 11:31:10.247741786 +0000 UTC m=+1212.540771078" observedRunningTime="2026-01-21 11:31:10.793689277 +0000 UTC m=+1213.086718589" watchObservedRunningTime="2026-01-21 11:31:10.796273247 +0000 UTC m=+1213.089302539" Jan 21 11:31:58 crc kubenswrapper[4824]: I0121 11:31:58.939313 4824 scope.go:117] "RemoveContainer" containerID="26ba4c23450a1625fb62c54cdc290126e5f1371f38bf091a20d00cb1c946e58f" Jan 21 11:32:16 crc kubenswrapper[4824]: I0121 11:32:16.065104 4824 patch_prober.go:28] interesting pod/machine-config-daemon-zqjg9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 21 11:32:16 crc kubenswrapper[4824]: I0121 11:32:16.065629 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 21 11:32:46 crc kubenswrapper[4824]: I0121 11:32:46.064777 4824 patch_prober.go:28] interesting pod/machine-config-daemon-zqjg9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 21 11:32:46 crc kubenswrapper[4824]: I0121 11:32:46.065308 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 21 11:32:58 crc kubenswrapper[4824]: I0121 11:32:58.985276 4824 scope.go:117] "RemoveContainer" containerID="7802d3564cb8577bdb2e0cb7acab4d7e0721818cb3c2ece34e851778aff1034e" Jan 21 11:32:59 crc kubenswrapper[4824]: I0121 11:32:59.004242 4824 scope.go:117] "RemoveContainer" containerID="eb6d7862490b4afbdbbe422c8984287dc94d6e4a39261b3846440fb920c542fb" Jan 21 11:32:59 crc kubenswrapper[4824]: I0121 11:32:59.036218 4824 scope.go:117] "RemoveContainer" containerID="68285af994e4fe82e817010ea6bff330d5655f091788a58ca5b521d0e79564f1" Jan 21 11:32:59 crc kubenswrapper[4824]: I0121 11:32:59.067153 4824 scope.go:117] "RemoveContainer" containerID="ff804ebe93733fa9e697132922f8e30c6ee75d058b7cb8afecec7da791271c35" Jan 21 11:33:16 crc kubenswrapper[4824]: I0121 11:33:16.064969 4824 patch_prober.go:28] interesting pod/machine-config-daemon-zqjg9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 21 11:33:16 crc kubenswrapper[4824]: I0121 11:33:16.065476 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 21 11:33:16 crc kubenswrapper[4824]: I0121 11:33:16.065514 4824 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" Jan 21 11:33:16 crc kubenswrapper[4824]: I0121 11:33:16.066048 4824 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f5273806c339b109c1d902705cf82c97535d3adc4a021ab1d42339f22a956af4"} pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 21 11:33:16 crc kubenswrapper[4824]: I0121 11:33:16.066100 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" containerName="machine-config-daemon" containerID="cri-o://f5273806c339b109c1d902705cf82c97535d3adc4a021ab1d42339f22a956af4" gracePeriod=600 Jan 21 11:33:16 crc kubenswrapper[4824]: I0121 11:33:16.616564 4824 generic.go:334] "Generic (PLEG): container finished" podID="33f3d922-4ffe-409b-a49a-d88c85898260" containerID="f5273806c339b109c1d902705cf82c97535d3adc4a021ab1d42339f22a956af4" exitCode=0 Jan 21 11:33:16 crc kubenswrapper[4824]: I0121 11:33:16.616616 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" event={"ID":"33f3d922-4ffe-409b-a49a-d88c85898260","Type":"ContainerDied","Data":"f5273806c339b109c1d902705cf82c97535d3adc4a021ab1d42339f22a956af4"} Jan 21 11:33:16 crc kubenswrapper[4824]: I0121 11:33:16.617035 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" event={"ID":"33f3d922-4ffe-409b-a49a-d88c85898260","Type":"ContainerStarted","Data":"1cd835fe36ebb6541740e5ab5fd843470d46dfb78975fd968e5cc9aaef462499"} Jan 21 11:33:16 crc kubenswrapper[4824]: I0121 11:33:16.617059 4824 scope.go:117] "RemoveContainer" containerID="74ad8e1889db466232cc3528a7e51d12f4be65b09faed3c15716a73518537690" Jan 21 11:34:08 crc kubenswrapper[4824]: I0121 11:34:08.971176 4824 generic.go:334] "Generic (PLEG): container finished" podID="300a037e-ac25-4a9e-9339-6cc707969f7c" containerID="36e8b1c82d028e5d8507608d6152420eb1389504eb7eb596bd23d52fdf2184e2" exitCode=0 Jan 21 11:34:08 crc kubenswrapper[4824]: I0121 11:34:08.971248 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-9z7z6" event={"ID":"300a037e-ac25-4a9e-9339-6cc707969f7c","Type":"ContainerDied","Data":"36e8b1c82d028e5d8507608d6152420eb1389504eb7eb596bd23d52fdf2184e2"} Jan 21 11:34:10 crc kubenswrapper[4824]: I0121 11:34:10.278164 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-9z7z6" Jan 21 11:34:10 crc kubenswrapper[4824]: I0121 11:34:10.303178 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/300a037e-ac25-4a9e-9339-6cc707969f7c-inventory\") pod \"300a037e-ac25-4a9e-9339-6cc707969f7c\" (UID: \"300a037e-ac25-4a9e-9339-6cc707969f7c\") " Jan 21 11:34:10 crc kubenswrapper[4824]: I0121 11:34:10.303300 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/300a037e-ac25-4a9e-9339-6cc707969f7c-ssh-key-openstack-edpm-ipam\") pod \"300a037e-ac25-4a9e-9339-6cc707969f7c\" (UID: \"300a037e-ac25-4a9e-9339-6cc707969f7c\") " Jan 21 11:34:10 crc kubenswrapper[4824]: I0121 11:34:10.303360 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/300a037e-ac25-4a9e-9339-6cc707969f7c-bootstrap-combined-ca-bundle\") pod \"300a037e-ac25-4a9e-9339-6cc707969f7c\" (UID: \"300a037e-ac25-4a9e-9339-6cc707969f7c\") " Jan 21 11:34:10 crc kubenswrapper[4824]: I0121 11:34:10.303471 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-srdjg\" (UniqueName: \"kubernetes.io/projected/300a037e-ac25-4a9e-9339-6cc707969f7c-kube-api-access-srdjg\") pod \"300a037e-ac25-4a9e-9339-6cc707969f7c\" (UID: \"300a037e-ac25-4a9e-9339-6cc707969f7c\") " Jan 21 11:34:10 crc kubenswrapper[4824]: I0121 11:34:10.308769 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/300a037e-ac25-4a9e-9339-6cc707969f7c-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "300a037e-ac25-4a9e-9339-6cc707969f7c" (UID: "300a037e-ac25-4a9e-9339-6cc707969f7c"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:34:10 crc kubenswrapper[4824]: I0121 11:34:10.308864 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/300a037e-ac25-4a9e-9339-6cc707969f7c-kube-api-access-srdjg" (OuterVolumeSpecName: "kube-api-access-srdjg") pod "300a037e-ac25-4a9e-9339-6cc707969f7c" (UID: "300a037e-ac25-4a9e-9339-6cc707969f7c"). InnerVolumeSpecName "kube-api-access-srdjg". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:34:10 crc kubenswrapper[4824]: I0121 11:34:10.325497 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/300a037e-ac25-4a9e-9339-6cc707969f7c-inventory" (OuterVolumeSpecName: "inventory") pod "300a037e-ac25-4a9e-9339-6cc707969f7c" (UID: "300a037e-ac25-4a9e-9339-6cc707969f7c"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:34:10 crc kubenswrapper[4824]: I0121 11:34:10.327398 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/300a037e-ac25-4a9e-9339-6cc707969f7c-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "300a037e-ac25-4a9e-9339-6cc707969f7c" (UID: "300a037e-ac25-4a9e-9339-6cc707969f7c"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:34:10 crc kubenswrapper[4824]: I0121 11:34:10.405519 4824 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/300a037e-ac25-4a9e-9339-6cc707969f7c-inventory\") on node \"crc\" DevicePath \"\"" Jan 21 11:34:10 crc kubenswrapper[4824]: I0121 11:34:10.405641 4824 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/300a037e-ac25-4a9e-9339-6cc707969f7c-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Jan 21 11:34:10 crc kubenswrapper[4824]: I0121 11:34:10.405700 4824 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/300a037e-ac25-4a9e-9339-6cc707969f7c-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 21 11:34:10 crc kubenswrapper[4824]: I0121 11:34:10.405766 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-srdjg\" (UniqueName: \"kubernetes.io/projected/300a037e-ac25-4a9e-9339-6cc707969f7c-kube-api-access-srdjg\") on node \"crc\" DevicePath \"\"" Jan 21 11:34:10 crc kubenswrapper[4824]: I0121 11:34:10.986479 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-9z7z6" event={"ID":"300a037e-ac25-4a9e-9339-6cc707969f7c","Type":"ContainerDied","Data":"3606afc5a84332a98e06b75db8f8ca12eba78a06f6dfd19a1acd5dec3cb653f4"} Jan 21 11:34:10 crc kubenswrapper[4824]: I0121 11:34:10.986867 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3606afc5a84332a98e06b75db8f8ca12eba78a06f6dfd19a1acd5dec3cb653f4" Jan 21 11:34:10 crc kubenswrapper[4824]: I0121 11:34:10.986524 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-9z7z6" Jan 21 11:34:11 crc kubenswrapper[4824]: I0121 11:34:11.047720 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-dpgt9"] Jan 21 11:34:11 crc kubenswrapper[4824]: E0121 11:34:11.048129 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="300a037e-ac25-4a9e-9339-6cc707969f7c" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Jan 21 11:34:11 crc kubenswrapper[4824]: I0121 11:34:11.048149 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="300a037e-ac25-4a9e-9339-6cc707969f7c" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Jan 21 11:34:11 crc kubenswrapper[4824]: I0121 11:34:11.048337 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="300a037e-ac25-4a9e-9339-6cc707969f7c" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Jan 21 11:34:11 crc kubenswrapper[4824]: I0121 11:34:11.049075 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-dpgt9" Jan 21 11:34:11 crc kubenswrapper[4824]: I0121 11:34:11.051351 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-tcqk6" Jan 21 11:34:11 crc kubenswrapper[4824]: I0121 11:34:11.051481 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Jan 21 11:34:11 crc kubenswrapper[4824]: I0121 11:34:11.051605 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Jan 21 11:34:11 crc kubenswrapper[4824]: I0121 11:34:11.052370 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Jan 21 11:34:11 crc kubenswrapper[4824]: I0121 11:34:11.055545 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-dpgt9"] Jan 21 11:34:11 crc kubenswrapper[4824]: I0121 11:34:11.119716 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/8dbaf72e-72a7-4073-a4c1-95cbdabaa64b-ssh-key-openstack-edpm-ipam\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-dpgt9\" (UID: \"8dbaf72e-72a7-4073-a4c1-95cbdabaa64b\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-dpgt9" Jan 21 11:34:11 crc kubenswrapper[4824]: I0121 11:34:11.120393 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v5hlg\" (UniqueName: \"kubernetes.io/projected/8dbaf72e-72a7-4073-a4c1-95cbdabaa64b-kube-api-access-v5hlg\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-dpgt9\" (UID: \"8dbaf72e-72a7-4073-a4c1-95cbdabaa64b\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-dpgt9" Jan 21 11:34:11 crc kubenswrapper[4824]: I0121 11:34:11.120483 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8dbaf72e-72a7-4073-a4c1-95cbdabaa64b-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-dpgt9\" (UID: \"8dbaf72e-72a7-4073-a4c1-95cbdabaa64b\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-dpgt9" Jan 21 11:34:11 crc kubenswrapper[4824]: I0121 11:34:11.222491 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/8dbaf72e-72a7-4073-a4c1-95cbdabaa64b-ssh-key-openstack-edpm-ipam\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-dpgt9\" (UID: \"8dbaf72e-72a7-4073-a4c1-95cbdabaa64b\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-dpgt9" Jan 21 11:34:11 crc kubenswrapper[4824]: I0121 11:34:11.222696 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v5hlg\" (UniqueName: \"kubernetes.io/projected/8dbaf72e-72a7-4073-a4c1-95cbdabaa64b-kube-api-access-v5hlg\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-dpgt9\" (UID: \"8dbaf72e-72a7-4073-a4c1-95cbdabaa64b\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-dpgt9" Jan 21 11:34:11 crc kubenswrapper[4824]: I0121 11:34:11.222765 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8dbaf72e-72a7-4073-a4c1-95cbdabaa64b-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-dpgt9\" (UID: \"8dbaf72e-72a7-4073-a4c1-95cbdabaa64b\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-dpgt9" Jan 21 11:34:11 crc kubenswrapper[4824]: I0121 11:34:11.225662 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8dbaf72e-72a7-4073-a4c1-95cbdabaa64b-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-dpgt9\" (UID: \"8dbaf72e-72a7-4073-a4c1-95cbdabaa64b\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-dpgt9" Jan 21 11:34:11 crc kubenswrapper[4824]: I0121 11:34:11.225771 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/8dbaf72e-72a7-4073-a4c1-95cbdabaa64b-ssh-key-openstack-edpm-ipam\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-dpgt9\" (UID: \"8dbaf72e-72a7-4073-a4c1-95cbdabaa64b\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-dpgt9" Jan 21 11:34:11 crc kubenswrapper[4824]: I0121 11:34:11.235755 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v5hlg\" (UniqueName: \"kubernetes.io/projected/8dbaf72e-72a7-4073-a4c1-95cbdabaa64b-kube-api-access-v5hlg\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-dpgt9\" (UID: \"8dbaf72e-72a7-4073-a4c1-95cbdabaa64b\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-dpgt9" Jan 21 11:34:11 crc kubenswrapper[4824]: I0121 11:34:11.381369 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-dpgt9" Jan 21 11:34:11 crc kubenswrapper[4824]: I0121 11:34:11.797190 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-dpgt9"] Jan 21 11:34:11 crc kubenswrapper[4824]: I0121 11:34:11.993368 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-dpgt9" event={"ID":"8dbaf72e-72a7-4073-a4c1-95cbdabaa64b","Type":"ContainerStarted","Data":"b5d574baa9ddaffa59e308d4f888821c6b55c1c17bdef6ce2814d173fcd2ffdc"} Jan 21 11:34:13 crc kubenswrapper[4824]: I0121 11:34:13.003746 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-dpgt9" event={"ID":"8dbaf72e-72a7-4073-a4c1-95cbdabaa64b","Type":"ContainerStarted","Data":"4d04c081499dd9f0ea8ea5831aa73fabb1773d706814fc10d98f8f23c9450fd3"} Jan 21 11:34:13 crc kubenswrapper[4824]: I0121 11:34:13.022512 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-dpgt9" podStartSLOduration=1.552950284 podStartE2EDuration="2.022497125s" podCreationTimestamp="2026-01-21 11:34:11 +0000 UTC" firstStartedPulling="2026-01-21 11:34:11.803496443 +0000 UTC m=+1394.096525736" lastFinishedPulling="2026-01-21 11:34:12.273043285 +0000 UTC m=+1394.566072577" observedRunningTime="2026-01-21 11:34:13.019043938 +0000 UTC m=+1395.312073229" watchObservedRunningTime="2026-01-21 11:34:13.022497125 +0000 UTC m=+1395.315526417" Jan 21 11:34:59 crc kubenswrapper[4824]: I0121 11:34:59.141123 4824 scope.go:117] "RemoveContainer" containerID="3f015c1fd2be4ed1835afc4169552ab51d97958ba2fb966dc794c0c63861420b" Jan 21 11:34:59 crc kubenswrapper[4824]: I0121 11:34:59.160375 4824 scope.go:117] "RemoveContainer" containerID="87edd33f33351d88d9cbd6b17c925a8893b337952321dc91c1ba77e5c174ae3c" Jan 21 11:34:59 crc kubenswrapper[4824]: I0121 11:34:59.192276 4824 scope.go:117] "RemoveContainer" containerID="d91972673e1ac4ea40f718ccf4ab2ec2575e9319007c3a3cc55cff5c9b8fc08a" Jan 21 11:34:59 crc kubenswrapper[4824]: I0121 11:34:59.209011 4824 scope.go:117] "RemoveContainer" containerID="97c437ef91be90c480bbbc66844b2f1bf63fc92b3d5efc241a8ff7639e0f322c" Jan 21 11:34:59 crc kubenswrapper[4824]: I0121 11:34:59.225644 4824 scope.go:117] "RemoveContainer" containerID="82b46e3d25bb50cb39acd7bbf34eb02ad735f71fcba98f606eba2559159f21c7" Jan 21 11:34:59 crc kubenswrapper[4824]: I0121 11:34:59.241507 4824 scope.go:117] "RemoveContainer" containerID="32fb68e8168ce229b5eee2857a8b6f19fa0565f1c232d572bde25ae83e70fb37" Jan 21 11:35:16 crc kubenswrapper[4824]: I0121 11:35:16.064714 4824 patch_prober.go:28] interesting pod/machine-config-daemon-zqjg9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 21 11:35:16 crc kubenswrapper[4824]: I0121 11:35:16.065456 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 21 11:35:35 crc kubenswrapper[4824]: I0121 11:35:35.559761 4824 generic.go:334] "Generic (PLEG): container finished" podID="8dbaf72e-72a7-4073-a4c1-95cbdabaa64b" containerID="4d04c081499dd9f0ea8ea5831aa73fabb1773d706814fc10d98f8f23c9450fd3" exitCode=0 Jan 21 11:35:35 crc kubenswrapper[4824]: I0121 11:35:35.559832 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-dpgt9" event={"ID":"8dbaf72e-72a7-4073-a4c1-95cbdabaa64b","Type":"ContainerDied","Data":"4d04c081499dd9f0ea8ea5831aa73fabb1773d706814fc10d98f8f23c9450fd3"} Jan 21 11:35:36 crc kubenswrapper[4824]: I0121 11:35:36.868476 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-dpgt9" Jan 21 11:35:36 crc kubenswrapper[4824]: I0121 11:35:36.956249 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/8dbaf72e-72a7-4073-a4c1-95cbdabaa64b-ssh-key-openstack-edpm-ipam\") pod \"8dbaf72e-72a7-4073-a4c1-95cbdabaa64b\" (UID: \"8dbaf72e-72a7-4073-a4c1-95cbdabaa64b\") " Jan 21 11:35:36 crc kubenswrapper[4824]: I0121 11:35:36.956335 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8dbaf72e-72a7-4073-a4c1-95cbdabaa64b-inventory\") pod \"8dbaf72e-72a7-4073-a4c1-95cbdabaa64b\" (UID: \"8dbaf72e-72a7-4073-a4c1-95cbdabaa64b\") " Jan 21 11:35:36 crc kubenswrapper[4824]: I0121 11:35:36.956411 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v5hlg\" (UniqueName: \"kubernetes.io/projected/8dbaf72e-72a7-4073-a4c1-95cbdabaa64b-kube-api-access-v5hlg\") pod \"8dbaf72e-72a7-4073-a4c1-95cbdabaa64b\" (UID: \"8dbaf72e-72a7-4073-a4c1-95cbdabaa64b\") " Jan 21 11:35:36 crc kubenswrapper[4824]: I0121 11:35:36.961463 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8dbaf72e-72a7-4073-a4c1-95cbdabaa64b-kube-api-access-v5hlg" (OuterVolumeSpecName: "kube-api-access-v5hlg") pod "8dbaf72e-72a7-4073-a4c1-95cbdabaa64b" (UID: "8dbaf72e-72a7-4073-a4c1-95cbdabaa64b"). InnerVolumeSpecName "kube-api-access-v5hlg". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:35:36 crc kubenswrapper[4824]: I0121 11:35:36.976976 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8dbaf72e-72a7-4073-a4c1-95cbdabaa64b-inventory" (OuterVolumeSpecName: "inventory") pod "8dbaf72e-72a7-4073-a4c1-95cbdabaa64b" (UID: "8dbaf72e-72a7-4073-a4c1-95cbdabaa64b"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:35:36 crc kubenswrapper[4824]: I0121 11:35:36.978441 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8dbaf72e-72a7-4073-a4c1-95cbdabaa64b-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "8dbaf72e-72a7-4073-a4c1-95cbdabaa64b" (UID: "8dbaf72e-72a7-4073-a4c1-95cbdabaa64b"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:35:37 crc kubenswrapper[4824]: I0121 11:35:37.059071 4824 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8dbaf72e-72a7-4073-a4c1-95cbdabaa64b-inventory\") on node \"crc\" DevicePath \"\"" Jan 21 11:35:37 crc kubenswrapper[4824]: I0121 11:35:37.059101 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v5hlg\" (UniqueName: \"kubernetes.io/projected/8dbaf72e-72a7-4073-a4c1-95cbdabaa64b-kube-api-access-v5hlg\") on node \"crc\" DevicePath \"\"" Jan 21 11:35:37 crc kubenswrapper[4824]: I0121 11:35:37.059113 4824 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/8dbaf72e-72a7-4073-a4c1-95cbdabaa64b-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Jan 21 11:35:37 crc kubenswrapper[4824]: I0121 11:35:37.574341 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-dpgt9" event={"ID":"8dbaf72e-72a7-4073-a4c1-95cbdabaa64b","Type":"ContainerDied","Data":"b5d574baa9ddaffa59e308d4f888821c6b55c1c17bdef6ce2814d173fcd2ffdc"} Jan 21 11:35:37 crc kubenswrapper[4824]: I0121 11:35:37.574578 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b5d574baa9ddaffa59e308d4f888821c6b55c1c17bdef6ce2814d173fcd2ffdc" Jan 21 11:35:37 crc kubenswrapper[4824]: I0121 11:35:37.574405 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-dpgt9" Jan 21 11:35:37 crc kubenswrapper[4824]: I0121 11:35:37.634051 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wqzrz"] Jan 21 11:35:37 crc kubenswrapper[4824]: E0121 11:35:37.634415 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8dbaf72e-72a7-4073-a4c1-95cbdabaa64b" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Jan 21 11:35:37 crc kubenswrapper[4824]: I0121 11:35:37.634433 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="8dbaf72e-72a7-4073-a4c1-95cbdabaa64b" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Jan 21 11:35:37 crc kubenswrapper[4824]: I0121 11:35:37.634652 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="8dbaf72e-72a7-4073-a4c1-95cbdabaa64b" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Jan 21 11:35:37 crc kubenswrapper[4824]: I0121 11:35:37.635257 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wqzrz" Jan 21 11:35:37 crc kubenswrapper[4824]: I0121 11:35:37.636588 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Jan 21 11:35:37 crc kubenswrapper[4824]: I0121 11:35:37.637184 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Jan 21 11:35:37 crc kubenswrapper[4824]: I0121 11:35:37.637518 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Jan 21 11:35:37 crc kubenswrapper[4824]: I0121 11:35:37.637544 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-tcqk6" Jan 21 11:35:37 crc kubenswrapper[4824]: I0121 11:35:37.644723 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wqzrz"] Jan 21 11:35:37 crc kubenswrapper[4824]: I0121 11:35:37.765621 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/b8c7ec7b-ee3f-46bc-95f8-4406713ba43a-ssh-key-openstack-edpm-ipam\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-wqzrz\" (UID: \"b8c7ec7b-ee3f-46bc-95f8-4406713ba43a\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wqzrz" Jan 21 11:35:37 crc kubenswrapper[4824]: I0121 11:35:37.765750 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b8c7ec7b-ee3f-46bc-95f8-4406713ba43a-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-wqzrz\" (UID: \"b8c7ec7b-ee3f-46bc-95f8-4406713ba43a\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wqzrz" Jan 21 11:35:37 crc kubenswrapper[4824]: I0121 11:35:37.765874 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m499z\" (UniqueName: \"kubernetes.io/projected/b8c7ec7b-ee3f-46bc-95f8-4406713ba43a-kube-api-access-m499z\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-wqzrz\" (UID: \"b8c7ec7b-ee3f-46bc-95f8-4406713ba43a\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wqzrz" Jan 21 11:35:37 crc kubenswrapper[4824]: I0121 11:35:37.868114 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b8c7ec7b-ee3f-46bc-95f8-4406713ba43a-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-wqzrz\" (UID: \"b8c7ec7b-ee3f-46bc-95f8-4406713ba43a\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wqzrz" Jan 21 11:35:37 crc kubenswrapper[4824]: I0121 11:35:37.868206 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m499z\" (UniqueName: \"kubernetes.io/projected/b8c7ec7b-ee3f-46bc-95f8-4406713ba43a-kube-api-access-m499z\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-wqzrz\" (UID: \"b8c7ec7b-ee3f-46bc-95f8-4406713ba43a\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wqzrz" Jan 21 11:35:37 crc kubenswrapper[4824]: I0121 11:35:37.868249 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/b8c7ec7b-ee3f-46bc-95f8-4406713ba43a-ssh-key-openstack-edpm-ipam\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-wqzrz\" (UID: \"b8c7ec7b-ee3f-46bc-95f8-4406713ba43a\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wqzrz" Jan 21 11:35:37 crc kubenswrapper[4824]: I0121 11:35:37.871690 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b8c7ec7b-ee3f-46bc-95f8-4406713ba43a-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-wqzrz\" (UID: \"b8c7ec7b-ee3f-46bc-95f8-4406713ba43a\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wqzrz" Jan 21 11:35:37 crc kubenswrapper[4824]: I0121 11:35:37.871761 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/b8c7ec7b-ee3f-46bc-95f8-4406713ba43a-ssh-key-openstack-edpm-ipam\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-wqzrz\" (UID: \"b8c7ec7b-ee3f-46bc-95f8-4406713ba43a\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wqzrz" Jan 21 11:35:37 crc kubenswrapper[4824]: I0121 11:35:37.881347 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m499z\" (UniqueName: \"kubernetes.io/projected/b8c7ec7b-ee3f-46bc-95f8-4406713ba43a-kube-api-access-m499z\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-wqzrz\" (UID: \"b8c7ec7b-ee3f-46bc-95f8-4406713ba43a\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wqzrz" Jan 21 11:35:37 crc kubenswrapper[4824]: I0121 11:35:37.954311 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wqzrz" Jan 21 11:35:38 crc kubenswrapper[4824]: I0121 11:35:38.043595 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-jbk7q"] Jan 21 11:35:38 crc kubenswrapper[4824]: I0121 11:35:38.061857 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-a0fd-account-create-update-zxjvg"] Jan 21 11:35:38 crc kubenswrapper[4824]: I0121 11:35:38.061884 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-q7wdg"] Jan 21 11:35:38 crc kubenswrapper[4824]: I0121 11:35:38.065430 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-383b-account-create-update-w99bv"] Jan 21 11:35:38 crc kubenswrapper[4824]: I0121 11:35:38.075039 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-a0fd-account-create-update-zxjvg"] Jan 21 11:35:38 crc kubenswrapper[4824]: I0121 11:35:38.082220 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-383b-account-create-update-w99bv"] Jan 21 11:35:38 crc kubenswrapper[4824]: I0121 11:35:38.089529 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-q7wdg"] Jan 21 11:35:38 crc kubenswrapper[4824]: I0121 11:35:38.095850 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-jbk7q"] Jan 21 11:35:38 crc kubenswrapper[4824]: I0121 11:35:38.388973 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wqzrz"] Jan 21 11:35:38 crc kubenswrapper[4824]: W0121 11:35:38.396022 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb8c7ec7b_ee3f_46bc_95f8_4406713ba43a.slice/crio-64488605dbdbc01335208815ea0dcf8dcc3c4bcaaf78a8acf2f9c7b7fc464289 WatchSource:0}: Error finding container 64488605dbdbc01335208815ea0dcf8dcc3c4bcaaf78a8acf2f9c7b7fc464289: Status 404 returned error can't find the container with id 64488605dbdbc01335208815ea0dcf8dcc3c4bcaaf78a8acf2f9c7b7fc464289 Jan 21 11:35:38 crc kubenswrapper[4824]: I0121 11:35:38.581337 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wqzrz" event={"ID":"b8c7ec7b-ee3f-46bc-95f8-4406713ba43a","Type":"ContainerStarted","Data":"64488605dbdbc01335208815ea0dcf8dcc3c4bcaaf78a8acf2f9c7b7fc464289"} Jan 21 11:35:39 crc kubenswrapper[4824]: I0121 11:35:39.588469 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wqzrz" event={"ID":"b8c7ec7b-ee3f-46bc-95f8-4406713ba43a","Type":"ContainerStarted","Data":"49f4bbc73004e2aa4fd9edf12e2c015dacbc34505e45f5cb367148bf340ae9dd"} Jan 21 11:35:39 crc kubenswrapper[4824]: I0121 11:35:39.600598 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wqzrz" podStartSLOduration=2.101713229 podStartE2EDuration="2.600585481s" podCreationTimestamp="2026-01-21 11:35:37 +0000 UTC" firstStartedPulling="2026-01-21 11:35:38.397522041 +0000 UTC m=+1480.690551333" lastFinishedPulling="2026-01-21 11:35:38.896394292 +0000 UTC m=+1481.189423585" observedRunningTime="2026-01-21 11:35:39.597452125 +0000 UTC m=+1481.890481417" watchObservedRunningTime="2026-01-21 11:35:39.600585481 +0000 UTC m=+1481.893614773" Jan 21 11:35:40 crc kubenswrapper[4824]: I0121 11:35:40.057651 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7c01cf79-8a28-4229-9257-90124e66d3cc" path="/var/lib/kubelet/pods/7c01cf79-8a28-4229-9257-90124e66d3cc/volumes" Jan 21 11:35:40 crc kubenswrapper[4824]: I0121 11:35:40.058294 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="885db96c-779f-47c5-a383-1821cb53977c" path="/var/lib/kubelet/pods/885db96c-779f-47c5-a383-1821cb53977c/volumes" Jan 21 11:35:40 crc kubenswrapper[4824]: I0121 11:35:40.058796 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0ca5b98-4b63-4b4e-a80f-d52977ae3d01" path="/var/lib/kubelet/pods/a0ca5b98-4b63-4b4e-a80f-d52977ae3d01/volumes" Jan 21 11:35:40 crc kubenswrapper[4824]: I0121 11:35:40.059334 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc853ab0-e92a-4ada-84fd-b59ac2af573d" path="/var/lib/kubelet/pods/bc853ab0-e92a-4ada-84fd-b59ac2af573d/volumes" Jan 21 11:35:44 crc kubenswrapper[4824]: I0121 11:35:44.022300 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-rlxbb"] Jan 21 11:35:44 crc kubenswrapper[4824]: I0121 11:35:44.029796 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-rlxbb"] Jan 21 11:35:44 crc kubenswrapper[4824]: I0121 11:35:44.037672 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-0697-account-create-update-9nvnh"] Jan 21 11:35:44 crc kubenswrapper[4824]: I0121 11:35:44.068316 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6fbc8b12-d965-44e9-822f-b95a5c5e88ab" path="/var/lib/kubelet/pods/6fbc8b12-d965-44e9-822f-b95a5c5e88ab/volumes" Jan 21 11:35:44 crc kubenswrapper[4824]: I0121 11:35:44.069047 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-0697-account-create-update-9nvnh"] Jan 21 11:35:46 crc kubenswrapper[4824]: I0121 11:35:46.056979 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="72452c85-3787-4bd5-b605-af5555c1fdaf" path="/var/lib/kubelet/pods/72452c85-3787-4bd5-b605-af5555c1fdaf/volumes" Jan 21 11:35:46 crc kubenswrapper[4824]: I0121 11:35:46.064791 4824 patch_prober.go:28] interesting pod/machine-config-daemon-zqjg9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 21 11:35:46 crc kubenswrapper[4824]: I0121 11:35:46.064859 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 21 11:35:59 crc kubenswrapper[4824]: I0121 11:35:59.045813 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/root-account-create-update-m6rpv"] Jan 21 11:35:59 crc kubenswrapper[4824]: I0121 11:35:59.052298 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/root-account-create-update-m6rpv"] Jan 21 11:35:59 crc kubenswrapper[4824]: I0121 11:35:59.288562 4824 scope.go:117] "RemoveContainer" containerID="aece7515cad2f167dd4b5a5e58d5aa5c0df426c12c8bbc743e03dad13feaedf7" Jan 21 11:35:59 crc kubenswrapper[4824]: I0121 11:35:59.318906 4824 scope.go:117] "RemoveContainer" containerID="0811ac195531b85d7ce2dbbbbfe5e4b4eec9c6e973dbfd91e9e942aecc0be98b" Jan 21 11:35:59 crc kubenswrapper[4824]: I0121 11:35:59.340093 4824 scope.go:117] "RemoveContainer" containerID="091c53575308b0937e0789468c0edd3eb80a63b653591b4cc1559689a8b4b286" Jan 21 11:35:59 crc kubenswrapper[4824]: I0121 11:35:59.380058 4824 scope.go:117] "RemoveContainer" containerID="7d54dd5634955be1b99e4810715039653beb466487ffbc563c911444570d3daa" Jan 21 11:35:59 crc kubenswrapper[4824]: I0121 11:35:59.432740 4824 scope.go:117] "RemoveContainer" containerID="be822998cfd0d1fe00b165b7bdf2ce2c669d12e3704320cfa573970aa648494d" Jan 21 11:35:59 crc kubenswrapper[4824]: I0121 11:35:59.461559 4824 scope.go:117] "RemoveContainer" containerID="e85df67ff5e42c058f160ae767051c544cbf1f8e15561d3884068b5517d68c3a" Jan 21 11:36:00 crc kubenswrapper[4824]: I0121 11:36:00.056758 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="76e33763-1a41-4640-8e4a-492ab92009e8" path="/var/lib/kubelet/pods/76e33763-1a41-4640-8e4a-492ab92009e8/volumes" Jan 21 11:36:04 crc kubenswrapper[4824]: I0121 11:36:04.020061 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-sx9lb"] Jan 21 11:36:04 crc kubenswrapper[4824]: I0121 11:36:04.026550 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-sx9lb"] Jan 21 11:36:04 crc kubenswrapper[4824]: I0121 11:36:04.057148 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ee7a30d4-dcec-4719-a629-dfaf588d4169" path="/var/lib/kubelet/pods/ee7a30d4-dcec-4719-a629-dfaf588d4169/volumes" Jan 21 11:36:15 crc kubenswrapper[4824]: I0121 11:36:15.025259 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-2619-account-create-update-9csh2"] Jan 21 11:36:15 crc kubenswrapper[4824]: I0121 11:36:15.035175 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-db-create-kk8kb"] Jan 21 11:36:15 crc kubenswrapper[4824]: I0121 11:36:15.046603 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-db-create-kk8kb"] Jan 21 11:36:15 crc kubenswrapper[4824]: I0121 11:36:15.060730 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-2619-account-create-update-9csh2"] Jan 21 11:36:16 crc kubenswrapper[4824]: I0121 11:36:16.056928 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8600373d-f466-43d4-92fc-1fa938f6e91b" path="/var/lib/kubelet/pods/8600373d-f466-43d4-92fc-1fa938f6e91b/volumes" Jan 21 11:36:16 crc kubenswrapper[4824]: I0121 11:36:16.057715 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fe6e7565-793b-4555-b6d5-758f87a5b9c7" path="/var/lib/kubelet/pods/fe6e7565-793b-4555-b6d5-758f87a5b9c7/volumes" Jan 21 11:36:16 crc kubenswrapper[4824]: I0121 11:36:16.064732 4824 patch_prober.go:28] interesting pod/machine-config-daemon-zqjg9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 21 11:36:16 crc kubenswrapper[4824]: I0121 11:36:16.064780 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 21 11:36:16 crc kubenswrapper[4824]: I0121 11:36:16.064816 4824 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" Jan 21 11:36:16 crc kubenswrapper[4824]: I0121 11:36:16.065223 4824 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"1cd835fe36ebb6541740e5ab5fd843470d46dfb78975fd968e5cc9aaef462499"} pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 21 11:36:16 crc kubenswrapper[4824]: I0121 11:36:16.065277 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" containerName="machine-config-daemon" containerID="cri-o://1cd835fe36ebb6541740e5ab5fd843470d46dfb78975fd968e5cc9aaef462499" gracePeriod=600 Jan 21 11:36:16 crc kubenswrapper[4824]: E0121 11:36:16.191313 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 11:36:16 crc kubenswrapper[4824]: I0121 11:36:16.827867 4824 generic.go:334] "Generic (PLEG): container finished" podID="33f3d922-4ffe-409b-a49a-d88c85898260" containerID="1cd835fe36ebb6541740e5ab5fd843470d46dfb78975fd968e5cc9aaef462499" exitCode=0 Jan 21 11:36:16 crc kubenswrapper[4824]: I0121 11:36:16.827920 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" event={"ID":"33f3d922-4ffe-409b-a49a-d88c85898260","Type":"ContainerDied","Data":"1cd835fe36ebb6541740e5ab5fd843470d46dfb78975fd968e5cc9aaef462499"} Jan 21 11:36:16 crc kubenswrapper[4824]: I0121 11:36:16.828133 4824 scope.go:117] "RemoveContainer" containerID="f5273806c339b109c1d902705cf82c97535d3adc4a021ab1d42339f22a956af4" Jan 21 11:36:16 crc kubenswrapper[4824]: I0121 11:36:16.828705 4824 scope.go:117] "RemoveContainer" containerID="1cd835fe36ebb6541740e5ab5fd843470d46dfb78975fd968e5cc9aaef462499" Jan 21 11:36:16 crc kubenswrapper[4824]: E0121 11:36:16.828981 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 11:36:18 crc kubenswrapper[4824]: I0121 11:36:18.096453 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-twgjw"] Jan 21 11:36:18 crc kubenswrapper[4824]: I0121 11:36:18.105753 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-60f6-account-create-update-pjh6n"] Jan 21 11:36:18 crc kubenswrapper[4824]: I0121 11:36:18.112131 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-ngtnj"] Jan 21 11:36:18 crc kubenswrapper[4824]: I0121 11:36:18.118227 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-60f6-account-create-update-pjh6n"] Jan 21 11:36:18 crc kubenswrapper[4824]: I0121 11:36:18.123986 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-0662-account-create-update-j7bhp"] Jan 21 11:36:18 crc kubenswrapper[4824]: I0121 11:36:18.128991 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-ngtnj"] Jan 21 11:36:18 crc kubenswrapper[4824]: I0121 11:36:18.133890 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-twgjw"] Jan 21 11:36:18 crc kubenswrapper[4824]: I0121 11:36:18.138859 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-3c9f-account-create-update-rtprr"] Jan 21 11:36:18 crc kubenswrapper[4824]: I0121 11:36:18.143853 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-0662-account-create-update-j7bhp"] Jan 21 11:36:18 crc kubenswrapper[4824]: I0121 11:36:18.148774 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-3c9f-account-create-update-rtprr"] Jan 21 11:36:18 crc kubenswrapper[4824]: I0121 11:36:18.153687 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-nskxc"] Jan 21 11:36:18 crc kubenswrapper[4824]: I0121 11:36:18.158610 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-nskxc"] Jan 21 11:36:18 crc kubenswrapper[4824]: I0121 11:36:18.239456 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-xhcp9"] Jan 21 11:36:18 crc kubenswrapper[4824]: I0121 11:36:18.241029 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xhcp9" Jan 21 11:36:18 crc kubenswrapper[4824]: I0121 11:36:18.247588 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-xhcp9"] Jan 21 11:36:18 crc kubenswrapper[4824]: I0121 11:36:18.379741 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/edf7c1b7-286d-4325-ade6-e13af1ba3b26-utilities\") pod \"redhat-marketplace-xhcp9\" (UID: \"edf7c1b7-286d-4325-ade6-e13af1ba3b26\") " pod="openshift-marketplace/redhat-marketplace-xhcp9" Jan 21 11:36:18 crc kubenswrapper[4824]: I0121 11:36:18.379876 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pn7n7\" (UniqueName: \"kubernetes.io/projected/edf7c1b7-286d-4325-ade6-e13af1ba3b26-kube-api-access-pn7n7\") pod \"redhat-marketplace-xhcp9\" (UID: \"edf7c1b7-286d-4325-ade6-e13af1ba3b26\") " pod="openshift-marketplace/redhat-marketplace-xhcp9" Jan 21 11:36:18 crc kubenswrapper[4824]: I0121 11:36:18.379942 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/edf7c1b7-286d-4325-ade6-e13af1ba3b26-catalog-content\") pod \"redhat-marketplace-xhcp9\" (UID: \"edf7c1b7-286d-4325-ade6-e13af1ba3b26\") " pod="openshift-marketplace/redhat-marketplace-xhcp9" Jan 21 11:36:18 crc kubenswrapper[4824]: I0121 11:36:18.481591 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/edf7c1b7-286d-4325-ade6-e13af1ba3b26-utilities\") pod \"redhat-marketplace-xhcp9\" (UID: \"edf7c1b7-286d-4325-ade6-e13af1ba3b26\") " pod="openshift-marketplace/redhat-marketplace-xhcp9" Jan 21 11:36:18 crc kubenswrapper[4824]: I0121 11:36:18.481700 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pn7n7\" (UniqueName: \"kubernetes.io/projected/edf7c1b7-286d-4325-ade6-e13af1ba3b26-kube-api-access-pn7n7\") pod \"redhat-marketplace-xhcp9\" (UID: \"edf7c1b7-286d-4325-ade6-e13af1ba3b26\") " pod="openshift-marketplace/redhat-marketplace-xhcp9" Jan 21 11:36:18 crc kubenswrapper[4824]: I0121 11:36:18.481764 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/edf7c1b7-286d-4325-ade6-e13af1ba3b26-catalog-content\") pod \"redhat-marketplace-xhcp9\" (UID: \"edf7c1b7-286d-4325-ade6-e13af1ba3b26\") " pod="openshift-marketplace/redhat-marketplace-xhcp9" Jan 21 11:36:18 crc kubenswrapper[4824]: I0121 11:36:18.482142 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/edf7c1b7-286d-4325-ade6-e13af1ba3b26-catalog-content\") pod \"redhat-marketplace-xhcp9\" (UID: \"edf7c1b7-286d-4325-ade6-e13af1ba3b26\") " pod="openshift-marketplace/redhat-marketplace-xhcp9" Jan 21 11:36:18 crc kubenswrapper[4824]: I0121 11:36:18.482360 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/edf7c1b7-286d-4325-ade6-e13af1ba3b26-utilities\") pod \"redhat-marketplace-xhcp9\" (UID: \"edf7c1b7-286d-4325-ade6-e13af1ba3b26\") " pod="openshift-marketplace/redhat-marketplace-xhcp9" Jan 21 11:36:18 crc kubenswrapper[4824]: I0121 11:36:18.498148 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pn7n7\" (UniqueName: \"kubernetes.io/projected/edf7c1b7-286d-4325-ade6-e13af1ba3b26-kube-api-access-pn7n7\") pod \"redhat-marketplace-xhcp9\" (UID: \"edf7c1b7-286d-4325-ade6-e13af1ba3b26\") " pod="openshift-marketplace/redhat-marketplace-xhcp9" Jan 21 11:36:18 crc kubenswrapper[4824]: I0121 11:36:18.554724 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xhcp9" Jan 21 11:36:19 crc kubenswrapper[4824]: I0121 11:36:19.072641 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-xhcp9"] Jan 21 11:36:19 crc kubenswrapper[4824]: I0121 11:36:19.852612 4824 generic.go:334] "Generic (PLEG): container finished" podID="edf7c1b7-286d-4325-ade6-e13af1ba3b26" containerID="88e966ca56affbabe3e790a444ff3a0ca4ea80b5ddef8fd97b42647930a06c5d" exitCode=0 Jan 21 11:36:19 crc kubenswrapper[4824]: I0121 11:36:19.852699 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xhcp9" event={"ID":"edf7c1b7-286d-4325-ade6-e13af1ba3b26","Type":"ContainerDied","Data":"88e966ca56affbabe3e790a444ff3a0ca4ea80b5ddef8fd97b42647930a06c5d"} Jan 21 11:36:19 crc kubenswrapper[4824]: I0121 11:36:19.852825 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xhcp9" event={"ID":"edf7c1b7-286d-4325-ade6-e13af1ba3b26","Type":"ContainerStarted","Data":"33b7c4d8b7630213d98e01a69d1c43ab8893a8320c312d1bd10ff4050bcd0312"} Jan 21 11:36:19 crc kubenswrapper[4824]: I0121 11:36:19.854209 4824 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 21 11:36:20 crc kubenswrapper[4824]: I0121 11:36:20.057453 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5c0ddc66-7696-4816-8e96-340aeb39d242" path="/var/lib/kubelet/pods/5c0ddc66-7696-4816-8e96-340aeb39d242/volumes" Jan 21 11:36:20 crc kubenswrapper[4824]: I0121 11:36:20.058275 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7b747df8-e61f-4689-a08c-0dc6688f1a0e" path="/var/lib/kubelet/pods/7b747df8-e61f-4689-a08c-0dc6688f1a0e/volumes" Jan 21 11:36:20 crc kubenswrapper[4824]: I0121 11:36:20.058815 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="93c8ce3d-a4f0-4cc7-8592-61653eb09d3c" path="/var/lib/kubelet/pods/93c8ce3d-a4f0-4cc7-8592-61653eb09d3c/volumes" Jan 21 11:36:20 crc kubenswrapper[4824]: I0121 11:36:20.059355 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9be37b35-7ad5-434a-8d16-c2f9f1661821" path="/var/lib/kubelet/pods/9be37b35-7ad5-434a-8d16-c2f9f1661821/volumes" Jan 21 11:36:20 crc kubenswrapper[4824]: I0121 11:36:20.060297 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a10a8f17-c27b-4544-8f9d-a6afd9991a8e" path="/var/lib/kubelet/pods/a10a8f17-c27b-4544-8f9d-a6afd9991a8e/volumes" Jan 21 11:36:20 crc kubenswrapper[4824]: I0121 11:36:20.060807 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b4773e7a-a50c-42e6-bb27-a25b0055b8e3" path="/var/lib/kubelet/pods/b4773e7a-a50c-42e6-bb27-a25b0055b8e3/volumes" Jan 21 11:36:20 crc kubenswrapper[4824]: I0121 11:36:20.861359 4824 generic.go:334] "Generic (PLEG): container finished" podID="edf7c1b7-286d-4325-ade6-e13af1ba3b26" containerID="17a58cf8bcafa90912c5e5b9c5c86b4fe93b7ac6ef62d41aae668feae7eddb20" exitCode=0 Jan 21 11:36:20 crc kubenswrapper[4824]: I0121 11:36:20.861569 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xhcp9" event={"ID":"edf7c1b7-286d-4325-ade6-e13af1ba3b26","Type":"ContainerDied","Data":"17a58cf8bcafa90912c5e5b9c5c86b4fe93b7ac6ef62d41aae668feae7eddb20"} Jan 21 11:36:21 crc kubenswrapper[4824]: I0121 11:36:21.030157 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-84f4b"] Jan 21 11:36:21 crc kubenswrapper[4824]: I0121 11:36:21.036435 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-84f4b"] Jan 21 11:36:21 crc kubenswrapper[4824]: I0121 11:36:21.869665 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xhcp9" event={"ID":"edf7c1b7-286d-4325-ade6-e13af1ba3b26","Type":"ContainerStarted","Data":"051a685415639cbce6fad623dcddce4af30d9f613a0a4899af67a843df696829"} Jan 21 11:36:21 crc kubenswrapper[4824]: I0121 11:36:21.887470 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-xhcp9" podStartSLOduration=2.4058206540000002 podStartE2EDuration="3.88743833s" podCreationTimestamp="2026-01-21 11:36:18 +0000 UTC" firstStartedPulling="2026-01-21 11:36:19.854000603 +0000 UTC m=+1522.147029895" lastFinishedPulling="2026-01-21 11:36:21.33561828 +0000 UTC m=+1523.628647571" observedRunningTime="2026-01-21 11:36:21.884042188 +0000 UTC m=+1524.177071480" watchObservedRunningTime="2026-01-21 11:36:21.88743833 +0000 UTC m=+1524.180467621" Jan 21 11:36:22 crc kubenswrapper[4824]: I0121 11:36:22.057836 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5240d35b-cbf1-472d-91da-debf418dd208" path="/var/lib/kubelet/pods/5240d35b-cbf1-472d-91da-debf418dd208/volumes" Jan 21 11:36:28 crc kubenswrapper[4824]: I0121 11:36:28.559230 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-xhcp9" Jan 21 11:36:28 crc kubenswrapper[4824]: I0121 11:36:28.559631 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-xhcp9" Jan 21 11:36:28 crc kubenswrapper[4824]: I0121 11:36:28.594355 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-xhcp9" Jan 21 11:36:28 crc kubenswrapper[4824]: I0121 11:36:28.955781 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-xhcp9" Jan 21 11:36:29 crc kubenswrapper[4824]: I0121 11:36:29.003675 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-xhcp9"] Jan 21 11:36:29 crc kubenswrapper[4824]: I0121 11:36:29.049295 4824 scope.go:117] "RemoveContainer" containerID="1cd835fe36ebb6541740e5ab5fd843470d46dfb78975fd968e5cc9aaef462499" Jan 21 11:36:29 crc kubenswrapper[4824]: E0121 11:36:29.049644 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 11:36:30 crc kubenswrapper[4824]: I0121 11:36:30.936296 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-xhcp9" podUID="edf7c1b7-286d-4325-ade6-e13af1ba3b26" containerName="registry-server" containerID="cri-o://051a685415639cbce6fad623dcddce4af30d9f613a0a4899af67a843df696829" gracePeriod=2 Jan 21 11:36:31 crc kubenswrapper[4824]: I0121 11:36:31.323840 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xhcp9" Jan 21 11:36:31 crc kubenswrapper[4824]: I0121 11:36:31.488391 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pn7n7\" (UniqueName: \"kubernetes.io/projected/edf7c1b7-286d-4325-ade6-e13af1ba3b26-kube-api-access-pn7n7\") pod \"edf7c1b7-286d-4325-ade6-e13af1ba3b26\" (UID: \"edf7c1b7-286d-4325-ade6-e13af1ba3b26\") " Jan 21 11:36:31 crc kubenswrapper[4824]: I0121 11:36:31.488858 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/edf7c1b7-286d-4325-ade6-e13af1ba3b26-catalog-content\") pod \"edf7c1b7-286d-4325-ade6-e13af1ba3b26\" (UID: \"edf7c1b7-286d-4325-ade6-e13af1ba3b26\") " Jan 21 11:36:31 crc kubenswrapper[4824]: I0121 11:36:31.489023 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/edf7c1b7-286d-4325-ade6-e13af1ba3b26-utilities\") pod \"edf7c1b7-286d-4325-ade6-e13af1ba3b26\" (UID: \"edf7c1b7-286d-4325-ade6-e13af1ba3b26\") " Jan 21 11:36:31 crc kubenswrapper[4824]: I0121 11:36:31.489597 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/edf7c1b7-286d-4325-ade6-e13af1ba3b26-utilities" (OuterVolumeSpecName: "utilities") pod "edf7c1b7-286d-4325-ade6-e13af1ba3b26" (UID: "edf7c1b7-286d-4325-ade6-e13af1ba3b26"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:36:31 crc kubenswrapper[4824]: I0121 11:36:31.489791 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/edf7c1b7-286d-4325-ade6-e13af1ba3b26-utilities\") on node \"crc\" DevicePath \"\"" Jan 21 11:36:31 crc kubenswrapper[4824]: I0121 11:36:31.494019 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/edf7c1b7-286d-4325-ade6-e13af1ba3b26-kube-api-access-pn7n7" (OuterVolumeSpecName: "kube-api-access-pn7n7") pod "edf7c1b7-286d-4325-ade6-e13af1ba3b26" (UID: "edf7c1b7-286d-4325-ade6-e13af1ba3b26"). InnerVolumeSpecName "kube-api-access-pn7n7". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:36:31 crc kubenswrapper[4824]: I0121 11:36:31.506989 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/edf7c1b7-286d-4325-ade6-e13af1ba3b26-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "edf7c1b7-286d-4325-ade6-e13af1ba3b26" (UID: "edf7c1b7-286d-4325-ade6-e13af1ba3b26"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:36:31 crc kubenswrapper[4824]: I0121 11:36:31.592119 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/edf7c1b7-286d-4325-ade6-e13af1ba3b26-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 21 11:36:31 crc kubenswrapper[4824]: I0121 11:36:31.592148 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pn7n7\" (UniqueName: \"kubernetes.io/projected/edf7c1b7-286d-4325-ade6-e13af1ba3b26-kube-api-access-pn7n7\") on node \"crc\" DevicePath \"\"" Jan 21 11:36:31 crc kubenswrapper[4824]: I0121 11:36:31.944632 4824 generic.go:334] "Generic (PLEG): container finished" podID="edf7c1b7-286d-4325-ade6-e13af1ba3b26" containerID="051a685415639cbce6fad623dcddce4af30d9f613a0a4899af67a843df696829" exitCode=0 Jan 21 11:36:31 crc kubenswrapper[4824]: I0121 11:36:31.944670 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xhcp9" event={"ID":"edf7c1b7-286d-4325-ade6-e13af1ba3b26","Type":"ContainerDied","Data":"051a685415639cbce6fad623dcddce4af30d9f613a0a4899af67a843df696829"} Jan 21 11:36:31 crc kubenswrapper[4824]: I0121 11:36:31.944693 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xhcp9" event={"ID":"edf7c1b7-286d-4325-ade6-e13af1ba3b26","Type":"ContainerDied","Data":"33b7c4d8b7630213d98e01a69d1c43ab8893a8320c312d1bd10ff4050bcd0312"} Jan 21 11:36:31 crc kubenswrapper[4824]: I0121 11:36:31.944707 4824 scope.go:117] "RemoveContainer" containerID="051a685415639cbce6fad623dcddce4af30d9f613a0a4899af67a843df696829" Jan 21 11:36:31 crc kubenswrapper[4824]: I0121 11:36:31.944802 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xhcp9" Jan 21 11:36:31 crc kubenswrapper[4824]: I0121 11:36:31.962714 4824 scope.go:117] "RemoveContainer" containerID="17a58cf8bcafa90912c5e5b9c5c86b4fe93b7ac6ef62d41aae668feae7eddb20" Jan 21 11:36:31 crc kubenswrapper[4824]: I0121 11:36:31.976194 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-xhcp9"] Jan 21 11:36:31 crc kubenswrapper[4824]: I0121 11:36:31.982635 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-xhcp9"] Jan 21 11:36:31 crc kubenswrapper[4824]: I0121 11:36:31.993183 4824 scope.go:117] "RemoveContainer" containerID="88e966ca56affbabe3e790a444ff3a0ca4ea80b5ddef8fd97b42647930a06c5d" Jan 21 11:36:32 crc kubenswrapper[4824]: I0121 11:36:32.016464 4824 scope.go:117] "RemoveContainer" containerID="051a685415639cbce6fad623dcddce4af30d9f613a0a4899af67a843df696829" Jan 21 11:36:32 crc kubenswrapper[4824]: E0121 11:36:32.019908 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"051a685415639cbce6fad623dcddce4af30d9f613a0a4899af67a843df696829\": container with ID starting with 051a685415639cbce6fad623dcddce4af30d9f613a0a4899af67a843df696829 not found: ID does not exist" containerID="051a685415639cbce6fad623dcddce4af30d9f613a0a4899af67a843df696829" Jan 21 11:36:32 crc kubenswrapper[4824]: I0121 11:36:32.019943 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"051a685415639cbce6fad623dcddce4af30d9f613a0a4899af67a843df696829"} err="failed to get container status \"051a685415639cbce6fad623dcddce4af30d9f613a0a4899af67a843df696829\": rpc error: code = NotFound desc = could not find container \"051a685415639cbce6fad623dcddce4af30d9f613a0a4899af67a843df696829\": container with ID starting with 051a685415639cbce6fad623dcddce4af30d9f613a0a4899af67a843df696829 not found: ID does not exist" Jan 21 11:36:32 crc kubenswrapper[4824]: I0121 11:36:32.020005 4824 scope.go:117] "RemoveContainer" containerID="17a58cf8bcafa90912c5e5b9c5c86b4fe93b7ac6ef62d41aae668feae7eddb20" Jan 21 11:36:32 crc kubenswrapper[4824]: E0121 11:36:32.020265 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"17a58cf8bcafa90912c5e5b9c5c86b4fe93b7ac6ef62d41aae668feae7eddb20\": container with ID starting with 17a58cf8bcafa90912c5e5b9c5c86b4fe93b7ac6ef62d41aae668feae7eddb20 not found: ID does not exist" containerID="17a58cf8bcafa90912c5e5b9c5c86b4fe93b7ac6ef62d41aae668feae7eddb20" Jan 21 11:36:32 crc kubenswrapper[4824]: I0121 11:36:32.020287 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"17a58cf8bcafa90912c5e5b9c5c86b4fe93b7ac6ef62d41aae668feae7eddb20"} err="failed to get container status \"17a58cf8bcafa90912c5e5b9c5c86b4fe93b7ac6ef62d41aae668feae7eddb20\": rpc error: code = NotFound desc = could not find container \"17a58cf8bcafa90912c5e5b9c5c86b4fe93b7ac6ef62d41aae668feae7eddb20\": container with ID starting with 17a58cf8bcafa90912c5e5b9c5c86b4fe93b7ac6ef62d41aae668feae7eddb20 not found: ID does not exist" Jan 21 11:36:32 crc kubenswrapper[4824]: I0121 11:36:32.020299 4824 scope.go:117] "RemoveContainer" containerID="88e966ca56affbabe3e790a444ff3a0ca4ea80b5ddef8fd97b42647930a06c5d" Jan 21 11:36:32 crc kubenswrapper[4824]: E0121 11:36:32.020892 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"88e966ca56affbabe3e790a444ff3a0ca4ea80b5ddef8fd97b42647930a06c5d\": container with ID starting with 88e966ca56affbabe3e790a444ff3a0ca4ea80b5ddef8fd97b42647930a06c5d not found: ID does not exist" containerID="88e966ca56affbabe3e790a444ff3a0ca4ea80b5ddef8fd97b42647930a06c5d" Jan 21 11:36:32 crc kubenswrapper[4824]: I0121 11:36:32.020924 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"88e966ca56affbabe3e790a444ff3a0ca4ea80b5ddef8fd97b42647930a06c5d"} err="failed to get container status \"88e966ca56affbabe3e790a444ff3a0ca4ea80b5ddef8fd97b42647930a06c5d\": rpc error: code = NotFound desc = could not find container \"88e966ca56affbabe3e790a444ff3a0ca4ea80b5ddef8fd97b42647930a06c5d\": container with ID starting with 88e966ca56affbabe3e790a444ff3a0ca4ea80b5ddef8fd97b42647930a06c5d not found: ID does not exist" Jan 21 11:36:32 crc kubenswrapper[4824]: I0121 11:36:32.056724 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="edf7c1b7-286d-4325-ade6-e13af1ba3b26" path="/var/lib/kubelet/pods/edf7c1b7-286d-4325-ade6-e13af1ba3b26/volumes" Jan 21 11:36:34 crc kubenswrapper[4824]: I0121 11:36:34.966525 4824 generic.go:334] "Generic (PLEG): container finished" podID="b8c7ec7b-ee3f-46bc-95f8-4406713ba43a" containerID="49f4bbc73004e2aa4fd9edf12e2c015dacbc34505e45f5cb367148bf340ae9dd" exitCode=0 Jan 21 11:36:34 crc kubenswrapper[4824]: I0121 11:36:34.966623 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wqzrz" event={"ID":"b8c7ec7b-ee3f-46bc-95f8-4406713ba43a","Type":"ContainerDied","Data":"49f4bbc73004e2aa4fd9edf12e2c015dacbc34505e45f5cb367148bf340ae9dd"} Jan 21 11:36:36 crc kubenswrapper[4824]: I0121 11:36:36.289841 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wqzrz" Jan 21 11:36:36 crc kubenswrapper[4824]: I0121 11:36:36.376312 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m499z\" (UniqueName: \"kubernetes.io/projected/b8c7ec7b-ee3f-46bc-95f8-4406713ba43a-kube-api-access-m499z\") pod \"b8c7ec7b-ee3f-46bc-95f8-4406713ba43a\" (UID: \"b8c7ec7b-ee3f-46bc-95f8-4406713ba43a\") " Jan 21 11:36:36 crc kubenswrapper[4824]: I0121 11:36:36.376516 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/b8c7ec7b-ee3f-46bc-95f8-4406713ba43a-ssh-key-openstack-edpm-ipam\") pod \"b8c7ec7b-ee3f-46bc-95f8-4406713ba43a\" (UID: \"b8c7ec7b-ee3f-46bc-95f8-4406713ba43a\") " Jan 21 11:36:36 crc kubenswrapper[4824]: I0121 11:36:36.376717 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b8c7ec7b-ee3f-46bc-95f8-4406713ba43a-inventory\") pod \"b8c7ec7b-ee3f-46bc-95f8-4406713ba43a\" (UID: \"b8c7ec7b-ee3f-46bc-95f8-4406713ba43a\") " Jan 21 11:36:36 crc kubenswrapper[4824]: I0121 11:36:36.382799 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b8c7ec7b-ee3f-46bc-95f8-4406713ba43a-kube-api-access-m499z" (OuterVolumeSpecName: "kube-api-access-m499z") pod "b8c7ec7b-ee3f-46bc-95f8-4406713ba43a" (UID: "b8c7ec7b-ee3f-46bc-95f8-4406713ba43a"). InnerVolumeSpecName "kube-api-access-m499z". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:36:36 crc kubenswrapper[4824]: I0121 11:36:36.400630 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b8c7ec7b-ee3f-46bc-95f8-4406713ba43a-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "b8c7ec7b-ee3f-46bc-95f8-4406713ba43a" (UID: "b8c7ec7b-ee3f-46bc-95f8-4406713ba43a"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:36:36 crc kubenswrapper[4824]: I0121 11:36:36.406806 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b8c7ec7b-ee3f-46bc-95f8-4406713ba43a-inventory" (OuterVolumeSpecName: "inventory") pod "b8c7ec7b-ee3f-46bc-95f8-4406713ba43a" (UID: "b8c7ec7b-ee3f-46bc-95f8-4406713ba43a"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:36:36 crc kubenswrapper[4824]: I0121 11:36:36.479157 4824 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/b8c7ec7b-ee3f-46bc-95f8-4406713ba43a-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Jan 21 11:36:36 crc kubenswrapper[4824]: I0121 11:36:36.479198 4824 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b8c7ec7b-ee3f-46bc-95f8-4406713ba43a-inventory\") on node \"crc\" DevicePath \"\"" Jan 21 11:36:36 crc kubenswrapper[4824]: I0121 11:36:36.479212 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m499z\" (UniqueName: \"kubernetes.io/projected/b8c7ec7b-ee3f-46bc-95f8-4406713ba43a-kube-api-access-m499z\") on node \"crc\" DevicePath \"\"" Jan 21 11:36:36 crc kubenswrapper[4824]: I0121 11:36:36.981090 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wqzrz" event={"ID":"b8c7ec7b-ee3f-46bc-95f8-4406713ba43a","Type":"ContainerDied","Data":"64488605dbdbc01335208815ea0dcf8dcc3c4bcaaf78a8acf2f9c7b7fc464289"} Jan 21 11:36:36 crc kubenswrapper[4824]: I0121 11:36:36.981397 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="64488605dbdbc01335208815ea0dcf8dcc3c4bcaaf78a8acf2f9c7b7fc464289" Jan 21 11:36:36 crc kubenswrapper[4824]: I0121 11:36:36.981165 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wqzrz" Jan 21 11:36:37 crc kubenswrapper[4824]: I0121 11:36:37.044079 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-5b6wz"] Jan 21 11:36:37 crc kubenswrapper[4824]: E0121 11:36:37.044477 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="edf7c1b7-286d-4325-ade6-e13af1ba3b26" containerName="extract-utilities" Jan 21 11:36:37 crc kubenswrapper[4824]: I0121 11:36:37.044498 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="edf7c1b7-286d-4325-ade6-e13af1ba3b26" containerName="extract-utilities" Jan 21 11:36:37 crc kubenswrapper[4824]: E0121 11:36:37.044507 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="edf7c1b7-286d-4325-ade6-e13af1ba3b26" containerName="registry-server" Jan 21 11:36:37 crc kubenswrapper[4824]: I0121 11:36:37.044513 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="edf7c1b7-286d-4325-ade6-e13af1ba3b26" containerName="registry-server" Jan 21 11:36:37 crc kubenswrapper[4824]: E0121 11:36:37.044532 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b8c7ec7b-ee3f-46bc-95f8-4406713ba43a" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Jan 21 11:36:37 crc kubenswrapper[4824]: I0121 11:36:37.044539 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="b8c7ec7b-ee3f-46bc-95f8-4406713ba43a" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Jan 21 11:36:37 crc kubenswrapper[4824]: E0121 11:36:37.044558 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="edf7c1b7-286d-4325-ade6-e13af1ba3b26" containerName="extract-content" Jan 21 11:36:37 crc kubenswrapper[4824]: I0121 11:36:37.044563 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="edf7c1b7-286d-4325-ade6-e13af1ba3b26" containerName="extract-content" Jan 21 11:36:37 crc kubenswrapper[4824]: I0121 11:36:37.044757 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="edf7c1b7-286d-4325-ade6-e13af1ba3b26" containerName="registry-server" Jan 21 11:36:37 crc kubenswrapper[4824]: I0121 11:36:37.044773 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="b8c7ec7b-ee3f-46bc-95f8-4406713ba43a" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Jan 21 11:36:37 crc kubenswrapper[4824]: I0121 11:36:37.045430 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-5b6wz" Jan 21 11:36:37 crc kubenswrapper[4824]: I0121 11:36:37.046861 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-tcqk6" Jan 21 11:36:37 crc kubenswrapper[4824]: I0121 11:36:37.047018 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Jan 21 11:36:37 crc kubenswrapper[4824]: I0121 11:36:37.047071 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Jan 21 11:36:37 crc kubenswrapper[4824]: I0121 11:36:37.047541 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Jan 21 11:36:37 crc kubenswrapper[4824]: I0121 11:36:37.049249 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-5b6wz"] Jan 21 11:36:37 crc kubenswrapper[4824]: I0121 11:36:37.191383 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-92v6n\" (UniqueName: \"kubernetes.io/projected/d672f153-6ef2-432d-829a-db63c8daab69-kube-api-access-92v6n\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-5b6wz\" (UID: \"d672f153-6ef2-432d-829a-db63c8daab69\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-5b6wz" Jan 21 11:36:37 crc kubenswrapper[4824]: I0121 11:36:37.191557 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/d672f153-6ef2-432d-829a-db63c8daab69-ssh-key-openstack-edpm-ipam\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-5b6wz\" (UID: \"d672f153-6ef2-432d-829a-db63c8daab69\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-5b6wz" Jan 21 11:36:37 crc kubenswrapper[4824]: I0121 11:36:37.191618 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d672f153-6ef2-432d-829a-db63c8daab69-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-5b6wz\" (UID: \"d672f153-6ef2-432d-829a-db63c8daab69\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-5b6wz" Jan 21 11:36:37 crc kubenswrapper[4824]: I0121 11:36:37.293647 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/d672f153-6ef2-432d-829a-db63c8daab69-ssh-key-openstack-edpm-ipam\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-5b6wz\" (UID: \"d672f153-6ef2-432d-829a-db63c8daab69\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-5b6wz" Jan 21 11:36:37 crc kubenswrapper[4824]: I0121 11:36:37.293731 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d672f153-6ef2-432d-829a-db63c8daab69-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-5b6wz\" (UID: \"d672f153-6ef2-432d-829a-db63c8daab69\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-5b6wz" Jan 21 11:36:37 crc kubenswrapper[4824]: I0121 11:36:37.293888 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-92v6n\" (UniqueName: \"kubernetes.io/projected/d672f153-6ef2-432d-829a-db63c8daab69-kube-api-access-92v6n\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-5b6wz\" (UID: \"d672f153-6ef2-432d-829a-db63c8daab69\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-5b6wz" Jan 21 11:36:37 crc kubenswrapper[4824]: I0121 11:36:37.297427 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d672f153-6ef2-432d-829a-db63c8daab69-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-5b6wz\" (UID: \"d672f153-6ef2-432d-829a-db63c8daab69\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-5b6wz" Jan 21 11:36:37 crc kubenswrapper[4824]: I0121 11:36:37.298865 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/d672f153-6ef2-432d-829a-db63c8daab69-ssh-key-openstack-edpm-ipam\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-5b6wz\" (UID: \"d672f153-6ef2-432d-829a-db63c8daab69\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-5b6wz" Jan 21 11:36:37 crc kubenswrapper[4824]: I0121 11:36:37.310242 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-92v6n\" (UniqueName: \"kubernetes.io/projected/d672f153-6ef2-432d-829a-db63c8daab69-kube-api-access-92v6n\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-5b6wz\" (UID: \"d672f153-6ef2-432d-829a-db63c8daab69\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-5b6wz" Jan 21 11:36:37 crc kubenswrapper[4824]: I0121 11:36:37.359744 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-5b6wz" Jan 21 11:36:37 crc kubenswrapper[4824]: I0121 11:36:37.803628 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-5b6wz"] Jan 21 11:36:37 crc kubenswrapper[4824]: I0121 11:36:37.989114 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-5b6wz" event={"ID":"d672f153-6ef2-432d-829a-db63c8daab69","Type":"ContainerStarted","Data":"73fe2f366123c47bd83aebfb3c09855e44f3bceec576e057fa2d17a95e4879fb"} Jan 21 11:36:39 crc kubenswrapper[4824]: I0121 11:36:39.001311 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-5b6wz" event={"ID":"d672f153-6ef2-432d-829a-db63c8daab69","Type":"ContainerStarted","Data":"7778ba4059495bd1d394bac5a6c5c17aa054fa3ecf50af111d3388b007e74498"} Jan 21 11:36:39 crc kubenswrapper[4824]: I0121 11:36:39.017068 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-5b6wz" podStartSLOduration=1.483571178 podStartE2EDuration="2.01705351s" podCreationTimestamp="2026-01-21 11:36:37 +0000 UTC" firstStartedPulling="2026-01-21 11:36:37.806452611 +0000 UTC m=+1540.099481892" lastFinishedPulling="2026-01-21 11:36:38.339934931 +0000 UTC m=+1540.632964224" observedRunningTime="2026-01-21 11:36:39.011750658 +0000 UTC m=+1541.304779951" watchObservedRunningTime="2026-01-21 11:36:39.01705351 +0000 UTC m=+1541.310082801" Jan 21 11:36:40 crc kubenswrapper[4824]: I0121 11:36:40.031710 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-t2txg"] Jan 21 11:36:40 crc kubenswrapper[4824]: I0121 11:36:40.038574 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-t2txg"] Jan 21 11:36:40 crc kubenswrapper[4824]: I0121 11:36:40.059722 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a19b6730-b7aa-4319-8127-c32d7874a471" path="/var/lib/kubelet/pods/a19b6730-b7aa-4319-8127-c32d7874a471/volumes" Jan 21 11:36:40 crc kubenswrapper[4824]: I0121 11:36:40.296453 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-m5zcg"] Jan 21 11:36:40 crc kubenswrapper[4824]: I0121 11:36:40.298268 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-m5zcg" Jan 21 11:36:40 crc kubenswrapper[4824]: I0121 11:36:40.312081 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-m5zcg"] Jan 21 11:36:40 crc kubenswrapper[4824]: I0121 11:36:40.449260 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6ba09de8-56d0-4a49-90fd-0901ef2a4983-utilities\") pod \"community-operators-m5zcg\" (UID: \"6ba09de8-56d0-4a49-90fd-0901ef2a4983\") " pod="openshift-marketplace/community-operators-m5zcg" Jan 21 11:36:40 crc kubenswrapper[4824]: I0121 11:36:40.449333 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wbmzk\" (UniqueName: \"kubernetes.io/projected/6ba09de8-56d0-4a49-90fd-0901ef2a4983-kube-api-access-wbmzk\") pod \"community-operators-m5zcg\" (UID: \"6ba09de8-56d0-4a49-90fd-0901ef2a4983\") " pod="openshift-marketplace/community-operators-m5zcg" Jan 21 11:36:40 crc kubenswrapper[4824]: I0121 11:36:40.449603 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6ba09de8-56d0-4a49-90fd-0901ef2a4983-catalog-content\") pod \"community-operators-m5zcg\" (UID: \"6ba09de8-56d0-4a49-90fd-0901ef2a4983\") " pod="openshift-marketplace/community-operators-m5zcg" Jan 21 11:36:40 crc kubenswrapper[4824]: I0121 11:36:40.551041 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6ba09de8-56d0-4a49-90fd-0901ef2a4983-catalog-content\") pod \"community-operators-m5zcg\" (UID: \"6ba09de8-56d0-4a49-90fd-0901ef2a4983\") " pod="openshift-marketplace/community-operators-m5zcg" Jan 21 11:36:40 crc kubenswrapper[4824]: I0121 11:36:40.551201 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6ba09de8-56d0-4a49-90fd-0901ef2a4983-utilities\") pod \"community-operators-m5zcg\" (UID: \"6ba09de8-56d0-4a49-90fd-0901ef2a4983\") " pod="openshift-marketplace/community-operators-m5zcg" Jan 21 11:36:40 crc kubenswrapper[4824]: I0121 11:36:40.551260 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wbmzk\" (UniqueName: \"kubernetes.io/projected/6ba09de8-56d0-4a49-90fd-0901ef2a4983-kube-api-access-wbmzk\") pod \"community-operators-m5zcg\" (UID: \"6ba09de8-56d0-4a49-90fd-0901ef2a4983\") " pod="openshift-marketplace/community-operators-m5zcg" Jan 21 11:36:40 crc kubenswrapper[4824]: I0121 11:36:40.551538 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6ba09de8-56d0-4a49-90fd-0901ef2a4983-catalog-content\") pod \"community-operators-m5zcg\" (UID: \"6ba09de8-56d0-4a49-90fd-0901ef2a4983\") " pod="openshift-marketplace/community-operators-m5zcg" Jan 21 11:36:40 crc kubenswrapper[4824]: I0121 11:36:40.551594 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6ba09de8-56d0-4a49-90fd-0901ef2a4983-utilities\") pod \"community-operators-m5zcg\" (UID: \"6ba09de8-56d0-4a49-90fd-0901ef2a4983\") " pod="openshift-marketplace/community-operators-m5zcg" Jan 21 11:36:40 crc kubenswrapper[4824]: I0121 11:36:40.568655 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wbmzk\" (UniqueName: \"kubernetes.io/projected/6ba09de8-56d0-4a49-90fd-0901ef2a4983-kube-api-access-wbmzk\") pod \"community-operators-m5zcg\" (UID: \"6ba09de8-56d0-4a49-90fd-0901ef2a4983\") " pod="openshift-marketplace/community-operators-m5zcg" Jan 21 11:36:40 crc kubenswrapper[4824]: I0121 11:36:40.613564 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-m5zcg" Jan 21 11:36:41 crc kubenswrapper[4824]: I0121 11:36:41.042703 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-m5zcg"] Jan 21 11:36:42 crc kubenswrapper[4824]: I0121 11:36:42.022318 4824 generic.go:334] "Generic (PLEG): container finished" podID="6ba09de8-56d0-4a49-90fd-0901ef2a4983" containerID="d80af5a4fc04c5587bf037fe2c047264f1d5ab586f1f6333fb85d1ff973ae63a" exitCode=0 Jan 21 11:36:42 crc kubenswrapper[4824]: I0121 11:36:42.022375 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-m5zcg" event={"ID":"6ba09de8-56d0-4a49-90fd-0901ef2a4983","Type":"ContainerDied","Data":"d80af5a4fc04c5587bf037fe2c047264f1d5ab586f1f6333fb85d1ff973ae63a"} Jan 21 11:36:42 crc kubenswrapper[4824]: I0121 11:36:42.022422 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-m5zcg" event={"ID":"6ba09de8-56d0-4a49-90fd-0901ef2a4983","Type":"ContainerStarted","Data":"a182f73b4f377f419a0b4bf8458079f91f332f55dfb874dca67b3bccb1efe150"} Jan 21 11:36:42 crc kubenswrapper[4824]: I0121 11:36:42.049899 4824 scope.go:117] "RemoveContainer" containerID="1cd835fe36ebb6541740e5ab5fd843470d46dfb78975fd968e5cc9aaef462499" Jan 21 11:36:42 crc kubenswrapper[4824]: E0121 11:36:42.050232 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 11:36:43 crc kubenswrapper[4824]: I0121 11:36:43.030347 4824 generic.go:334] "Generic (PLEG): container finished" podID="d672f153-6ef2-432d-829a-db63c8daab69" containerID="7778ba4059495bd1d394bac5a6c5c17aa054fa3ecf50af111d3388b007e74498" exitCode=0 Jan 21 11:36:43 crc kubenswrapper[4824]: I0121 11:36:43.030427 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-5b6wz" event={"ID":"d672f153-6ef2-432d-829a-db63c8daab69","Type":"ContainerDied","Data":"7778ba4059495bd1d394bac5a6c5c17aa054fa3ecf50af111d3388b007e74498"} Jan 21 11:36:43 crc kubenswrapper[4824]: I0121 11:36:43.032421 4824 generic.go:334] "Generic (PLEG): container finished" podID="6ba09de8-56d0-4a49-90fd-0901ef2a4983" containerID="a68b60b5bf8f3a21b953e3adf4b60c09768abd3169f9be0d1fda22fb920230c1" exitCode=0 Jan 21 11:36:43 crc kubenswrapper[4824]: I0121 11:36:43.032456 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-m5zcg" event={"ID":"6ba09de8-56d0-4a49-90fd-0901ef2a4983","Type":"ContainerDied","Data":"a68b60b5bf8f3a21b953e3adf4b60c09768abd3169f9be0d1fda22fb920230c1"} Jan 21 11:36:44 crc kubenswrapper[4824]: I0121 11:36:44.019703 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-6vgh6"] Jan 21 11:36:44 crc kubenswrapper[4824]: I0121 11:36:44.025793 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-6vgh6"] Jan 21 11:36:44 crc kubenswrapper[4824]: I0121 11:36:44.045464 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-m5zcg" event={"ID":"6ba09de8-56d0-4a49-90fd-0901ef2a4983","Type":"ContainerStarted","Data":"10fad9992ea62b761cf9c6209d485dd46387e62e86ec7454ba412c0853f216f8"} Jan 21 11:36:44 crc kubenswrapper[4824]: I0121 11:36:44.056547 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="08b8c363-a8c1-4dfe-8fd1-1af2d055a305" path="/var/lib/kubelet/pods/08b8c363-a8c1-4dfe-8fd1-1af2d055a305/volumes" Jan 21 11:36:44 crc kubenswrapper[4824]: I0121 11:36:44.065159 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-m5zcg" podStartSLOduration=2.596961728 podStartE2EDuration="4.065146825s" podCreationTimestamp="2026-01-21 11:36:40 +0000 UTC" firstStartedPulling="2026-01-21 11:36:42.024217976 +0000 UTC m=+1544.317247268" lastFinishedPulling="2026-01-21 11:36:43.492403073 +0000 UTC m=+1545.785432365" observedRunningTime="2026-01-21 11:36:44.059852029 +0000 UTC m=+1546.352881321" watchObservedRunningTime="2026-01-21 11:36:44.065146825 +0000 UTC m=+1546.358176116" Jan 21 11:36:44 crc kubenswrapper[4824]: I0121 11:36:44.394773 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-5b6wz" Jan 21 11:36:44 crc kubenswrapper[4824]: I0121 11:36:44.518477 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d672f153-6ef2-432d-829a-db63c8daab69-inventory\") pod \"d672f153-6ef2-432d-829a-db63c8daab69\" (UID: \"d672f153-6ef2-432d-829a-db63c8daab69\") " Jan 21 11:36:44 crc kubenswrapper[4824]: I0121 11:36:44.518840 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/d672f153-6ef2-432d-829a-db63c8daab69-ssh-key-openstack-edpm-ipam\") pod \"d672f153-6ef2-432d-829a-db63c8daab69\" (UID: \"d672f153-6ef2-432d-829a-db63c8daab69\") " Jan 21 11:36:44 crc kubenswrapper[4824]: I0121 11:36:44.519201 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-92v6n\" (UniqueName: \"kubernetes.io/projected/d672f153-6ef2-432d-829a-db63c8daab69-kube-api-access-92v6n\") pod \"d672f153-6ef2-432d-829a-db63c8daab69\" (UID: \"d672f153-6ef2-432d-829a-db63c8daab69\") " Jan 21 11:36:44 crc kubenswrapper[4824]: I0121 11:36:44.523054 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d672f153-6ef2-432d-829a-db63c8daab69-kube-api-access-92v6n" (OuterVolumeSpecName: "kube-api-access-92v6n") pod "d672f153-6ef2-432d-829a-db63c8daab69" (UID: "d672f153-6ef2-432d-829a-db63c8daab69"). InnerVolumeSpecName "kube-api-access-92v6n". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:36:44 crc kubenswrapper[4824]: I0121 11:36:44.539172 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d672f153-6ef2-432d-829a-db63c8daab69-inventory" (OuterVolumeSpecName: "inventory") pod "d672f153-6ef2-432d-829a-db63c8daab69" (UID: "d672f153-6ef2-432d-829a-db63c8daab69"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:36:44 crc kubenswrapper[4824]: I0121 11:36:44.540941 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d672f153-6ef2-432d-829a-db63c8daab69-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "d672f153-6ef2-432d-829a-db63c8daab69" (UID: "d672f153-6ef2-432d-829a-db63c8daab69"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:36:44 crc kubenswrapper[4824]: I0121 11:36:44.621611 4824 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/d672f153-6ef2-432d-829a-db63c8daab69-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Jan 21 11:36:44 crc kubenswrapper[4824]: I0121 11:36:44.621646 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-92v6n\" (UniqueName: \"kubernetes.io/projected/d672f153-6ef2-432d-829a-db63c8daab69-kube-api-access-92v6n\") on node \"crc\" DevicePath \"\"" Jan 21 11:36:44 crc kubenswrapper[4824]: I0121 11:36:44.621656 4824 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d672f153-6ef2-432d-829a-db63c8daab69-inventory\") on node \"crc\" DevicePath \"\"" Jan 21 11:36:45 crc kubenswrapper[4824]: I0121 11:36:45.053821 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-5b6wz" event={"ID":"d672f153-6ef2-432d-829a-db63c8daab69","Type":"ContainerDied","Data":"73fe2f366123c47bd83aebfb3c09855e44f3bceec576e057fa2d17a95e4879fb"} Jan 21 11:36:45 crc kubenswrapper[4824]: I0121 11:36:45.053863 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="73fe2f366123c47bd83aebfb3c09855e44f3bceec576e057fa2d17a95e4879fb" Jan 21 11:36:45 crc kubenswrapper[4824]: I0121 11:36:45.054029 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-5b6wz" Jan 21 11:36:45 crc kubenswrapper[4824]: I0121 11:36:45.106650 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-b74gh"] Jan 21 11:36:45 crc kubenswrapper[4824]: E0121 11:36:45.107021 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d672f153-6ef2-432d-829a-db63c8daab69" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Jan 21 11:36:45 crc kubenswrapper[4824]: I0121 11:36:45.107040 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="d672f153-6ef2-432d-829a-db63c8daab69" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Jan 21 11:36:45 crc kubenswrapper[4824]: I0121 11:36:45.107222 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="d672f153-6ef2-432d-829a-db63c8daab69" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Jan 21 11:36:45 crc kubenswrapper[4824]: I0121 11:36:45.107721 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-b74gh" Jan 21 11:36:45 crc kubenswrapper[4824]: I0121 11:36:45.109873 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-tcqk6" Jan 21 11:36:45 crc kubenswrapper[4824]: I0121 11:36:45.110072 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Jan 21 11:36:45 crc kubenswrapper[4824]: I0121 11:36:45.110350 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Jan 21 11:36:45 crc kubenswrapper[4824]: I0121 11:36:45.110498 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Jan 21 11:36:45 crc kubenswrapper[4824]: I0121 11:36:45.114342 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-b74gh"] Jan 21 11:36:45 crc kubenswrapper[4824]: I0121 11:36:45.231615 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7h2jt\" (UniqueName: \"kubernetes.io/projected/60fedbf7-07e6-41be-ada3-2a06f28835e3-kube-api-access-7h2jt\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-b74gh\" (UID: \"60fedbf7-07e6-41be-ada3-2a06f28835e3\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-b74gh" Jan 21 11:36:45 crc kubenswrapper[4824]: I0121 11:36:45.231914 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/60fedbf7-07e6-41be-ada3-2a06f28835e3-ssh-key-openstack-edpm-ipam\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-b74gh\" (UID: \"60fedbf7-07e6-41be-ada3-2a06f28835e3\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-b74gh" Jan 21 11:36:45 crc kubenswrapper[4824]: I0121 11:36:45.232267 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/60fedbf7-07e6-41be-ada3-2a06f28835e3-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-b74gh\" (UID: \"60fedbf7-07e6-41be-ada3-2a06f28835e3\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-b74gh" Jan 21 11:36:45 crc kubenswrapper[4824]: I0121 11:36:45.335170 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/60fedbf7-07e6-41be-ada3-2a06f28835e3-ssh-key-openstack-edpm-ipam\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-b74gh\" (UID: \"60fedbf7-07e6-41be-ada3-2a06f28835e3\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-b74gh" Jan 21 11:36:45 crc kubenswrapper[4824]: I0121 11:36:45.335587 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/60fedbf7-07e6-41be-ada3-2a06f28835e3-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-b74gh\" (UID: \"60fedbf7-07e6-41be-ada3-2a06f28835e3\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-b74gh" Jan 21 11:36:45 crc kubenswrapper[4824]: I0121 11:36:45.336179 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7h2jt\" (UniqueName: \"kubernetes.io/projected/60fedbf7-07e6-41be-ada3-2a06f28835e3-kube-api-access-7h2jt\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-b74gh\" (UID: \"60fedbf7-07e6-41be-ada3-2a06f28835e3\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-b74gh" Jan 21 11:36:45 crc kubenswrapper[4824]: I0121 11:36:45.338539 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/60fedbf7-07e6-41be-ada3-2a06f28835e3-ssh-key-openstack-edpm-ipam\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-b74gh\" (UID: \"60fedbf7-07e6-41be-ada3-2a06f28835e3\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-b74gh" Jan 21 11:36:45 crc kubenswrapper[4824]: I0121 11:36:45.338734 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/60fedbf7-07e6-41be-ada3-2a06f28835e3-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-b74gh\" (UID: \"60fedbf7-07e6-41be-ada3-2a06f28835e3\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-b74gh" Jan 21 11:36:45 crc kubenswrapper[4824]: I0121 11:36:45.350401 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7h2jt\" (UniqueName: \"kubernetes.io/projected/60fedbf7-07e6-41be-ada3-2a06f28835e3-kube-api-access-7h2jt\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-b74gh\" (UID: \"60fedbf7-07e6-41be-ada3-2a06f28835e3\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-b74gh" Jan 21 11:36:45 crc kubenswrapper[4824]: I0121 11:36:45.420590 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-b74gh" Jan 21 11:36:45 crc kubenswrapper[4824]: I0121 11:36:45.851751 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-b74gh"] Jan 21 11:36:45 crc kubenswrapper[4824]: W0121 11:36:45.852472 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod60fedbf7_07e6_41be_ada3_2a06f28835e3.slice/crio-bbe542f5684dd054dc36d25752bfcb215f03fd404bce76fae315f49118b04490 WatchSource:0}: Error finding container bbe542f5684dd054dc36d25752bfcb215f03fd404bce76fae315f49118b04490: Status 404 returned error can't find the container with id bbe542f5684dd054dc36d25752bfcb215f03fd404bce76fae315f49118b04490 Jan 21 11:36:46 crc kubenswrapper[4824]: I0121 11:36:46.026314 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-5l2gg"] Jan 21 11:36:46 crc kubenswrapper[4824]: I0121 11:36:46.032560 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-5l2gg"] Jan 21 11:36:46 crc kubenswrapper[4824]: I0121 11:36:46.057018 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a7e75011-d633-4ccc-951c-d019cb2100f9" path="/var/lib/kubelet/pods/a7e75011-d633-4ccc-951c-d019cb2100f9/volumes" Jan 21 11:36:46 crc kubenswrapper[4824]: I0121 11:36:46.061549 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-b74gh" event={"ID":"60fedbf7-07e6-41be-ada3-2a06f28835e3","Type":"ContainerStarted","Data":"bbe542f5684dd054dc36d25752bfcb215f03fd404bce76fae315f49118b04490"} Jan 21 11:36:47 crc kubenswrapper[4824]: I0121 11:36:47.070670 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-b74gh" event={"ID":"60fedbf7-07e6-41be-ada3-2a06f28835e3","Type":"ContainerStarted","Data":"495175a3ec4152174c9a4e2a8aae14a46ac8d180433f99dc41b0fab262a48f7c"} Jan 21 11:36:47 crc kubenswrapper[4824]: I0121 11:36:47.086885 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-b74gh" podStartSLOduration=1.540283025 podStartE2EDuration="2.08687221s" podCreationTimestamp="2026-01-21 11:36:45 +0000 UTC" firstStartedPulling="2026-01-21 11:36:45.854045762 +0000 UTC m=+1548.147075054" lastFinishedPulling="2026-01-21 11:36:46.400634947 +0000 UTC m=+1548.693664239" observedRunningTime="2026-01-21 11:36:47.081806126 +0000 UTC m=+1549.374835418" watchObservedRunningTime="2026-01-21 11:36:47.08687221 +0000 UTC m=+1549.379901502" Jan 21 11:36:48 crc kubenswrapper[4824]: I0121 11:36:48.019345 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-db-sync-jrsn8"] Jan 21 11:36:48 crc kubenswrapper[4824]: I0121 11:36:48.025309 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-db-sync-jrsn8"] Jan 21 11:36:48 crc kubenswrapper[4824]: I0121 11:36:48.056764 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cadf2d38-24f9-43ab-8e29-803d97a0ff06" path="/var/lib/kubelet/pods/cadf2d38-24f9-43ab-8e29-803d97a0ff06/volumes" Jan 21 11:36:50 crc kubenswrapper[4824]: I0121 11:36:50.613813 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-m5zcg" Jan 21 11:36:50 crc kubenswrapper[4824]: I0121 11:36:50.614070 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-m5zcg" Jan 21 11:36:50 crc kubenswrapper[4824]: I0121 11:36:50.644749 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-m5zcg" Jan 21 11:36:51 crc kubenswrapper[4824]: I0121 11:36:51.126235 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-m5zcg" Jan 21 11:36:51 crc kubenswrapper[4824]: I0121 11:36:51.165228 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-m5zcg"] Jan 21 11:36:53 crc kubenswrapper[4824]: I0121 11:36:53.106344 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-m5zcg" podUID="6ba09de8-56d0-4a49-90fd-0901ef2a4983" containerName="registry-server" containerID="cri-o://10fad9992ea62b761cf9c6209d485dd46387e62e86ec7454ba412c0853f216f8" gracePeriod=2 Jan 21 11:36:53 crc kubenswrapper[4824]: I0121 11:36:53.466308 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-m5zcg" Jan 21 11:36:53 crc kubenswrapper[4824]: I0121 11:36:53.668712 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6ba09de8-56d0-4a49-90fd-0901ef2a4983-catalog-content\") pod \"6ba09de8-56d0-4a49-90fd-0901ef2a4983\" (UID: \"6ba09de8-56d0-4a49-90fd-0901ef2a4983\") " Jan 21 11:36:53 crc kubenswrapper[4824]: I0121 11:36:53.669150 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wbmzk\" (UniqueName: \"kubernetes.io/projected/6ba09de8-56d0-4a49-90fd-0901ef2a4983-kube-api-access-wbmzk\") pod \"6ba09de8-56d0-4a49-90fd-0901ef2a4983\" (UID: \"6ba09de8-56d0-4a49-90fd-0901ef2a4983\") " Jan 21 11:36:53 crc kubenswrapper[4824]: I0121 11:36:53.669218 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6ba09de8-56d0-4a49-90fd-0901ef2a4983-utilities\") pod \"6ba09de8-56d0-4a49-90fd-0901ef2a4983\" (UID: \"6ba09de8-56d0-4a49-90fd-0901ef2a4983\") " Jan 21 11:36:53 crc kubenswrapper[4824]: I0121 11:36:53.669646 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6ba09de8-56d0-4a49-90fd-0901ef2a4983-utilities" (OuterVolumeSpecName: "utilities") pod "6ba09de8-56d0-4a49-90fd-0901ef2a4983" (UID: "6ba09de8-56d0-4a49-90fd-0901ef2a4983"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:36:53 crc kubenswrapper[4824]: I0121 11:36:53.673727 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ba09de8-56d0-4a49-90fd-0901ef2a4983-kube-api-access-wbmzk" (OuterVolumeSpecName: "kube-api-access-wbmzk") pod "6ba09de8-56d0-4a49-90fd-0901ef2a4983" (UID: "6ba09de8-56d0-4a49-90fd-0901ef2a4983"). InnerVolumeSpecName "kube-api-access-wbmzk". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:36:53 crc kubenswrapper[4824]: I0121 11:36:53.707884 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6ba09de8-56d0-4a49-90fd-0901ef2a4983-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6ba09de8-56d0-4a49-90fd-0901ef2a4983" (UID: "6ba09de8-56d0-4a49-90fd-0901ef2a4983"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:36:53 crc kubenswrapper[4824]: I0121 11:36:53.770750 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wbmzk\" (UniqueName: \"kubernetes.io/projected/6ba09de8-56d0-4a49-90fd-0901ef2a4983-kube-api-access-wbmzk\") on node \"crc\" DevicePath \"\"" Jan 21 11:36:53 crc kubenswrapper[4824]: I0121 11:36:53.770775 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6ba09de8-56d0-4a49-90fd-0901ef2a4983-utilities\") on node \"crc\" DevicePath \"\"" Jan 21 11:36:53 crc kubenswrapper[4824]: I0121 11:36:53.770785 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6ba09de8-56d0-4a49-90fd-0901ef2a4983-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 21 11:36:54 crc kubenswrapper[4824]: I0121 11:36:54.114562 4824 generic.go:334] "Generic (PLEG): container finished" podID="6ba09de8-56d0-4a49-90fd-0901ef2a4983" containerID="10fad9992ea62b761cf9c6209d485dd46387e62e86ec7454ba412c0853f216f8" exitCode=0 Jan 21 11:36:54 crc kubenswrapper[4824]: I0121 11:36:54.114610 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-m5zcg" event={"ID":"6ba09de8-56d0-4a49-90fd-0901ef2a4983","Type":"ContainerDied","Data":"10fad9992ea62b761cf9c6209d485dd46387e62e86ec7454ba412c0853f216f8"} Jan 21 11:36:54 crc kubenswrapper[4824]: I0121 11:36:54.114641 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-m5zcg" event={"ID":"6ba09de8-56d0-4a49-90fd-0901ef2a4983","Type":"ContainerDied","Data":"a182f73b4f377f419a0b4bf8458079f91f332f55dfb874dca67b3bccb1efe150"} Jan 21 11:36:54 crc kubenswrapper[4824]: I0121 11:36:54.114658 4824 scope.go:117] "RemoveContainer" containerID="10fad9992ea62b761cf9c6209d485dd46387e62e86ec7454ba412c0853f216f8" Jan 21 11:36:54 crc kubenswrapper[4824]: I0121 11:36:54.115497 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-m5zcg" Jan 21 11:36:54 crc kubenswrapper[4824]: I0121 11:36:54.130942 4824 scope.go:117] "RemoveContainer" containerID="a68b60b5bf8f3a21b953e3adf4b60c09768abd3169f9be0d1fda22fb920230c1" Jan 21 11:36:54 crc kubenswrapper[4824]: I0121 11:36:54.133289 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-m5zcg"] Jan 21 11:36:54 crc kubenswrapper[4824]: I0121 11:36:54.140434 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-m5zcg"] Jan 21 11:36:54 crc kubenswrapper[4824]: I0121 11:36:54.144794 4824 scope.go:117] "RemoveContainer" containerID="d80af5a4fc04c5587bf037fe2c047264f1d5ab586f1f6333fb85d1ff973ae63a" Jan 21 11:36:54 crc kubenswrapper[4824]: I0121 11:36:54.176018 4824 scope.go:117] "RemoveContainer" containerID="10fad9992ea62b761cf9c6209d485dd46387e62e86ec7454ba412c0853f216f8" Jan 21 11:36:54 crc kubenswrapper[4824]: E0121 11:36:54.176329 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"10fad9992ea62b761cf9c6209d485dd46387e62e86ec7454ba412c0853f216f8\": container with ID starting with 10fad9992ea62b761cf9c6209d485dd46387e62e86ec7454ba412c0853f216f8 not found: ID does not exist" containerID="10fad9992ea62b761cf9c6209d485dd46387e62e86ec7454ba412c0853f216f8" Jan 21 11:36:54 crc kubenswrapper[4824]: I0121 11:36:54.176362 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"10fad9992ea62b761cf9c6209d485dd46387e62e86ec7454ba412c0853f216f8"} err="failed to get container status \"10fad9992ea62b761cf9c6209d485dd46387e62e86ec7454ba412c0853f216f8\": rpc error: code = NotFound desc = could not find container \"10fad9992ea62b761cf9c6209d485dd46387e62e86ec7454ba412c0853f216f8\": container with ID starting with 10fad9992ea62b761cf9c6209d485dd46387e62e86ec7454ba412c0853f216f8 not found: ID does not exist" Jan 21 11:36:54 crc kubenswrapper[4824]: I0121 11:36:54.176383 4824 scope.go:117] "RemoveContainer" containerID="a68b60b5bf8f3a21b953e3adf4b60c09768abd3169f9be0d1fda22fb920230c1" Jan 21 11:36:54 crc kubenswrapper[4824]: E0121 11:36:54.176688 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a68b60b5bf8f3a21b953e3adf4b60c09768abd3169f9be0d1fda22fb920230c1\": container with ID starting with a68b60b5bf8f3a21b953e3adf4b60c09768abd3169f9be0d1fda22fb920230c1 not found: ID does not exist" containerID="a68b60b5bf8f3a21b953e3adf4b60c09768abd3169f9be0d1fda22fb920230c1" Jan 21 11:36:54 crc kubenswrapper[4824]: I0121 11:36:54.176710 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a68b60b5bf8f3a21b953e3adf4b60c09768abd3169f9be0d1fda22fb920230c1"} err="failed to get container status \"a68b60b5bf8f3a21b953e3adf4b60c09768abd3169f9be0d1fda22fb920230c1\": rpc error: code = NotFound desc = could not find container \"a68b60b5bf8f3a21b953e3adf4b60c09768abd3169f9be0d1fda22fb920230c1\": container with ID starting with a68b60b5bf8f3a21b953e3adf4b60c09768abd3169f9be0d1fda22fb920230c1 not found: ID does not exist" Jan 21 11:36:54 crc kubenswrapper[4824]: I0121 11:36:54.176725 4824 scope.go:117] "RemoveContainer" containerID="d80af5a4fc04c5587bf037fe2c047264f1d5ab586f1f6333fb85d1ff973ae63a" Jan 21 11:36:54 crc kubenswrapper[4824]: E0121 11:36:54.177218 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d80af5a4fc04c5587bf037fe2c047264f1d5ab586f1f6333fb85d1ff973ae63a\": container with ID starting with d80af5a4fc04c5587bf037fe2c047264f1d5ab586f1f6333fb85d1ff973ae63a not found: ID does not exist" containerID="d80af5a4fc04c5587bf037fe2c047264f1d5ab586f1f6333fb85d1ff973ae63a" Jan 21 11:36:54 crc kubenswrapper[4824]: I0121 11:36:54.177257 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d80af5a4fc04c5587bf037fe2c047264f1d5ab586f1f6333fb85d1ff973ae63a"} err="failed to get container status \"d80af5a4fc04c5587bf037fe2c047264f1d5ab586f1f6333fb85d1ff973ae63a\": rpc error: code = NotFound desc = could not find container \"d80af5a4fc04c5587bf037fe2c047264f1d5ab586f1f6333fb85d1ff973ae63a\": container with ID starting with d80af5a4fc04c5587bf037fe2c047264f1d5ab586f1f6333fb85d1ff973ae63a not found: ID does not exist" Jan 21 11:36:55 crc kubenswrapper[4824]: I0121 11:36:55.049361 4824 scope.go:117] "RemoveContainer" containerID="1cd835fe36ebb6541740e5ab5fd843470d46dfb78975fd968e5cc9aaef462499" Jan 21 11:36:55 crc kubenswrapper[4824]: E0121 11:36:55.049582 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 11:36:56 crc kubenswrapper[4824]: I0121 11:36:56.057570 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ba09de8-56d0-4a49-90fd-0901ef2a4983" path="/var/lib/kubelet/pods/6ba09de8-56d0-4a49-90fd-0901ef2a4983/volumes" Jan 21 11:36:56 crc kubenswrapper[4824]: I0121 11:36:56.479929 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-j5kv8"] Jan 21 11:36:56 crc kubenswrapper[4824]: E0121 11:36:56.480329 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ba09de8-56d0-4a49-90fd-0901ef2a4983" containerName="extract-utilities" Jan 21 11:36:56 crc kubenswrapper[4824]: I0121 11:36:56.480348 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ba09de8-56d0-4a49-90fd-0901ef2a4983" containerName="extract-utilities" Jan 21 11:36:56 crc kubenswrapper[4824]: E0121 11:36:56.480362 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ba09de8-56d0-4a49-90fd-0901ef2a4983" containerName="registry-server" Jan 21 11:36:56 crc kubenswrapper[4824]: I0121 11:36:56.480368 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ba09de8-56d0-4a49-90fd-0901ef2a4983" containerName="registry-server" Jan 21 11:36:56 crc kubenswrapper[4824]: E0121 11:36:56.480382 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ba09de8-56d0-4a49-90fd-0901ef2a4983" containerName="extract-content" Jan 21 11:36:56 crc kubenswrapper[4824]: I0121 11:36:56.480388 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ba09de8-56d0-4a49-90fd-0901ef2a4983" containerName="extract-content" Jan 21 11:36:56 crc kubenswrapper[4824]: I0121 11:36:56.480562 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="6ba09de8-56d0-4a49-90fd-0901ef2a4983" containerName="registry-server" Jan 21 11:36:56 crc kubenswrapper[4824]: I0121 11:36:56.481770 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-j5kv8" Jan 21 11:36:56 crc kubenswrapper[4824]: I0121 11:36:56.487581 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-j5kv8"] Jan 21 11:36:56 crc kubenswrapper[4824]: I0121 11:36:56.515782 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5633f007-e1df-4760-b397-db8d77a2c43e-catalog-content\") pod \"redhat-operators-j5kv8\" (UID: \"5633f007-e1df-4760-b397-db8d77a2c43e\") " pod="openshift-marketplace/redhat-operators-j5kv8" Jan 21 11:36:56 crc kubenswrapper[4824]: I0121 11:36:56.515816 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5633f007-e1df-4760-b397-db8d77a2c43e-utilities\") pod \"redhat-operators-j5kv8\" (UID: \"5633f007-e1df-4760-b397-db8d77a2c43e\") " pod="openshift-marketplace/redhat-operators-j5kv8" Jan 21 11:36:56 crc kubenswrapper[4824]: I0121 11:36:56.515875 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9rqjc\" (UniqueName: \"kubernetes.io/projected/5633f007-e1df-4760-b397-db8d77a2c43e-kube-api-access-9rqjc\") pod \"redhat-operators-j5kv8\" (UID: \"5633f007-e1df-4760-b397-db8d77a2c43e\") " pod="openshift-marketplace/redhat-operators-j5kv8" Jan 21 11:36:56 crc kubenswrapper[4824]: I0121 11:36:56.617349 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5633f007-e1df-4760-b397-db8d77a2c43e-catalog-content\") pod \"redhat-operators-j5kv8\" (UID: \"5633f007-e1df-4760-b397-db8d77a2c43e\") " pod="openshift-marketplace/redhat-operators-j5kv8" Jan 21 11:36:56 crc kubenswrapper[4824]: I0121 11:36:56.617381 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5633f007-e1df-4760-b397-db8d77a2c43e-utilities\") pod \"redhat-operators-j5kv8\" (UID: \"5633f007-e1df-4760-b397-db8d77a2c43e\") " pod="openshift-marketplace/redhat-operators-j5kv8" Jan 21 11:36:56 crc kubenswrapper[4824]: I0121 11:36:56.617437 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9rqjc\" (UniqueName: \"kubernetes.io/projected/5633f007-e1df-4760-b397-db8d77a2c43e-kube-api-access-9rqjc\") pod \"redhat-operators-j5kv8\" (UID: \"5633f007-e1df-4760-b397-db8d77a2c43e\") " pod="openshift-marketplace/redhat-operators-j5kv8" Jan 21 11:36:56 crc kubenswrapper[4824]: I0121 11:36:56.617795 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5633f007-e1df-4760-b397-db8d77a2c43e-catalog-content\") pod \"redhat-operators-j5kv8\" (UID: \"5633f007-e1df-4760-b397-db8d77a2c43e\") " pod="openshift-marketplace/redhat-operators-j5kv8" Jan 21 11:36:56 crc kubenswrapper[4824]: I0121 11:36:56.617816 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5633f007-e1df-4760-b397-db8d77a2c43e-utilities\") pod \"redhat-operators-j5kv8\" (UID: \"5633f007-e1df-4760-b397-db8d77a2c43e\") " pod="openshift-marketplace/redhat-operators-j5kv8" Jan 21 11:36:56 crc kubenswrapper[4824]: I0121 11:36:56.632361 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9rqjc\" (UniqueName: \"kubernetes.io/projected/5633f007-e1df-4760-b397-db8d77a2c43e-kube-api-access-9rqjc\") pod \"redhat-operators-j5kv8\" (UID: \"5633f007-e1df-4760-b397-db8d77a2c43e\") " pod="openshift-marketplace/redhat-operators-j5kv8" Jan 21 11:36:56 crc kubenswrapper[4824]: I0121 11:36:56.807850 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-j5kv8" Jan 21 11:36:57 crc kubenswrapper[4824]: I0121 11:36:57.224675 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-j5kv8"] Jan 21 11:36:58 crc kubenswrapper[4824]: I0121 11:36:58.142104 4824 generic.go:334] "Generic (PLEG): container finished" podID="5633f007-e1df-4760-b397-db8d77a2c43e" containerID="e0d42905860419b0009413ac290505728fdbe1fd0d99cf72d3a49305ae5fd29d" exitCode=0 Jan 21 11:36:58 crc kubenswrapper[4824]: I0121 11:36:58.142169 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j5kv8" event={"ID":"5633f007-e1df-4760-b397-db8d77a2c43e","Type":"ContainerDied","Data":"e0d42905860419b0009413ac290505728fdbe1fd0d99cf72d3a49305ae5fd29d"} Jan 21 11:36:58 crc kubenswrapper[4824]: I0121 11:36:58.142321 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j5kv8" event={"ID":"5633f007-e1df-4760-b397-db8d77a2c43e","Type":"ContainerStarted","Data":"9730f16fe571b01572a5f22435b85174e662bfa2a84a0568a751de97d620c25c"} Jan 21 11:36:59 crc kubenswrapper[4824]: I0121 11:36:59.596399 4824 scope.go:117] "RemoveContainer" containerID="2176bc084f2009e4d942790b8c3474f286b639f648b347b28da26c0a5a3ccec2" Jan 21 11:36:59 crc kubenswrapper[4824]: I0121 11:36:59.628053 4824 scope.go:117] "RemoveContainer" containerID="49abf3af76b6411b5212e8cfa13e8c4145a6700e1d310d3281c5d847c1a1f457" Jan 21 11:36:59 crc kubenswrapper[4824]: I0121 11:36:59.663006 4824 scope.go:117] "RemoveContainer" containerID="81f52e55093614da3aed12f82750111917b47ee8401eb58ccc0da58488e3019f" Jan 21 11:36:59 crc kubenswrapper[4824]: I0121 11:36:59.683373 4824 scope.go:117] "RemoveContainer" containerID="45d2f067a75588f2f74c579f26d8f9af2b6a872de4eb9d340b34e9fee4726547" Jan 21 11:36:59 crc kubenswrapper[4824]: I0121 11:36:59.716479 4824 scope.go:117] "RemoveContainer" containerID="e12a5e35428ef122a88a06ec7747afe8e712c9a7807b731859f6ec595579f719" Jan 21 11:36:59 crc kubenswrapper[4824]: I0121 11:36:59.751702 4824 scope.go:117] "RemoveContainer" containerID="2534da156868c09c8dd8dee0424c34fab5c7b1401695a79ef623bde9f1ef00f0" Jan 21 11:36:59 crc kubenswrapper[4824]: I0121 11:36:59.781722 4824 scope.go:117] "RemoveContainer" containerID="3d07104d28aceb5ad31acec840c18d559f5c98c4390845b748361ff21f72d204" Jan 21 11:36:59 crc kubenswrapper[4824]: I0121 11:36:59.808032 4824 scope.go:117] "RemoveContainer" containerID="cc788777a506818eef849722d9795fcf44c54eb36cb8cfbc69470609fce80bd0" Jan 21 11:36:59 crc kubenswrapper[4824]: I0121 11:36:59.822147 4824 scope.go:117] "RemoveContainer" containerID="5dcf0246f4ab2f2f4f3230d6a47ec6ac1102fcb34b4edb4d89a73c99d792f5ff" Jan 21 11:36:59 crc kubenswrapper[4824]: I0121 11:36:59.846032 4824 scope.go:117] "RemoveContainer" containerID="5f4a1e6bcd0c138e59cd1e69d47cce9046bae28ab6b9ce57bd7be2ad0c6a4e61" Jan 21 11:36:59 crc kubenswrapper[4824]: I0121 11:36:59.865161 4824 scope.go:117] "RemoveContainer" containerID="c1a92fb4372889a2ab5383adea30e43a1fa74caae3039082b2ef5482212d2758" Jan 21 11:36:59 crc kubenswrapper[4824]: I0121 11:36:59.880623 4824 scope.go:117] "RemoveContainer" containerID="b6f52bcc83bf92c2ef07afe3fa11fe445d957e311fc35ba4f1767a43394658f9" Jan 21 11:36:59 crc kubenswrapper[4824]: I0121 11:36:59.895429 4824 scope.go:117] "RemoveContainer" containerID="f658b544252ce5134cf2b3d03b008e5a8cf0c0a33b02b6a5554a5288bb79bbfe" Jan 21 11:36:59 crc kubenswrapper[4824]: I0121 11:36:59.909618 4824 scope.go:117] "RemoveContainer" containerID="25e2c38241236d848b515e556d4747f5cb7cbca9e5eba1926722e4f2e38b424d" Jan 21 11:36:59 crc kubenswrapper[4824]: I0121 11:36:59.922127 4824 scope.go:117] "RemoveContainer" containerID="2b4aa0578cb659bd1fe3cb0351e682b347beae954c60ccf215edaf9c2da1abd4" Jan 21 11:37:00 crc kubenswrapper[4824]: I0121 11:37:00.027358 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-ngq69"] Jan 21 11:37:00 crc kubenswrapper[4824]: I0121 11:37:00.034618 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-rlp2t"] Jan 21 11:37:00 crc kubenswrapper[4824]: I0121 11:37:00.040672 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-ngq69"] Jan 21 11:37:00 crc kubenswrapper[4824]: I0121 11:37:00.046372 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-rlp2t"] Jan 21 11:37:00 crc kubenswrapper[4824]: I0121 11:37:00.057264 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0a6c0628-4522-4bb7-8a82-cc2e019eca2a" path="/var/lib/kubelet/pods/0a6c0628-4522-4bb7-8a82-cc2e019eca2a/volumes" Jan 21 11:37:00 crc kubenswrapper[4824]: I0121 11:37:00.058001 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="262733b9-4179-49e3-aee9-b62197cc89ba" path="/var/lib/kubelet/pods/262733b9-4179-49e3-aee9-b62197cc89ba/volumes" Jan 21 11:37:00 crc kubenswrapper[4824]: I0121 11:37:00.155930 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j5kv8" event={"ID":"5633f007-e1df-4760-b397-db8d77a2c43e","Type":"ContainerStarted","Data":"35c46439bf62c0ed1bc145d1bc4f703e7629a47f9ed6b218029824d204f3e00c"} Jan 21 11:37:02 crc kubenswrapper[4824]: I0121 11:37:02.170297 4824 generic.go:334] "Generic (PLEG): container finished" podID="5633f007-e1df-4760-b397-db8d77a2c43e" containerID="35c46439bf62c0ed1bc145d1bc4f703e7629a47f9ed6b218029824d204f3e00c" exitCode=0 Jan 21 11:37:02 crc kubenswrapper[4824]: I0121 11:37:02.170390 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j5kv8" event={"ID":"5633f007-e1df-4760-b397-db8d77a2c43e","Type":"ContainerDied","Data":"35c46439bf62c0ed1bc145d1bc4f703e7629a47f9ed6b218029824d204f3e00c"} Jan 21 11:37:03 crc kubenswrapper[4824]: I0121 11:37:03.179136 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j5kv8" event={"ID":"5633f007-e1df-4760-b397-db8d77a2c43e","Type":"ContainerStarted","Data":"843f54d9af678173aef81e708b00180ebf68828556bd2cd3ac5845fbe7b50c31"} Jan 21 11:37:03 crc kubenswrapper[4824]: I0121 11:37:03.198284 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-j5kv8" podStartSLOduration=2.641608016 podStartE2EDuration="7.198270088s" podCreationTimestamp="2026-01-21 11:36:56 +0000 UTC" firstStartedPulling="2026-01-21 11:36:58.145689972 +0000 UTC m=+1560.438719254" lastFinishedPulling="2026-01-21 11:37:02.702352034 +0000 UTC m=+1564.995381326" observedRunningTime="2026-01-21 11:37:03.198088274 +0000 UTC m=+1565.491117567" watchObservedRunningTime="2026-01-21 11:37:03.198270088 +0000 UTC m=+1565.491299380" Jan 21 11:37:06 crc kubenswrapper[4824]: I0121 11:37:06.808412 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-j5kv8" Jan 21 11:37:06 crc kubenswrapper[4824]: I0121 11:37:06.808640 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-j5kv8" Jan 21 11:37:06 crc kubenswrapper[4824]: I0121 11:37:06.841686 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-j5kv8" Jan 21 11:37:07 crc kubenswrapper[4824]: I0121 11:37:07.232541 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-j5kv8" Jan 21 11:37:07 crc kubenswrapper[4824]: I0121 11:37:07.271597 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-j5kv8"] Jan 21 11:37:09 crc kubenswrapper[4824]: I0121 11:37:09.213939 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-j5kv8" podUID="5633f007-e1df-4760-b397-db8d77a2c43e" containerName="registry-server" containerID="cri-o://843f54d9af678173aef81e708b00180ebf68828556bd2cd3ac5845fbe7b50c31" gracePeriod=2 Jan 21 11:37:09 crc kubenswrapper[4824]: I0121 11:37:09.570372 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-j5kv8" Jan 21 11:37:09 crc kubenswrapper[4824]: I0121 11:37:09.637181 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9rqjc\" (UniqueName: \"kubernetes.io/projected/5633f007-e1df-4760-b397-db8d77a2c43e-kube-api-access-9rqjc\") pod \"5633f007-e1df-4760-b397-db8d77a2c43e\" (UID: \"5633f007-e1df-4760-b397-db8d77a2c43e\") " Jan 21 11:37:09 crc kubenswrapper[4824]: I0121 11:37:09.637277 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5633f007-e1df-4760-b397-db8d77a2c43e-utilities\") pod \"5633f007-e1df-4760-b397-db8d77a2c43e\" (UID: \"5633f007-e1df-4760-b397-db8d77a2c43e\") " Jan 21 11:37:09 crc kubenswrapper[4824]: I0121 11:37:09.637448 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5633f007-e1df-4760-b397-db8d77a2c43e-catalog-content\") pod \"5633f007-e1df-4760-b397-db8d77a2c43e\" (UID: \"5633f007-e1df-4760-b397-db8d77a2c43e\") " Jan 21 11:37:09 crc kubenswrapper[4824]: I0121 11:37:09.637912 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5633f007-e1df-4760-b397-db8d77a2c43e-utilities" (OuterVolumeSpecName: "utilities") pod "5633f007-e1df-4760-b397-db8d77a2c43e" (UID: "5633f007-e1df-4760-b397-db8d77a2c43e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:37:09 crc kubenswrapper[4824]: I0121 11:37:09.641463 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5633f007-e1df-4760-b397-db8d77a2c43e-kube-api-access-9rqjc" (OuterVolumeSpecName: "kube-api-access-9rqjc") pod "5633f007-e1df-4760-b397-db8d77a2c43e" (UID: "5633f007-e1df-4760-b397-db8d77a2c43e"). InnerVolumeSpecName "kube-api-access-9rqjc". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:37:09 crc kubenswrapper[4824]: I0121 11:37:09.724568 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5633f007-e1df-4760-b397-db8d77a2c43e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5633f007-e1df-4760-b397-db8d77a2c43e" (UID: "5633f007-e1df-4760-b397-db8d77a2c43e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:37:09 crc kubenswrapper[4824]: I0121 11:37:09.739006 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5633f007-e1df-4760-b397-db8d77a2c43e-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 21 11:37:09 crc kubenswrapper[4824]: I0121 11:37:09.739029 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9rqjc\" (UniqueName: \"kubernetes.io/projected/5633f007-e1df-4760-b397-db8d77a2c43e-kube-api-access-9rqjc\") on node \"crc\" DevicePath \"\"" Jan 21 11:37:09 crc kubenswrapper[4824]: I0121 11:37:09.739051 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5633f007-e1df-4760-b397-db8d77a2c43e-utilities\") on node \"crc\" DevicePath \"\"" Jan 21 11:37:10 crc kubenswrapper[4824]: I0121 11:37:10.049776 4824 scope.go:117] "RemoveContainer" containerID="1cd835fe36ebb6541740e5ab5fd843470d46dfb78975fd968e5cc9aaef462499" Jan 21 11:37:10 crc kubenswrapper[4824]: E0121 11:37:10.050460 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 11:37:10 crc kubenswrapper[4824]: I0121 11:37:10.223198 4824 generic.go:334] "Generic (PLEG): container finished" podID="5633f007-e1df-4760-b397-db8d77a2c43e" containerID="843f54d9af678173aef81e708b00180ebf68828556bd2cd3ac5845fbe7b50c31" exitCode=0 Jan 21 11:37:10 crc kubenswrapper[4824]: I0121 11:37:10.223233 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j5kv8" event={"ID":"5633f007-e1df-4760-b397-db8d77a2c43e","Type":"ContainerDied","Data":"843f54d9af678173aef81e708b00180ebf68828556bd2cd3ac5845fbe7b50c31"} Jan 21 11:37:10 crc kubenswrapper[4824]: I0121 11:37:10.223254 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j5kv8" event={"ID":"5633f007-e1df-4760-b397-db8d77a2c43e","Type":"ContainerDied","Data":"9730f16fe571b01572a5f22435b85174e662bfa2a84a0568a751de97d620c25c"} Jan 21 11:37:10 crc kubenswrapper[4824]: I0121 11:37:10.223268 4824 scope.go:117] "RemoveContainer" containerID="843f54d9af678173aef81e708b00180ebf68828556bd2cd3ac5845fbe7b50c31" Jan 21 11:37:10 crc kubenswrapper[4824]: I0121 11:37:10.223368 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-j5kv8" Jan 21 11:37:10 crc kubenswrapper[4824]: I0121 11:37:10.247566 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-j5kv8"] Jan 21 11:37:10 crc kubenswrapper[4824]: I0121 11:37:10.254254 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-j5kv8"] Jan 21 11:37:10 crc kubenswrapper[4824]: I0121 11:37:10.257692 4824 scope.go:117] "RemoveContainer" containerID="35c46439bf62c0ed1bc145d1bc4f703e7629a47f9ed6b218029824d204f3e00c" Jan 21 11:37:10 crc kubenswrapper[4824]: I0121 11:37:10.274689 4824 scope.go:117] "RemoveContainer" containerID="e0d42905860419b0009413ac290505728fdbe1fd0d99cf72d3a49305ae5fd29d" Jan 21 11:37:10 crc kubenswrapper[4824]: I0121 11:37:10.304873 4824 scope.go:117] "RemoveContainer" containerID="843f54d9af678173aef81e708b00180ebf68828556bd2cd3ac5845fbe7b50c31" Jan 21 11:37:10 crc kubenswrapper[4824]: E0121 11:37:10.305366 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"843f54d9af678173aef81e708b00180ebf68828556bd2cd3ac5845fbe7b50c31\": container with ID starting with 843f54d9af678173aef81e708b00180ebf68828556bd2cd3ac5845fbe7b50c31 not found: ID does not exist" containerID="843f54d9af678173aef81e708b00180ebf68828556bd2cd3ac5845fbe7b50c31" Jan 21 11:37:10 crc kubenswrapper[4824]: I0121 11:37:10.305399 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"843f54d9af678173aef81e708b00180ebf68828556bd2cd3ac5845fbe7b50c31"} err="failed to get container status \"843f54d9af678173aef81e708b00180ebf68828556bd2cd3ac5845fbe7b50c31\": rpc error: code = NotFound desc = could not find container \"843f54d9af678173aef81e708b00180ebf68828556bd2cd3ac5845fbe7b50c31\": container with ID starting with 843f54d9af678173aef81e708b00180ebf68828556bd2cd3ac5845fbe7b50c31 not found: ID does not exist" Jan 21 11:37:10 crc kubenswrapper[4824]: I0121 11:37:10.305437 4824 scope.go:117] "RemoveContainer" containerID="35c46439bf62c0ed1bc145d1bc4f703e7629a47f9ed6b218029824d204f3e00c" Jan 21 11:37:10 crc kubenswrapper[4824]: E0121 11:37:10.305721 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"35c46439bf62c0ed1bc145d1bc4f703e7629a47f9ed6b218029824d204f3e00c\": container with ID starting with 35c46439bf62c0ed1bc145d1bc4f703e7629a47f9ed6b218029824d204f3e00c not found: ID does not exist" containerID="35c46439bf62c0ed1bc145d1bc4f703e7629a47f9ed6b218029824d204f3e00c" Jan 21 11:37:10 crc kubenswrapper[4824]: I0121 11:37:10.305746 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"35c46439bf62c0ed1bc145d1bc4f703e7629a47f9ed6b218029824d204f3e00c"} err="failed to get container status \"35c46439bf62c0ed1bc145d1bc4f703e7629a47f9ed6b218029824d204f3e00c\": rpc error: code = NotFound desc = could not find container \"35c46439bf62c0ed1bc145d1bc4f703e7629a47f9ed6b218029824d204f3e00c\": container with ID starting with 35c46439bf62c0ed1bc145d1bc4f703e7629a47f9ed6b218029824d204f3e00c not found: ID does not exist" Jan 21 11:37:10 crc kubenswrapper[4824]: I0121 11:37:10.305760 4824 scope.go:117] "RemoveContainer" containerID="e0d42905860419b0009413ac290505728fdbe1fd0d99cf72d3a49305ae5fd29d" Jan 21 11:37:10 crc kubenswrapper[4824]: E0121 11:37:10.305939 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e0d42905860419b0009413ac290505728fdbe1fd0d99cf72d3a49305ae5fd29d\": container with ID starting with e0d42905860419b0009413ac290505728fdbe1fd0d99cf72d3a49305ae5fd29d not found: ID does not exist" containerID="e0d42905860419b0009413ac290505728fdbe1fd0d99cf72d3a49305ae5fd29d" Jan 21 11:37:10 crc kubenswrapper[4824]: I0121 11:37:10.305976 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e0d42905860419b0009413ac290505728fdbe1fd0d99cf72d3a49305ae5fd29d"} err="failed to get container status \"e0d42905860419b0009413ac290505728fdbe1fd0d99cf72d3a49305ae5fd29d\": rpc error: code = NotFound desc = could not find container \"e0d42905860419b0009413ac290505728fdbe1fd0d99cf72d3a49305ae5fd29d\": container with ID starting with e0d42905860419b0009413ac290505728fdbe1fd0d99cf72d3a49305ae5fd29d not found: ID does not exist" Jan 21 11:37:12 crc kubenswrapper[4824]: I0121 11:37:12.057916 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5633f007-e1df-4760-b397-db8d77a2c43e" path="/var/lib/kubelet/pods/5633f007-e1df-4760-b397-db8d77a2c43e/volumes" Jan 21 11:37:15 crc kubenswrapper[4824]: I0121 11:37:15.258214 4824 generic.go:334] "Generic (PLEG): container finished" podID="60fedbf7-07e6-41be-ada3-2a06f28835e3" containerID="495175a3ec4152174c9a4e2a8aae14a46ac8d180433f99dc41b0fab262a48f7c" exitCode=0 Jan 21 11:37:15 crc kubenswrapper[4824]: I0121 11:37:15.258263 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-b74gh" event={"ID":"60fedbf7-07e6-41be-ada3-2a06f28835e3","Type":"ContainerDied","Data":"495175a3ec4152174c9a4e2a8aae14a46ac8d180433f99dc41b0fab262a48f7c"} Jan 21 11:37:16 crc kubenswrapper[4824]: I0121 11:37:16.561947 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-b74gh" Jan 21 11:37:16 crc kubenswrapper[4824]: I0121 11:37:16.657053 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/60fedbf7-07e6-41be-ada3-2a06f28835e3-inventory\") pod \"60fedbf7-07e6-41be-ada3-2a06f28835e3\" (UID: \"60fedbf7-07e6-41be-ada3-2a06f28835e3\") " Jan 21 11:37:16 crc kubenswrapper[4824]: I0121 11:37:16.657127 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/60fedbf7-07e6-41be-ada3-2a06f28835e3-ssh-key-openstack-edpm-ipam\") pod \"60fedbf7-07e6-41be-ada3-2a06f28835e3\" (UID: \"60fedbf7-07e6-41be-ada3-2a06f28835e3\") " Jan 21 11:37:16 crc kubenswrapper[4824]: I0121 11:37:16.657285 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7h2jt\" (UniqueName: \"kubernetes.io/projected/60fedbf7-07e6-41be-ada3-2a06f28835e3-kube-api-access-7h2jt\") pod \"60fedbf7-07e6-41be-ada3-2a06f28835e3\" (UID: \"60fedbf7-07e6-41be-ada3-2a06f28835e3\") " Jan 21 11:37:16 crc kubenswrapper[4824]: I0121 11:37:16.661437 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/60fedbf7-07e6-41be-ada3-2a06f28835e3-kube-api-access-7h2jt" (OuterVolumeSpecName: "kube-api-access-7h2jt") pod "60fedbf7-07e6-41be-ada3-2a06f28835e3" (UID: "60fedbf7-07e6-41be-ada3-2a06f28835e3"). InnerVolumeSpecName "kube-api-access-7h2jt". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:37:16 crc kubenswrapper[4824]: I0121 11:37:16.677571 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/60fedbf7-07e6-41be-ada3-2a06f28835e3-inventory" (OuterVolumeSpecName: "inventory") pod "60fedbf7-07e6-41be-ada3-2a06f28835e3" (UID: "60fedbf7-07e6-41be-ada3-2a06f28835e3"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:37:16 crc kubenswrapper[4824]: I0121 11:37:16.678212 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/60fedbf7-07e6-41be-ada3-2a06f28835e3-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "60fedbf7-07e6-41be-ada3-2a06f28835e3" (UID: "60fedbf7-07e6-41be-ada3-2a06f28835e3"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:37:16 crc kubenswrapper[4824]: I0121 11:37:16.759058 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7h2jt\" (UniqueName: \"kubernetes.io/projected/60fedbf7-07e6-41be-ada3-2a06f28835e3-kube-api-access-7h2jt\") on node \"crc\" DevicePath \"\"" Jan 21 11:37:16 crc kubenswrapper[4824]: I0121 11:37:16.759096 4824 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/60fedbf7-07e6-41be-ada3-2a06f28835e3-inventory\") on node \"crc\" DevicePath \"\"" Jan 21 11:37:16 crc kubenswrapper[4824]: I0121 11:37:16.759105 4824 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/60fedbf7-07e6-41be-ada3-2a06f28835e3-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Jan 21 11:37:17 crc kubenswrapper[4824]: I0121 11:37:17.271373 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-b74gh" event={"ID":"60fedbf7-07e6-41be-ada3-2a06f28835e3","Type":"ContainerDied","Data":"bbe542f5684dd054dc36d25752bfcb215f03fd404bce76fae315f49118b04490"} Jan 21 11:37:17 crc kubenswrapper[4824]: I0121 11:37:17.271645 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bbe542f5684dd054dc36d25752bfcb215f03fd404bce76fae315f49118b04490" Jan 21 11:37:17 crc kubenswrapper[4824]: I0121 11:37:17.271402 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-b74gh" Jan 21 11:37:17 crc kubenswrapper[4824]: I0121 11:37:17.403634 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-4bd5n"] Jan 21 11:37:17 crc kubenswrapper[4824]: E0121 11:37:17.404041 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5633f007-e1df-4760-b397-db8d77a2c43e" containerName="extract-utilities" Jan 21 11:37:17 crc kubenswrapper[4824]: I0121 11:37:17.404058 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="5633f007-e1df-4760-b397-db8d77a2c43e" containerName="extract-utilities" Jan 21 11:37:17 crc kubenswrapper[4824]: E0121 11:37:17.404081 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5633f007-e1df-4760-b397-db8d77a2c43e" containerName="extract-content" Jan 21 11:37:17 crc kubenswrapper[4824]: I0121 11:37:17.404088 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="5633f007-e1df-4760-b397-db8d77a2c43e" containerName="extract-content" Jan 21 11:37:17 crc kubenswrapper[4824]: E0121 11:37:17.404098 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="60fedbf7-07e6-41be-ada3-2a06f28835e3" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Jan 21 11:37:17 crc kubenswrapper[4824]: I0121 11:37:17.404106 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="60fedbf7-07e6-41be-ada3-2a06f28835e3" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Jan 21 11:37:17 crc kubenswrapper[4824]: E0121 11:37:17.404140 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5633f007-e1df-4760-b397-db8d77a2c43e" containerName="registry-server" Jan 21 11:37:17 crc kubenswrapper[4824]: I0121 11:37:17.404146 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="5633f007-e1df-4760-b397-db8d77a2c43e" containerName="registry-server" Jan 21 11:37:17 crc kubenswrapper[4824]: I0121 11:37:17.404320 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="60fedbf7-07e6-41be-ada3-2a06f28835e3" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Jan 21 11:37:17 crc kubenswrapper[4824]: I0121 11:37:17.404341 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="5633f007-e1df-4760-b397-db8d77a2c43e" containerName="registry-server" Jan 21 11:37:17 crc kubenswrapper[4824]: I0121 11:37:17.404873 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-4bd5n" Jan 21 11:37:17 crc kubenswrapper[4824]: I0121 11:37:17.406439 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Jan 21 11:37:17 crc kubenswrapper[4824]: I0121 11:37:17.406745 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Jan 21 11:37:17 crc kubenswrapper[4824]: I0121 11:37:17.406871 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-tcqk6" Jan 21 11:37:17 crc kubenswrapper[4824]: I0121 11:37:17.407095 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Jan 21 11:37:17 crc kubenswrapper[4824]: I0121 11:37:17.409843 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-4bd5n"] Jan 21 11:37:17 crc kubenswrapper[4824]: I0121 11:37:17.469375 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4wqvn\" (UniqueName: \"kubernetes.io/projected/8d46d1eb-4230-49c7-8376-698e7689ad2b-kube-api-access-4wqvn\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-4bd5n\" (UID: \"8d46d1eb-4230-49c7-8376-698e7689ad2b\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-4bd5n" Jan 21 11:37:17 crc kubenswrapper[4824]: I0121 11:37:17.469442 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/8d46d1eb-4230-49c7-8376-698e7689ad2b-ssh-key-openstack-edpm-ipam\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-4bd5n\" (UID: \"8d46d1eb-4230-49c7-8376-698e7689ad2b\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-4bd5n" Jan 21 11:37:17 crc kubenswrapper[4824]: I0121 11:37:17.469582 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8d46d1eb-4230-49c7-8376-698e7689ad2b-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-4bd5n\" (UID: \"8d46d1eb-4230-49c7-8376-698e7689ad2b\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-4bd5n" Jan 21 11:37:17 crc kubenswrapper[4824]: I0121 11:37:17.571234 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/8d46d1eb-4230-49c7-8376-698e7689ad2b-ssh-key-openstack-edpm-ipam\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-4bd5n\" (UID: \"8d46d1eb-4230-49c7-8376-698e7689ad2b\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-4bd5n" Jan 21 11:37:17 crc kubenswrapper[4824]: I0121 11:37:17.571399 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8d46d1eb-4230-49c7-8376-698e7689ad2b-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-4bd5n\" (UID: \"8d46d1eb-4230-49c7-8376-698e7689ad2b\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-4bd5n" Jan 21 11:37:17 crc kubenswrapper[4824]: I0121 11:37:17.571536 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4wqvn\" (UniqueName: \"kubernetes.io/projected/8d46d1eb-4230-49c7-8376-698e7689ad2b-kube-api-access-4wqvn\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-4bd5n\" (UID: \"8d46d1eb-4230-49c7-8376-698e7689ad2b\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-4bd5n" Jan 21 11:37:17 crc kubenswrapper[4824]: I0121 11:37:17.575742 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8d46d1eb-4230-49c7-8376-698e7689ad2b-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-4bd5n\" (UID: \"8d46d1eb-4230-49c7-8376-698e7689ad2b\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-4bd5n" Jan 21 11:37:17 crc kubenswrapper[4824]: I0121 11:37:17.576528 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/8d46d1eb-4230-49c7-8376-698e7689ad2b-ssh-key-openstack-edpm-ipam\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-4bd5n\" (UID: \"8d46d1eb-4230-49c7-8376-698e7689ad2b\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-4bd5n" Jan 21 11:37:17 crc kubenswrapper[4824]: I0121 11:37:17.584665 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4wqvn\" (UniqueName: \"kubernetes.io/projected/8d46d1eb-4230-49c7-8376-698e7689ad2b-kube-api-access-4wqvn\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-4bd5n\" (UID: \"8d46d1eb-4230-49c7-8376-698e7689ad2b\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-4bd5n" Jan 21 11:37:17 crc kubenswrapper[4824]: I0121 11:37:17.728516 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-4bd5n" Jan 21 11:37:18 crc kubenswrapper[4824]: I0121 11:37:18.176009 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-4bd5n"] Jan 21 11:37:18 crc kubenswrapper[4824]: I0121 11:37:18.280005 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-4bd5n" event={"ID":"8d46d1eb-4230-49c7-8376-698e7689ad2b","Type":"ContainerStarted","Data":"7c498c4871058e9b8aef05f8de415d1229252f8805909c3848b5ff58ff925394"} Jan 21 11:37:18 crc kubenswrapper[4824]: I0121 11:37:18.551782 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-rpk5p"] Jan 21 11:37:18 crc kubenswrapper[4824]: I0121 11:37:18.553577 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rpk5p" Jan 21 11:37:18 crc kubenswrapper[4824]: I0121 11:37:18.561114 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-rpk5p"] Jan 21 11:37:18 crc kubenswrapper[4824]: I0121 11:37:18.586948 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/16274e36-c051-4ef8-9120-07b9996f43d6-catalog-content\") pod \"certified-operators-rpk5p\" (UID: \"16274e36-c051-4ef8-9120-07b9996f43d6\") " pod="openshift-marketplace/certified-operators-rpk5p" Jan 21 11:37:18 crc kubenswrapper[4824]: I0121 11:37:18.587079 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zbn97\" (UniqueName: \"kubernetes.io/projected/16274e36-c051-4ef8-9120-07b9996f43d6-kube-api-access-zbn97\") pod \"certified-operators-rpk5p\" (UID: \"16274e36-c051-4ef8-9120-07b9996f43d6\") " pod="openshift-marketplace/certified-operators-rpk5p" Jan 21 11:37:18 crc kubenswrapper[4824]: I0121 11:37:18.587138 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/16274e36-c051-4ef8-9120-07b9996f43d6-utilities\") pod \"certified-operators-rpk5p\" (UID: \"16274e36-c051-4ef8-9120-07b9996f43d6\") " pod="openshift-marketplace/certified-operators-rpk5p" Jan 21 11:37:18 crc kubenswrapper[4824]: I0121 11:37:18.688405 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/16274e36-c051-4ef8-9120-07b9996f43d6-catalog-content\") pod \"certified-operators-rpk5p\" (UID: \"16274e36-c051-4ef8-9120-07b9996f43d6\") " pod="openshift-marketplace/certified-operators-rpk5p" Jan 21 11:37:18 crc kubenswrapper[4824]: I0121 11:37:18.688507 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zbn97\" (UniqueName: \"kubernetes.io/projected/16274e36-c051-4ef8-9120-07b9996f43d6-kube-api-access-zbn97\") pod \"certified-operators-rpk5p\" (UID: \"16274e36-c051-4ef8-9120-07b9996f43d6\") " pod="openshift-marketplace/certified-operators-rpk5p" Jan 21 11:37:18 crc kubenswrapper[4824]: I0121 11:37:18.688574 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/16274e36-c051-4ef8-9120-07b9996f43d6-utilities\") pod \"certified-operators-rpk5p\" (UID: \"16274e36-c051-4ef8-9120-07b9996f43d6\") " pod="openshift-marketplace/certified-operators-rpk5p" Jan 21 11:37:18 crc kubenswrapper[4824]: I0121 11:37:18.688924 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/16274e36-c051-4ef8-9120-07b9996f43d6-utilities\") pod \"certified-operators-rpk5p\" (UID: \"16274e36-c051-4ef8-9120-07b9996f43d6\") " pod="openshift-marketplace/certified-operators-rpk5p" Jan 21 11:37:18 crc kubenswrapper[4824]: I0121 11:37:18.689169 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/16274e36-c051-4ef8-9120-07b9996f43d6-catalog-content\") pod \"certified-operators-rpk5p\" (UID: \"16274e36-c051-4ef8-9120-07b9996f43d6\") " pod="openshift-marketplace/certified-operators-rpk5p" Jan 21 11:37:18 crc kubenswrapper[4824]: I0121 11:37:18.704879 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zbn97\" (UniqueName: \"kubernetes.io/projected/16274e36-c051-4ef8-9120-07b9996f43d6-kube-api-access-zbn97\") pod \"certified-operators-rpk5p\" (UID: \"16274e36-c051-4ef8-9120-07b9996f43d6\") " pod="openshift-marketplace/certified-operators-rpk5p" Jan 21 11:37:18 crc kubenswrapper[4824]: I0121 11:37:18.872690 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rpk5p" Jan 21 11:37:19 crc kubenswrapper[4824]: I0121 11:37:19.288344 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-4bd5n" event={"ID":"8d46d1eb-4230-49c7-8376-698e7689ad2b","Type":"ContainerStarted","Data":"985d679df703288004096b440faf28c305917ee0a67a88fa7530222f8ee9d524"} Jan 21 11:37:19 crc kubenswrapper[4824]: I0121 11:37:19.303785 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-4bd5n" podStartSLOduration=1.8169619959999999 podStartE2EDuration="2.303770552s" podCreationTimestamp="2026-01-21 11:37:17 +0000 UTC" firstStartedPulling="2026-01-21 11:37:18.178888249 +0000 UTC m=+1580.471917541" lastFinishedPulling="2026-01-21 11:37:18.665696805 +0000 UTC m=+1580.958726097" observedRunningTime="2026-01-21 11:37:19.303537893 +0000 UTC m=+1581.596567185" watchObservedRunningTime="2026-01-21 11:37:19.303770552 +0000 UTC m=+1581.596799844" Jan 21 11:37:19 crc kubenswrapper[4824]: I0121 11:37:19.331024 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-rpk5p"] Jan 21 11:37:19 crc kubenswrapper[4824]: W0121 11:37:19.332155 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod16274e36_c051_4ef8_9120_07b9996f43d6.slice/crio-afc27a2346681f4afcd3cafb71b79b494dc2cff90b0fdb1ed7c606d5655d8650 WatchSource:0}: Error finding container afc27a2346681f4afcd3cafb71b79b494dc2cff90b0fdb1ed7c606d5655d8650: Status 404 returned error can't find the container with id afc27a2346681f4afcd3cafb71b79b494dc2cff90b0fdb1ed7c606d5655d8650 Jan 21 11:37:20 crc kubenswrapper[4824]: I0121 11:37:20.295488 4824 generic.go:334] "Generic (PLEG): container finished" podID="16274e36-c051-4ef8-9120-07b9996f43d6" containerID="e7d107ba375871e7f4f968f5b67758b58b2443b557b831e2e914f1350a98b41e" exitCode=0 Jan 21 11:37:20 crc kubenswrapper[4824]: I0121 11:37:20.295925 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rpk5p" event={"ID":"16274e36-c051-4ef8-9120-07b9996f43d6","Type":"ContainerDied","Data":"e7d107ba375871e7f4f968f5b67758b58b2443b557b831e2e914f1350a98b41e"} Jan 21 11:37:20 crc kubenswrapper[4824]: I0121 11:37:20.295987 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rpk5p" event={"ID":"16274e36-c051-4ef8-9120-07b9996f43d6","Type":"ContainerStarted","Data":"afc27a2346681f4afcd3cafb71b79b494dc2cff90b0fdb1ed7c606d5655d8650"} Jan 21 11:37:21 crc kubenswrapper[4824]: I0121 11:37:21.303813 4824 generic.go:334] "Generic (PLEG): container finished" podID="16274e36-c051-4ef8-9120-07b9996f43d6" containerID="c2019e8af01faf9c1722337cb07cb9e935dee62198d44250661012214b6f797b" exitCode=0 Jan 21 11:37:21 crc kubenswrapper[4824]: I0121 11:37:21.303847 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rpk5p" event={"ID":"16274e36-c051-4ef8-9120-07b9996f43d6","Type":"ContainerDied","Data":"c2019e8af01faf9c1722337cb07cb9e935dee62198d44250661012214b6f797b"} Jan 21 11:37:22 crc kubenswrapper[4824]: I0121 11:37:22.048585 4824 scope.go:117] "RemoveContainer" containerID="1cd835fe36ebb6541740e5ab5fd843470d46dfb78975fd968e5cc9aaef462499" Jan 21 11:37:22 crc kubenswrapper[4824]: E0121 11:37:22.049000 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 11:37:22 crc kubenswrapper[4824]: I0121 11:37:22.311728 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rpk5p" event={"ID":"16274e36-c051-4ef8-9120-07b9996f43d6","Type":"ContainerStarted","Data":"d199aa059ed50154802c051f60f56269888fa64fb50f8a9db325afdc3ceea9ff"} Jan 21 11:37:22 crc kubenswrapper[4824]: I0121 11:37:22.324545 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-rpk5p" podStartSLOduration=2.827084818 podStartE2EDuration="4.324528962s" podCreationTimestamp="2026-01-21 11:37:18 +0000 UTC" firstStartedPulling="2026-01-21 11:37:20.297062348 +0000 UTC m=+1582.590091640" lastFinishedPulling="2026-01-21 11:37:21.794506493 +0000 UTC m=+1584.087535784" observedRunningTime="2026-01-21 11:37:22.322279104 +0000 UTC m=+1584.615308396" watchObservedRunningTime="2026-01-21 11:37:22.324528962 +0000 UTC m=+1584.617558254" Jan 21 11:37:28 crc kubenswrapper[4824]: I0121 11:37:28.873442 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-rpk5p" Jan 21 11:37:28 crc kubenswrapper[4824]: I0121 11:37:28.874083 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-rpk5p" Jan 21 11:37:28 crc kubenswrapper[4824]: I0121 11:37:28.911833 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-rpk5p" Jan 21 11:37:29 crc kubenswrapper[4824]: I0121 11:37:29.401837 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-rpk5p" Jan 21 11:37:29 crc kubenswrapper[4824]: I0121 11:37:29.442121 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-rpk5p"] Jan 21 11:37:31 crc kubenswrapper[4824]: I0121 11:37:31.034978 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-4h8mx"] Jan 21 11:37:31 crc kubenswrapper[4824]: I0121 11:37:31.044447 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-16cb-account-create-update-qjbx7"] Jan 21 11:37:31 crc kubenswrapper[4824]: I0121 11:37:31.052089 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-16cb-account-create-update-qjbx7"] Jan 21 11:37:31 crc kubenswrapper[4824]: I0121 11:37:31.060283 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-4h8mx"] Jan 21 11:37:31 crc kubenswrapper[4824]: I0121 11:37:31.382212 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-rpk5p" podUID="16274e36-c051-4ef8-9120-07b9996f43d6" containerName="registry-server" containerID="cri-o://d199aa059ed50154802c051f60f56269888fa64fb50f8a9db325afdc3ceea9ff" gracePeriod=2 Jan 21 11:37:32 crc kubenswrapper[4824]: I0121 11:37:32.021989 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-91fe-account-create-update-4xgv2"] Jan 21 11:37:32 crc kubenswrapper[4824]: I0121 11:37:32.029138 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-90fb-account-create-update-m4n4l"] Jan 21 11:37:32 crc kubenswrapper[4824]: I0121 11:37:32.037057 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-dhvfz"] Jan 21 11:37:32 crc kubenswrapper[4824]: I0121 11:37:32.042816 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-90fb-account-create-update-m4n4l"] Jan 21 11:37:32 crc kubenswrapper[4824]: I0121 11:37:32.057255 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="34afc8af-6ce5-4ee0-917b-326bd60d6237" path="/var/lib/kubelet/pods/34afc8af-6ce5-4ee0-917b-326bd60d6237/volumes" Jan 21 11:37:32 crc kubenswrapper[4824]: I0121 11:37:32.057985 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="91261d07-56c8-4da3-8dd5-6f23587d77be" path="/var/lib/kubelet/pods/91261d07-56c8-4da3-8dd5-6f23587d77be/volumes" Jan 21 11:37:32 crc kubenswrapper[4824]: I0121 11:37:32.058703 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b24eefe0-9cb1-4672-85d8-6dfe2babebff" path="/var/lib/kubelet/pods/b24eefe0-9cb1-4672-85d8-6dfe2babebff/volumes" Jan 21 11:37:32 crc kubenswrapper[4824]: I0121 11:37:32.059312 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-91fe-account-create-update-4xgv2"] Jan 21 11:37:32 crc kubenswrapper[4824]: I0121 11:37:32.059331 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-dhvfz"] Jan 21 11:37:32 crc kubenswrapper[4824]: I0121 11:37:32.239705 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rpk5p" Jan 21 11:37:32 crc kubenswrapper[4824]: I0121 11:37:32.350363 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/16274e36-c051-4ef8-9120-07b9996f43d6-utilities\") pod \"16274e36-c051-4ef8-9120-07b9996f43d6\" (UID: \"16274e36-c051-4ef8-9120-07b9996f43d6\") " Jan 21 11:37:32 crc kubenswrapper[4824]: I0121 11:37:32.350637 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zbn97\" (UniqueName: \"kubernetes.io/projected/16274e36-c051-4ef8-9120-07b9996f43d6-kube-api-access-zbn97\") pod \"16274e36-c051-4ef8-9120-07b9996f43d6\" (UID: \"16274e36-c051-4ef8-9120-07b9996f43d6\") " Jan 21 11:37:32 crc kubenswrapper[4824]: I0121 11:37:32.350783 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/16274e36-c051-4ef8-9120-07b9996f43d6-catalog-content\") pod \"16274e36-c051-4ef8-9120-07b9996f43d6\" (UID: \"16274e36-c051-4ef8-9120-07b9996f43d6\") " Jan 21 11:37:32 crc kubenswrapper[4824]: I0121 11:37:32.351181 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/16274e36-c051-4ef8-9120-07b9996f43d6-utilities" (OuterVolumeSpecName: "utilities") pod "16274e36-c051-4ef8-9120-07b9996f43d6" (UID: "16274e36-c051-4ef8-9120-07b9996f43d6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:37:32 crc kubenswrapper[4824]: I0121 11:37:32.351412 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/16274e36-c051-4ef8-9120-07b9996f43d6-utilities\") on node \"crc\" DevicePath \"\"" Jan 21 11:37:32 crc kubenswrapper[4824]: I0121 11:37:32.355157 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/16274e36-c051-4ef8-9120-07b9996f43d6-kube-api-access-zbn97" (OuterVolumeSpecName: "kube-api-access-zbn97") pod "16274e36-c051-4ef8-9120-07b9996f43d6" (UID: "16274e36-c051-4ef8-9120-07b9996f43d6"). InnerVolumeSpecName "kube-api-access-zbn97". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:37:32 crc kubenswrapper[4824]: I0121 11:37:32.385777 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/16274e36-c051-4ef8-9120-07b9996f43d6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "16274e36-c051-4ef8-9120-07b9996f43d6" (UID: "16274e36-c051-4ef8-9120-07b9996f43d6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:37:32 crc kubenswrapper[4824]: I0121 11:37:32.392202 4824 generic.go:334] "Generic (PLEG): container finished" podID="16274e36-c051-4ef8-9120-07b9996f43d6" containerID="d199aa059ed50154802c051f60f56269888fa64fb50f8a9db325afdc3ceea9ff" exitCode=0 Jan 21 11:37:32 crc kubenswrapper[4824]: I0121 11:37:32.392258 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rpk5p" Jan 21 11:37:32 crc kubenswrapper[4824]: I0121 11:37:32.392250 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rpk5p" event={"ID":"16274e36-c051-4ef8-9120-07b9996f43d6","Type":"ContainerDied","Data":"d199aa059ed50154802c051f60f56269888fa64fb50f8a9db325afdc3ceea9ff"} Jan 21 11:37:32 crc kubenswrapper[4824]: I0121 11:37:32.392385 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rpk5p" event={"ID":"16274e36-c051-4ef8-9120-07b9996f43d6","Type":"ContainerDied","Data":"afc27a2346681f4afcd3cafb71b79b494dc2cff90b0fdb1ed7c606d5655d8650"} Jan 21 11:37:32 crc kubenswrapper[4824]: I0121 11:37:32.392403 4824 scope.go:117] "RemoveContainer" containerID="d199aa059ed50154802c051f60f56269888fa64fb50f8a9db325afdc3ceea9ff" Jan 21 11:37:32 crc kubenswrapper[4824]: I0121 11:37:32.413832 4824 scope.go:117] "RemoveContainer" containerID="c2019e8af01faf9c1722337cb07cb9e935dee62198d44250661012214b6f797b" Jan 21 11:37:32 crc kubenswrapper[4824]: I0121 11:37:32.420168 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-rpk5p"] Jan 21 11:37:32 crc kubenswrapper[4824]: I0121 11:37:32.429227 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-rpk5p"] Jan 21 11:37:32 crc kubenswrapper[4824]: I0121 11:37:32.444179 4824 scope.go:117] "RemoveContainer" containerID="e7d107ba375871e7f4f968f5b67758b58b2443b557b831e2e914f1350a98b41e" Jan 21 11:37:32 crc kubenswrapper[4824]: I0121 11:37:32.453540 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zbn97\" (UniqueName: \"kubernetes.io/projected/16274e36-c051-4ef8-9120-07b9996f43d6-kube-api-access-zbn97\") on node \"crc\" DevicePath \"\"" Jan 21 11:37:32 crc kubenswrapper[4824]: I0121 11:37:32.453568 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/16274e36-c051-4ef8-9120-07b9996f43d6-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 21 11:37:32 crc kubenswrapper[4824]: I0121 11:37:32.463719 4824 scope.go:117] "RemoveContainer" containerID="d199aa059ed50154802c051f60f56269888fa64fb50f8a9db325afdc3ceea9ff" Jan 21 11:37:32 crc kubenswrapper[4824]: E0121 11:37:32.464083 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d199aa059ed50154802c051f60f56269888fa64fb50f8a9db325afdc3ceea9ff\": container with ID starting with d199aa059ed50154802c051f60f56269888fa64fb50f8a9db325afdc3ceea9ff not found: ID does not exist" containerID="d199aa059ed50154802c051f60f56269888fa64fb50f8a9db325afdc3ceea9ff" Jan 21 11:37:32 crc kubenswrapper[4824]: I0121 11:37:32.464128 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d199aa059ed50154802c051f60f56269888fa64fb50f8a9db325afdc3ceea9ff"} err="failed to get container status \"d199aa059ed50154802c051f60f56269888fa64fb50f8a9db325afdc3ceea9ff\": rpc error: code = NotFound desc = could not find container \"d199aa059ed50154802c051f60f56269888fa64fb50f8a9db325afdc3ceea9ff\": container with ID starting with d199aa059ed50154802c051f60f56269888fa64fb50f8a9db325afdc3ceea9ff not found: ID does not exist" Jan 21 11:37:32 crc kubenswrapper[4824]: I0121 11:37:32.464151 4824 scope.go:117] "RemoveContainer" containerID="c2019e8af01faf9c1722337cb07cb9e935dee62198d44250661012214b6f797b" Jan 21 11:37:32 crc kubenswrapper[4824]: E0121 11:37:32.464468 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c2019e8af01faf9c1722337cb07cb9e935dee62198d44250661012214b6f797b\": container with ID starting with c2019e8af01faf9c1722337cb07cb9e935dee62198d44250661012214b6f797b not found: ID does not exist" containerID="c2019e8af01faf9c1722337cb07cb9e935dee62198d44250661012214b6f797b" Jan 21 11:37:32 crc kubenswrapper[4824]: I0121 11:37:32.464498 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c2019e8af01faf9c1722337cb07cb9e935dee62198d44250661012214b6f797b"} err="failed to get container status \"c2019e8af01faf9c1722337cb07cb9e935dee62198d44250661012214b6f797b\": rpc error: code = NotFound desc = could not find container \"c2019e8af01faf9c1722337cb07cb9e935dee62198d44250661012214b6f797b\": container with ID starting with c2019e8af01faf9c1722337cb07cb9e935dee62198d44250661012214b6f797b not found: ID does not exist" Jan 21 11:37:32 crc kubenswrapper[4824]: I0121 11:37:32.464529 4824 scope.go:117] "RemoveContainer" containerID="e7d107ba375871e7f4f968f5b67758b58b2443b557b831e2e914f1350a98b41e" Jan 21 11:37:32 crc kubenswrapper[4824]: E0121 11:37:32.464767 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e7d107ba375871e7f4f968f5b67758b58b2443b557b831e2e914f1350a98b41e\": container with ID starting with e7d107ba375871e7f4f968f5b67758b58b2443b557b831e2e914f1350a98b41e not found: ID does not exist" containerID="e7d107ba375871e7f4f968f5b67758b58b2443b557b831e2e914f1350a98b41e" Jan 21 11:37:32 crc kubenswrapper[4824]: I0121 11:37:32.464797 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e7d107ba375871e7f4f968f5b67758b58b2443b557b831e2e914f1350a98b41e"} err="failed to get container status \"e7d107ba375871e7f4f968f5b67758b58b2443b557b831e2e914f1350a98b41e\": rpc error: code = NotFound desc = could not find container \"e7d107ba375871e7f4f968f5b67758b58b2443b557b831e2e914f1350a98b41e\": container with ID starting with e7d107ba375871e7f4f968f5b67758b58b2443b557b831e2e914f1350a98b41e not found: ID does not exist" Jan 21 11:37:33 crc kubenswrapper[4824]: I0121 11:37:33.021515 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-f2xfg"] Jan 21 11:37:33 crc kubenswrapper[4824]: I0121 11:37:33.029401 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-f2xfg"] Jan 21 11:37:33 crc kubenswrapper[4824]: I0121 11:37:33.049610 4824 scope.go:117] "RemoveContainer" containerID="1cd835fe36ebb6541740e5ab5fd843470d46dfb78975fd968e5cc9aaef462499" Jan 21 11:37:33 crc kubenswrapper[4824]: E0121 11:37:33.050443 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 11:37:34 crc kubenswrapper[4824]: I0121 11:37:34.056924 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="12e7657f-78fe-4b5c-b614-4449e94366b9" path="/var/lib/kubelet/pods/12e7657f-78fe-4b5c-b614-4449e94366b9/volumes" Jan 21 11:37:34 crc kubenswrapper[4824]: I0121 11:37:34.057706 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="16274e36-c051-4ef8-9120-07b9996f43d6" path="/var/lib/kubelet/pods/16274e36-c051-4ef8-9120-07b9996f43d6/volumes" Jan 21 11:37:34 crc kubenswrapper[4824]: I0121 11:37:34.058373 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="60da7490-4aa1-4880-bfcb-e51a4ed99f91" path="/var/lib/kubelet/pods/60da7490-4aa1-4880-bfcb-e51a4ed99f91/volumes" Jan 21 11:37:34 crc kubenswrapper[4824]: I0121 11:37:34.059300 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d5162c70-9359-42d5-a02d-16140ac4e366" path="/var/lib/kubelet/pods/d5162c70-9359-42d5-a02d-16140ac4e366/volumes" Jan 21 11:37:48 crc kubenswrapper[4824]: I0121 11:37:48.054136 4824 scope.go:117] "RemoveContainer" containerID="1cd835fe36ebb6541740e5ab5fd843470d46dfb78975fd968e5cc9aaef462499" Jan 21 11:37:48 crc kubenswrapper[4824]: E0121 11:37:48.054733 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 11:37:52 crc kubenswrapper[4824]: I0121 11:37:52.027162 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-k6zm8"] Jan 21 11:37:52 crc kubenswrapper[4824]: I0121 11:37:52.033440 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-k6zm8"] Jan 21 11:37:52 crc kubenswrapper[4824]: I0121 11:37:52.057211 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9c73eda8-d376-4e18-9b33-f5ac4d10d091" path="/var/lib/kubelet/pods/9c73eda8-d376-4e18-9b33-f5ac4d10d091/volumes" Jan 21 11:37:52 crc kubenswrapper[4824]: I0121 11:37:52.525261 4824 generic.go:334] "Generic (PLEG): container finished" podID="8d46d1eb-4230-49c7-8376-698e7689ad2b" containerID="985d679df703288004096b440faf28c305917ee0a67a88fa7530222f8ee9d524" exitCode=0 Jan 21 11:37:52 crc kubenswrapper[4824]: I0121 11:37:52.525304 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-4bd5n" event={"ID":"8d46d1eb-4230-49c7-8376-698e7689ad2b","Type":"ContainerDied","Data":"985d679df703288004096b440faf28c305917ee0a67a88fa7530222f8ee9d524"} Jan 21 11:37:53 crc kubenswrapper[4824]: I0121 11:37:53.872410 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-4bd5n" Jan 21 11:37:54 crc kubenswrapper[4824]: I0121 11:37:54.032852 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4wqvn\" (UniqueName: \"kubernetes.io/projected/8d46d1eb-4230-49c7-8376-698e7689ad2b-kube-api-access-4wqvn\") pod \"8d46d1eb-4230-49c7-8376-698e7689ad2b\" (UID: \"8d46d1eb-4230-49c7-8376-698e7689ad2b\") " Jan 21 11:37:54 crc kubenswrapper[4824]: I0121 11:37:54.033083 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/8d46d1eb-4230-49c7-8376-698e7689ad2b-ssh-key-openstack-edpm-ipam\") pod \"8d46d1eb-4230-49c7-8376-698e7689ad2b\" (UID: \"8d46d1eb-4230-49c7-8376-698e7689ad2b\") " Jan 21 11:37:54 crc kubenswrapper[4824]: I0121 11:37:54.033110 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8d46d1eb-4230-49c7-8376-698e7689ad2b-inventory\") pod \"8d46d1eb-4230-49c7-8376-698e7689ad2b\" (UID: \"8d46d1eb-4230-49c7-8376-698e7689ad2b\") " Jan 21 11:37:54 crc kubenswrapper[4824]: I0121 11:37:54.037347 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8d46d1eb-4230-49c7-8376-698e7689ad2b-kube-api-access-4wqvn" (OuterVolumeSpecName: "kube-api-access-4wqvn") pod "8d46d1eb-4230-49c7-8376-698e7689ad2b" (UID: "8d46d1eb-4230-49c7-8376-698e7689ad2b"). InnerVolumeSpecName "kube-api-access-4wqvn". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:37:54 crc kubenswrapper[4824]: I0121 11:37:54.053813 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8d46d1eb-4230-49c7-8376-698e7689ad2b-inventory" (OuterVolumeSpecName: "inventory") pod "8d46d1eb-4230-49c7-8376-698e7689ad2b" (UID: "8d46d1eb-4230-49c7-8376-698e7689ad2b"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:37:54 crc kubenswrapper[4824]: I0121 11:37:54.054056 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8d46d1eb-4230-49c7-8376-698e7689ad2b-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "8d46d1eb-4230-49c7-8376-698e7689ad2b" (UID: "8d46d1eb-4230-49c7-8376-698e7689ad2b"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:37:54 crc kubenswrapper[4824]: I0121 11:37:54.135443 4824 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/8d46d1eb-4230-49c7-8376-698e7689ad2b-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Jan 21 11:37:54 crc kubenswrapper[4824]: I0121 11:37:54.135467 4824 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8d46d1eb-4230-49c7-8376-698e7689ad2b-inventory\") on node \"crc\" DevicePath \"\"" Jan 21 11:37:54 crc kubenswrapper[4824]: I0121 11:37:54.135476 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4wqvn\" (UniqueName: \"kubernetes.io/projected/8d46d1eb-4230-49c7-8376-698e7689ad2b-kube-api-access-4wqvn\") on node \"crc\" DevicePath \"\"" Jan 21 11:37:54 crc kubenswrapper[4824]: I0121 11:37:54.538354 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-4bd5n" event={"ID":"8d46d1eb-4230-49c7-8376-698e7689ad2b","Type":"ContainerDied","Data":"7c498c4871058e9b8aef05f8de415d1229252f8805909c3848b5ff58ff925394"} Jan 21 11:37:54 crc kubenswrapper[4824]: I0121 11:37:54.538390 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7c498c4871058e9b8aef05f8de415d1229252f8805909c3848b5ff58ff925394" Jan 21 11:37:54 crc kubenswrapper[4824]: I0121 11:37:54.538396 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-4bd5n" Jan 21 11:37:54 crc kubenswrapper[4824]: I0121 11:37:54.594857 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-wbrwb"] Jan 21 11:37:54 crc kubenswrapper[4824]: E0121 11:37:54.595228 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="16274e36-c051-4ef8-9120-07b9996f43d6" containerName="registry-server" Jan 21 11:37:54 crc kubenswrapper[4824]: I0121 11:37:54.595247 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="16274e36-c051-4ef8-9120-07b9996f43d6" containerName="registry-server" Jan 21 11:37:54 crc kubenswrapper[4824]: E0121 11:37:54.595269 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="16274e36-c051-4ef8-9120-07b9996f43d6" containerName="extract-content" Jan 21 11:37:54 crc kubenswrapper[4824]: I0121 11:37:54.595275 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="16274e36-c051-4ef8-9120-07b9996f43d6" containerName="extract-content" Jan 21 11:37:54 crc kubenswrapper[4824]: E0121 11:37:54.595284 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="16274e36-c051-4ef8-9120-07b9996f43d6" containerName="extract-utilities" Jan 21 11:37:54 crc kubenswrapper[4824]: I0121 11:37:54.595290 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="16274e36-c051-4ef8-9120-07b9996f43d6" containerName="extract-utilities" Jan 21 11:37:54 crc kubenswrapper[4824]: E0121 11:37:54.595298 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d46d1eb-4230-49c7-8376-698e7689ad2b" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Jan 21 11:37:54 crc kubenswrapper[4824]: I0121 11:37:54.595305 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d46d1eb-4230-49c7-8376-698e7689ad2b" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Jan 21 11:37:54 crc kubenswrapper[4824]: I0121 11:37:54.595461 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="16274e36-c051-4ef8-9120-07b9996f43d6" containerName="registry-server" Jan 21 11:37:54 crc kubenswrapper[4824]: I0121 11:37:54.595473 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="8d46d1eb-4230-49c7-8376-698e7689ad2b" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Jan 21 11:37:54 crc kubenswrapper[4824]: I0121 11:37:54.596023 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-wbrwb" Jan 21 11:37:54 crc kubenswrapper[4824]: I0121 11:37:54.597732 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Jan 21 11:37:54 crc kubenswrapper[4824]: I0121 11:37:54.597784 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Jan 21 11:37:54 crc kubenswrapper[4824]: I0121 11:37:54.597794 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-tcqk6" Jan 21 11:37:54 crc kubenswrapper[4824]: I0121 11:37:54.598290 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Jan 21 11:37:54 crc kubenswrapper[4824]: I0121 11:37:54.608211 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-wbrwb"] Jan 21 11:37:54 crc kubenswrapper[4824]: I0121 11:37:54.744885 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/4529344f-3dfd-4858-83f1-4abffe47f2ad-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-wbrwb\" (UID: \"4529344f-3dfd-4858-83f1-4abffe47f2ad\") " pod="openstack/ssh-known-hosts-edpm-deployment-wbrwb" Jan 21 11:37:54 crc kubenswrapper[4824]: I0121 11:37:54.745105 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c2wph\" (UniqueName: \"kubernetes.io/projected/4529344f-3dfd-4858-83f1-4abffe47f2ad-kube-api-access-c2wph\") pod \"ssh-known-hosts-edpm-deployment-wbrwb\" (UID: \"4529344f-3dfd-4858-83f1-4abffe47f2ad\") " pod="openstack/ssh-known-hosts-edpm-deployment-wbrwb" Jan 21 11:37:54 crc kubenswrapper[4824]: I0121 11:37:54.745391 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/4529344f-3dfd-4858-83f1-4abffe47f2ad-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-wbrwb\" (UID: \"4529344f-3dfd-4858-83f1-4abffe47f2ad\") " pod="openstack/ssh-known-hosts-edpm-deployment-wbrwb" Jan 21 11:37:54 crc kubenswrapper[4824]: I0121 11:37:54.847289 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c2wph\" (UniqueName: \"kubernetes.io/projected/4529344f-3dfd-4858-83f1-4abffe47f2ad-kube-api-access-c2wph\") pod \"ssh-known-hosts-edpm-deployment-wbrwb\" (UID: \"4529344f-3dfd-4858-83f1-4abffe47f2ad\") " pod="openstack/ssh-known-hosts-edpm-deployment-wbrwb" Jan 21 11:37:54 crc kubenswrapper[4824]: I0121 11:37:54.847435 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/4529344f-3dfd-4858-83f1-4abffe47f2ad-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-wbrwb\" (UID: \"4529344f-3dfd-4858-83f1-4abffe47f2ad\") " pod="openstack/ssh-known-hosts-edpm-deployment-wbrwb" Jan 21 11:37:54 crc kubenswrapper[4824]: I0121 11:37:54.847469 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/4529344f-3dfd-4858-83f1-4abffe47f2ad-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-wbrwb\" (UID: \"4529344f-3dfd-4858-83f1-4abffe47f2ad\") " pod="openstack/ssh-known-hosts-edpm-deployment-wbrwb" Jan 21 11:37:54 crc kubenswrapper[4824]: I0121 11:37:54.850197 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/4529344f-3dfd-4858-83f1-4abffe47f2ad-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-wbrwb\" (UID: \"4529344f-3dfd-4858-83f1-4abffe47f2ad\") " pod="openstack/ssh-known-hosts-edpm-deployment-wbrwb" Jan 21 11:37:54 crc kubenswrapper[4824]: I0121 11:37:54.850339 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/4529344f-3dfd-4858-83f1-4abffe47f2ad-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-wbrwb\" (UID: \"4529344f-3dfd-4858-83f1-4abffe47f2ad\") " pod="openstack/ssh-known-hosts-edpm-deployment-wbrwb" Jan 21 11:37:54 crc kubenswrapper[4824]: I0121 11:37:54.860271 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c2wph\" (UniqueName: \"kubernetes.io/projected/4529344f-3dfd-4858-83f1-4abffe47f2ad-kube-api-access-c2wph\") pod \"ssh-known-hosts-edpm-deployment-wbrwb\" (UID: \"4529344f-3dfd-4858-83f1-4abffe47f2ad\") " pod="openstack/ssh-known-hosts-edpm-deployment-wbrwb" Jan 21 11:37:54 crc kubenswrapper[4824]: I0121 11:37:54.909885 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-wbrwb" Jan 21 11:37:55 crc kubenswrapper[4824]: I0121 11:37:55.347142 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-wbrwb"] Jan 21 11:37:55 crc kubenswrapper[4824]: I0121 11:37:55.545227 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-wbrwb" event={"ID":"4529344f-3dfd-4858-83f1-4abffe47f2ad","Type":"ContainerStarted","Data":"e68763debb1695544c631b3afa9c054e02ec0885d12cf70e1432d634dc00418b"} Jan 21 11:37:56 crc kubenswrapper[4824]: I0121 11:37:56.552237 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-wbrwb" event={"ID":"4529344f-3dfd-4858-83f1-4abffe47f2ad","Type":"ContainerStarted","Data":"39da814edb81d4f0345bc2ff3a9d25851f522015824a48410564608294c239ce"} Jan 21 11:37:56 crc kubenswrapper[4824]: I0121 11:37:56.566409 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-edpm-deployment-wbrwb" podStartSLOduration=2.02052861 podStartE2EDuration="2.566393808s" podCreationTimestamp="2026-01-21 11:37:54 +0000 UTC" firstStartedPulling="2026-01-21 11:37:55.35336279 +0000 UTC m=+1617.646392082" lastFinishedPulling="2026-01-21 11:37:55.899227988 +0000 UTC m=+1618.192257280" observedRunningTime="2026-01-21 11:37:56.563917693 +0000 UTC m=+1618.856946985" watchObservedRunningTime="2026-01-21 11:37:56.566393808 +0000 UTC m=+1618.859423100" Jan 21 11:38:00 crc kubenswrapper[4824]: I0121 11:38:00.157831 4824 scope.go:117] "RemoveContainer" containerID="6fa0076da1636efc5cc0178f55ca211d393f930b6c63c3d85cac4eedd9f2d288" Jan 21 11:38:00 crc kubenswrapper[4824]: I0121 11:38:00.182142 4824 scope.go:117] "RemoveContainer" containerID="a9bd46298904e9e7c4cccbdd1cecf51e20eaa4ef17b0e64f0ae86c7cca0ec79d" Jan 21 11:38:00 crc kubenswrapper[4824]: I0121 11:38:00.212348 4824 scope.go:117] "RemoveContainer" containerID="4c9c6f12e02388718e6b9fff3020bdecb9cf28f3ef2f988918e2aeb78dac7d68" Jan 21 11:38:00 crc kubenswrapper[4824]: I0121 11:38:00.249485 4824 scope.go:117] "RemoveContainer" containerID="b5f108624f476c15b99bcdcb7c95850ef2819568712069d3c2536bdcb6daa49b" Jan 21 11:38:00 crc kubenswrapper[4824]: I0121 11:38:00.285934 4824 scope.go:117] "RemoveContainer" containerID="c2bb2e93e558c6a9e9ce5d9ae43e9812275d36c293fd644336fd7e9fcc3ba6c0" Jan 21 11:38:00 crc kubenswrapper[4824]: I0121 11:38:00.315237 4824 scope.go:117] "RemoveContainer" containerID="7ab5f4aea14ccac0c6081bda9c7469c2c70bff8d5179b2221cca7a70044cada5" Jan 21 11:38:00 crc kubenswrapper[4824]: I0121 11:38:00.333150 4824 scope.go:117] "RemoveContainer" containerID="06b41897897bde2d466371c85738035a8ab9ee6530d1dfecdba6e71d57264bfe" Jan 21 11:38:00 crc kubenswrapper[4824]: I0121 11:38:00.349098 4824 scope.go:117] "RemoveContainer" containerID="7d5e936f6ab78e748abd4aabb374f95965e7eaaf526fd1a755b675eed79144ec" Jan 21 11:38:00 crc kubenswrapper[4824]: I0121 11:38:00.365314 4824 scope.go:117] "RemoveContainer" containerID="edd12500feb323e4dab204d7803bc5160bb50ef2b5116e5ba7686e65e4e155fc" Jan 21 11:38:01 crc kubenswrapper[4824]: I0121 11:38:01.589875 4824 generic.go:334] "Generic (PLEG): container finished" podID="4529344f-3dfd-4858-83f1-4abffe47f2ad" containerID="39da814edb81d4f0345bc2ff3a9d25851f522015824a48410564608294c239ce" exitCode=0 Jan 21 11:38:01 crc kubenswrapper[4824]: I0121 11:38:01.589948 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-wbrwb" event={"ID":"4529344f-3dfd-4858-83f1-4abffe47f2ad","Type":"ContainerDied","Data":"39da814edb81d4f0345bc2ff3a9d25851f522015824a48410564608294c239ce"} Jan 21 11:38:02 crc kubenswrapper[4824]: I0121 11:38:02.049031 4824 scope.go:117] "RemoveContainer" containerID="1cd835fe36ebb6541740e5ab5fd843470d46dfb78975fd968e5cc9aaef462499" Jan 21 11:38:02 crc kubenswrapper[4824]: E0121 11:38:02.049302 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 11:38:02 crc kubenswrapper[4824]: I0121 11:38:02.909214 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-wbrwb" Jan 21 11:38:03 crc kubenswrapper[4824]: I0121 11:38:03.079461 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/4529344f-3dfd-4858-83f1-4abffe47f2ad-ssh-key-openstack-edpm-ipam\") pod \"4529344f-3dfd-4858-83f1-4abffe47f2ad\" (UID: \"4529344f-3dfd-4858-83f1-4abffe47f2ad\") " Jan 21 11:38:03 crc kubenswrapper[4824]: I0121 11:38:03.079527 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/4529344f-3dfd-4858-83f1-4abffe47f2ad-inventory-0\") pod \"4529344f-3dfd-4858-83f1-4abffe47f2ad\" (UID: \"4529344f-3dfd-4858-83f1-4abffe47f2ad\") " Jan 21 11:38:03 crc kubenswrapper[4824]: I0121 11:38:03.079584 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c2wph\" (UniqueName: \"kubernetes.io/projected/4529344f-3dfd-4858-83f1-4abffe47f2ad-kube-api-access-c2wph\") pod \"4529344f-3dfd-4858-83f1-4abffe47f2ad\" (UID: \"4529344f-3dfd-4858-83f1-4abffe47f2ad\") " Jan 21 11:38:03 crc kubenswrapper[4824]: I0121 11:38:03.083866 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4529344f-3dfd-4858-83f1-4abffe47f2ad-kube-api-access-c2wph" (OuterVolumeSpecName: "kube-api-access-c2wph") pod "4529344f-3dfd-4858-83f1-4abffe47f2ad" (UID: "4529344f-3dfd-4858-83f1-4abffe47f2ad"). InnerVolumeSpecName "kube-api-access-c2wph". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:38:03 crc kubenswrapper[4824]: I0121 11:38:03.099833 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4529344f-3dfd-4858-83f1-4abffe47f2ad-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "4529344f-3dfd-4858-83f1-4abffe47f2ad" (UID: "4529344f-3dfd-4858-83f1-4abffe47f2ad"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:38:03 crc kubenswrapper[4824]: I0121 11:38:03.099927 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4529344f-3dfd-4858-83f1-4abffe47f2ad-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "4529344f-3dfd-4858-83f1-4abffe47f2ad" (UID: "4529344f-3dfd-4858-83f1-4abffe47f2ad"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:38:03 crc kubenswrapper[4824]: I0121 11:38:03.181724 4824 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/4529344f-3dfd-4858-83f1-4abffe47f2ad-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Jan 21 11:38:03 crc kubenswrapper[4824]: I0121 11:38:03.181751 4824 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/4529344f-3dfd-4858-83f1-4abffe47f2ad-inventory-0\") on node \"crc\" DevicePath \"\"" Jan 21 11:38:03 crc kubenswrapper[4824]: I0121 11:38:03.181761 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c2wph\" (UniqueName: \"kubernetes.io/projected/4529344f-3dfd-4858-83f1-4abffe47f2ad-kube-api-access-c2wph\") on node \"crc\" DevicePath \"\"" Jan 21 11:38:03 crc kubenswrapper[4824]: I0121 11:38:03.604940 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-wbrwb" event={"ID":"4529344f-3dfd-4858-83f1-4abffe47f2ad","Type":"ContainerDied","Data":"e68763debb1695544c631b3afa9c054e02ec0885d12cf70e1432d634dc00418b"} Jan 21 11:38:03 crc kubenswrapper[4824]: I0121 11:38:03.604985 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-wbrwb" Jan 21 11:38:03 crc kubenswrapper[4824]: I0121 11:38:03.604992 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e68763debb1695544c631b3afa9c054e02ec0885d12cf70e1432d634dc00418b" Jan 21 11:38:03 crc kubenswrapper[4824]: I0121 11:38:03.659398 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-pn4sw"] Jan 21 11:38:03 crc kubenswrapper[4824]: E0121 11:38:03.659787 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4529344f-3dfd-4858-83f1-4abffe47f2ad" containerName="ssh-known-hosts-edpm-deployment" Jan 21 11:38:03 crc kubenswrapper[4824]: I0121 11:38:03.659805 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="4529344f-3dfd-4858-83f1-4abffe47f2ad" containerName="ssh-known-hosts-edpm-deployment" Jan 21 11:38:03 crc kubenswrapper[4824]: I0121 11:38:03.660003 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="4529344f-3dfd-4858-83f1-4abffe47f2ad" containerName="ssh-known-hosts-edpm-deployment" Jan 21 11:38:03 crc kubenswrapper[4824]: I0121 11:38:03.660569 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-pn4sw" Jan 21 11:38:03 crc kubenswrapper[4824]: I0121 11:38:03.662416 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Jan 21 11:38:03 crc kubenswrapper[4824]: I0121 11:38:03.662668 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Jan 21 11:38:03 crc kubenswrapper[4824]: I0121 11:38:03.662795 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Jan 21 11:38:03 crc kubenswrapper[4824]: I0121 11:38:03.663634 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-tcqk6" Jan 21 11:38:03 crc kubenswrapper[4824]: I0121 11:38:03.675130 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-pn4sw"] Jan 21 11:38:03 crc kubenswrapper[4824]: I0121 11:38:03.792754 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ad533443-01b7-4860-9920-93ed67f6b52f-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-pn4sw\" (UID: \"ad533443-01b7-4860-9920-93ed67f6b52f\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-pn4sw" Jan 21 11:38:03 crc kubenswrapper[4824]: I0121 11:38:03.793129 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/ad533443-01b7-4860-9920-93ed67f6b52f-ssh-key-openstack-edpm-ipam\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-pn4sw\" (UID: \"ad533443-01b7-4860-9920-93ed67f6b52f\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-pn4sw" Jan 21 11:38:03 crc kubenswrapper[4824]: I0121 11:38:03.793159 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mph2g\" (UniqueName: \"kubernetes.io/projected/ad533443-01b7-4860-9920-93ed67f6b52f-kube-api-access-mph2g\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-pn4sw\" (UID: \"ad533443-01b7-4860-9920-93ed67f6b52f\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-pn4sw" Jan 21 11:38:03 crc kubenswrapper[4824]: I0121 11:38:03.894635 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/ad533443-01b7-4860-9920-93ed67f6b52f-ssh-key-openstack-edpm-ipam\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-pn4sw\" (UID: \"ad533443-01b7-4860-9920-93ed67f6b52f\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-pn4sw" Jan 21 11:38:03 crc kubenswrapper[4824]: I0121 11:38:03.894687 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mph2g\" (UniqueName: \"kubernetes.io/projected/ad533443-01b7-4860-9920-93ed67f6b52f-kube-api-access-mph2g\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-pn4sw\" (UID: \"ad533443-01b7-4860-9920-93ed67f6b52f\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-pn4sw" Jan 21 11:38:03 crc kubenswrapper[4824]: I0121 11:38:03.894775 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ad533443-01b7-4860-9920-93ed67f6b52f-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-pn4sw\" (UID: \"ad533443-01b7-4860-9920-93ed67f6b52f\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-pn4sw" Jan 21 11:38:03 crc kubenswrapper[4824]: I0121 11:38:03.897790 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/ad533443-01b7-4860-9920-93ed67f6b52f-ssh-key-openstack-edpm-ipam\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-pn4sw\" (UID: \"ad533443-01b7-4860-9920-93ed67f6b52f\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-pn4sw" Jan 21 11:38:03 crc kubenswrapper[4824]: I0121 11:38:03.903546 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ad533443-01b7-4860-9920-93ed67f6b52f-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-pn4sw\" (UID: \"ad533443-01b7-4860-9920-93ed67f6b52f\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-pn4sw" Jan 21 11:38:03 crc kubenswrapper[4824]: I0121 11:38:03.908866 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mph2g\" (UniqueName: \"kubernetes.io/projected/ad533443-01b7-4860-9920-93ed67f6b52f-kube-api-access-mph2g\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-pn4sw\" (UID: \"ad533443-01b7-4860-9920-93ed67f6b52f\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-pn4sw" Jan 21 11:38:03 crc kubenswrapper[4824]: I0121 11:38:03.975840 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-pn4sw" Jan 21 11:38:04 crc kubenswrapper[4824]: I0121 11:38:04.391522 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-pn4sw"] Jan 21 11:38:04 crc kubenswrapper[4824]: I0121 11:38:04.612248 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-pn4sw" event={"ID":"ad533443-01b7-4860-9920-93ed67f6b52f","Type":"ContainerStarted","Data":"fb01de3abedf34fe5f459bd3c02fff34247e31b70d197937119d7945bb60ea39"} Jan 21 11:38:05 crc kubenswrapper[4824]: I0121 11:38:05.620903 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-pn4sw" event={"ID":"ad533443-01b7-4860-9920-93ed67f6b52f","Type":"ContainerStarted","Data":"946d859a672c7d1f509346749f7eeff11c24665737fbec52db44d7f634ff05ac"} Jan 21 11:38:05 crc kubenswrapper[4824]: I0121 11:38:05.640001 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-pn4sw" podStartSLOduration=2.098352483 podStartE2EDuration="2.639988266s" podCreationTimestamp="2026-01-21 11:38:03 +0000 UTC" firstStartedPulling="2026-01-21 11:38:04.398894178 +0000 UTC m=+1626.691923471" lastFinishedPulling="2026-01-21 11:38:04.940529962 +0000 UTC m=+1627.233559254" observedRunningTime="2026-01-21 11:38:05.635028412 +0000 UTC m=+1627.928057704" watchObservedRunningTime="2026-01-21 11:38:05.639988266 +0000 UTC m=+1627.933017557" Jan 21 11:38:10 crc kubenswrapper[4824]: I0121 11:38:10.663065 4824 generic.go:334] "Generic (PLEG): container finished" podID="ad533443-01b7-4860-9920-93ed67f6b52f" containerID="946d859a672c7d1f509346749f7eeff11c24665737fbec52db44d7f634ff05ac" exitCode=0 Jan 21 11:38:10 crc kubenswrapper[4824]: I0121 11:38:10.663139 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-pn4sw" event={"ID":"ad533443-01b7-4860-9920-93ed67f6b52f","Type":"ContainerDied","Data":"946d859a672c7d1f509346749f7eeff11c24665737fbec52db44d7f634ff05ac"} Jan 21 11:38:11 crc kubenswrapper[4824]: I0121 11:38:11.981471 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-pn4sw" Jan 21 11:38:12 crc kubenswrapper[4824]: I0121 11:38:12.127035 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/ad533443-01b7-4860-9920-93ed67f6b52f-ssh-key-openstack-edpm-ipam\") pod \"ad533443-01b7-4860-9920-93ed67f6b52f\" (UID: \"ad533443-01b7-4860-9920-93ed67f6b52f\") " Jan 21 11:38:12 crc kubenswrapper[4824]: I0121 11:38:12.127232 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mph2g\" (UniqueName: \"kubernetes.io/projected/ad533443-01b7-4860-9920-93ed67f6b52f-kube-api-access-mph2g\") pod \"ad533443-01b7-4860-9920-93ed67f6b52f\" (UID: \"ad533443-01b7-4860-9920-93ed67f6b52f\") " Jan 21 11:38:12 crc kubenswrapper[4824]: I0121 11:38:12.127273 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ad533443-01b7-4860-9920-93ed67f6b52f-inventory\") pod \"ad533443-01b7-4860-9920-93ed67f6b52f\" (UID: \"ad533443-01b7-4860-9920-93ed67f6b52f\") " Jan 21 11:38:12 crc kubenswrapper[4824]: I0121 11:38:12.131353 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ad533443-01b7-4860-9920-93ed67f6b52f-kube-api-access-mph2g" (OuterVolumeSpecName: "kube-api-access-mph2g") pod "ad533443-01b7-4860-9920-93ed67f6b52f" (UID: "ad533443-01b7-4860-9920-93ed67f6b52f"). InnerVolumeSpecName "kube-api-access-mph2g". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:38:12 crc kubenswrapper[4824]: I0121 11:38:12.147453 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ad533443-01b7-4860-9920-93ed67f6b52f-inventory" (OuterVolumeSpecName: "inventory") pod "ad533443-01b7-4860-9920-93ed67f6b52f" (UID: "ad533443-01b7-4860-9920-93ed67f6b52f"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:38:12 crc kubenswrapper[4824]: I0121 11:38:12.152311 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ad533443-01b7-4860-9920-93ed67f6b52f-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "ad533443-01b7-4860-9920-93ed67f6b52f" (UID: "ad533443-01b7-4860-9920-93ed67f6b52f"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:38:12 crc kubenswrapper[4824]: I0121 11:38:12.228909 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mph2g\" (UniqueName: \"kubernetes.io/projected/ad533443-01b7-4860-9920-93ed67f6b52f-kube-api-access-mph2g\") on node \"crc\" DevicePath \"\"" Jan 21 11:38:12 crc kubenswrapper[4824]: I0121 11:38:12.228933 4824 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ad533443-01b7-4860-9920-93ed67f6b52f-inventory\") on node \"crc\" DevicePath \"\"" Jan 21 11:38:12 crc kubenswrapper[4824]: I0121 11:38:12.228943 4824 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/ad533443-01b7-4860-9920-93ed67f6b52f-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Jan 21 11:38:12 crc kubenswrapper[4824]: I0121 11:38:12.677038 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-pn4sw" event={"ID":"ad533443-01b7-4860-9920-93ed67f6b52f","Type":"ContainerDied","Data":"fb01de3abedf34fe5f459bd3c02fff34247e31b70d197937119d7945bb60ea39"} Jan 21 11:38:12 crc kubenswrapper[4824]: I0121 11:38:12.677065 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-pn4sw" Jan 21 11:38:12 crc kubenswrapper[4824]: I0121 11:38:12.677073 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fb01de3abedf34fe5f459bd3c02fff34247e31b70d197937119d7945bb60ea39" Jan 21 11:38:12 crc kubenswrapper[4824]: I0121 11:38:12.722765 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-8w57w"] Jan 21 11:38:12 crc kubenswrapper[4824]: E0121 11:38:12.723125 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad533443-01b7-4860-9920-93ed67f6b52f" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Jan 21 11:38:12 crc kubenswrapper[4824]: I0121 11:38:12.723142 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad533443-01b7-4860-9920-93ed67f6b52f" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Jan 21 11:38:12 crc kubenswrapper[4824]: I0121 11:38:12.723370 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="ad533443-01b7-4860-9920-93ed67f6b52f" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Jan 21 11:38:12 crc kubenswrapper[4824]: I0121 11:38:12.723871 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-8w57w" Jan 21 11:38:12 crc kubenswrapper[4824]: I0121 11:38:12.726731 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Jan 21 11:38:12 crc kubenswrapper[4824]: I0121 11:38:12.726801 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Jan 21 11:38:12 crc kubenswrapper[4824]: I0121 11:38:12.727084 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Jan 21 11:38:12 crc kubenswrapper[4824]: I0121 11:38:12.727116 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-tcqk6" Jan 21 11:38:12 crc kubenswrapper[4824]: I0121 11:38:12.730777 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-8w57w"] Jan 21 11:38:12 crc kubenswrapper[4824]: I0121 11:38:12.735177 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/890acdab-9f31-4afc-80e4-c1df308ac5a6-ssh-key-openstack-edpm-ipam\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-8w57w\" (UID: \"890acdab-9f31-4afc-80e4-c1df308ac5a6\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-8w57w" Jan 21 11:38:12 crc kubenswrapper[4824]: I0121 11:38:12.735469 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d25d6\" (UniqueName: \"kubernetes.io/projected/890acdab-9f31-4afc-80e4-c1df308ac5a6-kube-api-access-d25d6\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-8w57w\" (UID: \"890acdab-9f31-4afc-80e4-c1df308ac5a6\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-8w57w" Jan 21 11:38:12 crc kubenswrapper[4824]: I0121 11:38:12.735633 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/890acdab-9f31-4afc-80e4-c1df308ac5a6-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-8w57w\" (UID: \"890acdab-9f31-4afc-80e4-c1df308ac5a6\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-8w57w" Jan 21 11:38:12 crc kubenswrapper[4824]: I0121 11:38:12.837246 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/890acdab-9f31-4afc-80e4-c1df308ac5a6-ssh-key-openstack-edpm-ipam\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-8w57w\" (UID: \"890acdab-9f31-4afc-80e4-c1df308ac5a6\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-8w57w" Jan 21 11:38:12 crc kubenswrapper[4824]: I0121 11:38:12.837330 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d25d6\" (UniqueName: \"kubernetes.io/projected/890acdab-9f31-4afc-80e4-c1df308ac5a6-kube-api-access-d25d6\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-8w57w\" (UID: \"890acdab-9f31-4afc-80e4-c1df308ac5a6\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-8w57w" Jan 21 11:38:12 crc kubenswrapper[4824]: I0121 11:38:12.837552 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/890acdab-9f31-4afc-80e4-c1df308ac5a6-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-8w57w\" (UID: \"890acdab-9f31-4afc-80e4-c1df308ac5a6\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-8w57w" Jan 21 11:38:12 crc kubenswrapper[4824]: I0121 11:38:12.841431 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/890acdab-9f31-4afc-80e4-c1df308ac5a6-ssh-key-openstack-edpm-ipam\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-8w57w\" (UID: \"890acdab-9f31-4afc-80e4-c1df308ac5a6\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-8w57w" Jan 21 11:38:12 crc kubenswrapper[4824]: I0121 11:38:12.843083 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/890acdab-9f31-4afc-80e4-c1df308ac5a6-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-8w57w\" (UID: \"890acdab-9f31-4afc-80e4-c1df308ac5a6\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-8w57w" Jan 21 11:38:12 crc kubenswrapper[4824]: I0121 11:38:12.851048 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d25d6\" (UniqueName: \"kubernetes.io/projected/890acdab-9f31-4afc-80e4-c1df308ac5a6-kube-api-access-d25d6\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-8w57w\" (UID: \"890acdab-9f31-4afc-80e4-c1df308ac5a6\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-8w57w" Jan 21 11:38:13 crc kubenswrapper[4824]: I0121 11:38:13.036854 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-8w57w" Jan 21 11:38:13 crc kubenswrapper[4824]: I0121 11:38:13.349595 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-8w57w"] Jan 21 11:38:13 crc kubenswrapper[4824]: I0121 11:38:13.684106 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-8w57w" event={"ID":"890acdab-9f31-4afc-80e4-c1df308ac5a6","Type":"ContainerStarted","Data":"b95a36f4b2a1281221a694900f274ed364cc1352c3fe01992775f1762a0a247c"} Jan 21 11:38:14 crc kubenswrapper[4824]: I0121 11:38:14.021652 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-g9vq5"] Jan 21 11:38:14 crc kubenswrapper[4824]: I0121 11:38:14.027544 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-g9vq5"] Jan 21 11:38:14 crc kubenswrapper[4824]: I0121 11:38:14.057287 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="baa34da9-3711-4cd4-a5eb-ee5814ac51a7" path="/var/lib/kubelet/pods/baa34da9-3711-4cd4-a5eb-ee5814ac51a7/volumes" Jan 21 11:38:14 crc kubenswrapper[4824]: I0121 11:38:14.691410 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-8w57w" event={"ID":"890acdab-9f31-4afc-80e4-c1df308ac5a6","Type":"ContainerStarted","Data":"882b40080daa5269b988b3f409290f0fb60c6467f759ffaa24e78d2d41b4111c"} Jan 21 11:38:14 crc kubenswrapper[4824]: I0121 11:38:14.706632 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-8w57w" podStartSLOduration=2.190013548 podStartE2EDuration="2.706616202s" podCreationTimestamp="2026-01-21 11:38:12 +0000 UTC" firstStartedPulling="2026-01-21 11:38:13.350115932 +0000 UTC m=+1635.643145224" lastFinishedPulling="2026-01-21 11:38:13.866718587 +0000 UTC m=+1636.159747878" observedRunningTime="2026-01-21 11:38:14.706190348 +0000 UTC m=+1636.999219640" watchObservedRunningTime="2026-01-21 11:38:14.706616202 +0000 UTC m=+1636.999645494" Jan 21 11:38:15 crc kubenswrapper[4824]: I0121 11:38:15.020304 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-7gtcg"] Jan 21 11:38:15 crc kubenswrapper[4824]: I0121 11:38:15.026822 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-7gtcg"] Jan 21 11:38:16 crc kubenswrapper[4824]: I0121 11:38:16.057006 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="135a8573-b55a-4c5f-9cb2-a7c3adea9720" path="/var/lib/kubelet/pods/135a8573-b55a-4c5f-9cb2-a7c3adea9720/volumes" Jan 21 11:38:17 crc kubenswrapper[4824]: I0121 11:38:17.049445 4824 scope.go:117] "RemoveContainer" containerID="1cd835fe36ebb6541740e5ab5fd843470d46dfb78975fd968e5cc9aaef462499" Jan 21 11:38:17 crc kubenswrapper[4824]: E0121 11:38:17.049685 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 11:38:20 crc kubenswrapper[4824]: I0121 11:38:20.729176 4824 generic.go:334] "Generic (PLEG): container finished" podID="890acdab-9f31-4afc-80e4-c1df308ac5a6" containerID="882b40080daa5269b988b3f409290f0fb60c6467f759ffaa24e78d2d41b4111c" exitCode=0 Jan 21 11:38:20 crc kubenswrapper[4824]: I0121 11:38:20.729254 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-8w57w" event={"ID":"890acdab-9f31-4afc-80e4-c1df308ac5a6","Type":"ContainerDied","Data":"882b40080daa5269b988b3f409290f0fb60c6467f759ffaa24e78d2d41b4111c"} Jan 21 11:38:22 crc kubenswrapper[4824]: I0121 11:38:22.056575 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-8w57w" Jan 21 11:38:22 crc kubenswrapper[4824]: I0121 11:38:22.098169 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d25d6\" (UniqueName: \"kubernetes.io/projected/890acdab-9f31-4afc-80e4-c1df308ac5a6-kube-api-access-d25d6\") pod \"890acdab-9f31-4afc-80e4-c1df308ac5a6\" (UID: \"890acdab-9f31-4afc-80e4-c1df308ac5a6\") " Jan 21 11:38:22 crc kubenswrapper[4824]: I0121 11:38:22.098213 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/890acdab-9f31-4afc-80e4-c1df308ac5a6-inventory\") pod \"890acdab-9f31-4afc-80e4-c1df308ac5a6\" (UID: \"890acdab-9f31-4afc-80e4-c1df308ac5a6\") " Jan 21 11:38:22 crc kubenswrapper[4824]: I0121 11:38:22.098247 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/890acdab-9f31-4afc-80e4-c1df308ac5a6-ssh-key-openstack-edpm-ipam\") pod \"890acdab-9f31-4afc-80e4-c1df308ac5a6\" (UID: \"890acdab-9f31-4afc-80e4-c1df308ac5a6\") " Jan 21 11:38:22 crc kubenswrapper[4824]: I0121 11:38:22.102509 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/890acdab-9f31-4afc-80e4-c1df308ac5a6-kube-api-access-d25d6" (OuterVolumeSpecName: "kube-api-access-d25d6") pod "890acdab-9f31-4afc-80e4-c1df308ac5a6" (UID: "890acdab-9f31-4afc-80e4-c1df308ac5a6"). InnerVolumeSpecName "kube-api-access-d25d6". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:38:22 crc kubenswrapper[4824]: I0121 11:38:22.119061 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/890acdab-9f31-4afc-80e4-c1df308ac5a6-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "890acdab-9f31-4afc-80e4-c1df308ac5a6" (UID: "890acdab-9f31-4afc-80e4-c1df308ac5a6"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:38:22 crc kubenswrapper[4824]: I0121 11:38:22.119107 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/890acdab-9f31-4afc-80e4-c1df308ac5a6-inventory" (OuterVolumeSpecName: "inventory") pod "890acdab-9f31-4afc-80e4-c1df308ac5a6" (UID: "890acdab-9f31-4afc-80e4-c1df308ac5a6"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:38:22 crc kubenswrapper[4824]: I0121 11:38:22.200609 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d25d6\" (UniqueName: \"kubernetes.io/projected/890acdab-9f31-4afc-80e4-c1df308ac5a6-kube-api-access-d25d6\") on node \"crc\" DevicePath \"\"" Jan 21 11:38:22 crc kubenswrapper[4824]: I0121 11:38:22.200633 4824 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/890acdab-9f31-4afc-80e4-c1df308ac5a6-inventory\") on node \"crc\" DevicePath \"\"" Jan 21 11:38:22 crc kubenswrapper[4824]: I0121 11:38:22.200643 4824 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/890acdab-9f31-4afc-80e4-c1df308ac5a6-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Jan 21 11:38:22 crc kubenswrapper[4824]: I0121 11:38:22.741283 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-8w57w" event={"ID":"890acdab-9f31-4afc-80e4-c1df308ac5a6","Type":"ContainerDied","Data":"b95a36f4b2a1281221a694900f274ed364cc1352c3fe01992775f1762a0a247c"} Jan 21 11:38:22 crc kubenswrapper[4824]: I0121 11:38:22.741316 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b95a36f4b2a1281221a694900f274ed364cc1352c3fe01992775f1762a0a247c" Jan 21 11:38:22 crc kubenswrapper[4824]: I0121 11:38:22.741533 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-8w57w" Jan 21 11:38:22 crc kubenswrapper[4824]: I0121 11:38:22.792499 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2"] Jan 21 11:38:22 crc kubenswrapper[4824]: E0121 11:38:22.792847 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="890acdab-9f31-4afc-80e4-c1df308ac5a6" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Jan 21 11:38:22 crc kubenswrapper[4824]: I0121 11:38:22.792866 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="890acdab-9f31-4afc-80e4-c1df308ac5a6" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Jan 21 11:38:22 crc kubenswrapper[4824]: I0121 11:38:22.793046 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="890acdab-9f31-4afc-80e4-c1df308ac5a6" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Jan 21 11:38:22 crc kubenswrapper[4824]: I0121 11:38:22.793580 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2" Jan 21 11:38:22 crc kubenswrapper[4824]: I0121 11:38:22.796523 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Jan 21 11:38:22 crc kubenswrapper[4824]: I0121 11:38:22.796901 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-telemetry-default-certs-0" Jan 21 11:38:22 crc kubenswrapper[4824]: I0121 11:38:22.796913 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Jan 21 11:38:22 crc kubenswrapper[4824]: I0121 11:38:22.797239 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-neutron-metadata-default-certs-0" Jan 21 11:38:22 crc kubenswrapper[4824]: I0121 11:38:22.797289 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-ovn-default-certs-0" Jan 21 11:38:22 crc kubenswrapper[4824]: I0121 11:38:22.797334 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-tcqk6" Jan 21 11:38:22 crc kubenswrapper[4824]: I0121 11:38:22.797350 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Jan 21 11:38:22 crc kubenswrapper[4824]: I0121 11:38:22.797586 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-libvirt-default-certs-0" Jan 21 11:38:22 crc kubenswrapper[4824]: I0121 11:38:22.804059 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2"] Jan 21 11:38:22 crc kubenswrapper[4824]: I0121 11:38:22.807091 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b821ea8d-9327-4623-b5be-3da7d3872cab-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2\" (UID: \"b821ea8d-9327-4623-b5be-3da7d3872cab\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2" Jan 21 11:38:22 crc kubenswrapper[4824]: I0121 11:38:22.807315 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b821ea8d-9327-4623-b5be-3da7d3872cab-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2\" (UID: \"b821ea8d-9327-4623-b5be-3da7d3872cab\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2" Jan 21 11:38:22 crc kubenswrapper[4824]: I0121 11:38:22.807339 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b821ea8d-9327-4623-b5be-3da7d3872cab-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2\" (UID: \"b821ea8d-9327-4623-b5be-3da7d3872cab\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2" Jan 21 11:38:22 crc kubenswrapper[4824]: I0121 11:38:22.807379 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b821ea8d-9327-4623-b5be-3da7d3872cab-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2\" (UID: \"b821ea8d-9327-4623-b5be-3da7d3872cab\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2" Jan 21 11:38:22 crc kubenswrapper[4824]: I0121 11:38:22.807461 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b821ea8d-9327-4623-b5be-3da7d3872cab-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2\" (UID: \"b821ea8d-9327-4623-b5be-3da7d3872cab\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2" Jan 21 11:38:22 crc kubenswrapper[4824]: I0121 11:38:22.807621 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b821ea8d-9327-4623-b5be-3da7d3872cab-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2\" (UID: \"b821ea8d-9327-4623-b5be-3da7d3872cab\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2" Jan 21 11:38:22 crc kubenswrapper[4824]: I0121 11:38:22.807713 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b821ea8d-9327-4623-b5be-3da7d3872cab-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2\" (UID: \"b821ea8d-9327-4623-b5be-3da7d3872cab\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2" Jan 21 11:38:22 crc kubenswrapper[4824]: I0121 11:38:22.807785 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dhxkm\" (UniqueName: \"kubernetes.io/projected/b821ea8d-9327-4623-b5be-3da7d3872cab-kube-api-access-dhxkm\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2\" (UID: \"b821ea8d-9327-4623-b5be-3da7d3872cab\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2" Jan 21 11:38:22 crc kubenswrapper[4824]: I0121 11:38:22.807927 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b821ea8d-9327-4623-b5be-3da7d3872cab-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2\" (UID: \"b821ea8d-9327-4623-b5be-3da7d3872cab\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2" Jan 21 11:38:22 crc kubenswrapper[4824]: I0121 11:38:22.808089 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b821ea8d-9327-4623-b5be-3da7d3872cab-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2\" (UID: \"b821ea8d-9327-4623-b5be-3da7d3872cab\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2" Jan 21 11:38:22 crc kubenswrapper[4824]: I0121 11:38:22.808339 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b821ea8d-9327-4623-b5be-3da7d3872cab-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2\" (UID: \"b821ea8d-9327-4623-b5be-3da7d3872cab\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2" Jan 21 11:38:22 crc kubenswrapper[4824]: I0121 11:38:22.808380 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b821ea8d-9327-4623-b5be-3da7d3872cab-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2\" (UID: \"b821ea8d-9327-4623-b5be-3da7d3872cab\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2" Jan 21 11:38:22 crc kubenswrapper[4824]: I0121 11:38:22.808412 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b821ea8d-9327-4623-b5be-3da7d3872cab-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2\" (UID: \"b821ea8d-9327-4623-b5be-3da7d3872cab\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2" Jan 21 11:38:22 crc kubenswrapper[4824]: I0121 11:38:22.808457 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/b821ea8d-9327-4623-b5be-3da7d3872cab-ssh-key-openstack-edpm-ipam\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2\" (UID: \"b821ea8d-9327-4623-b5be-3da7d3872cab\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2" Jan 21 11:38:22 crc kubenswrapper[4824]: I0121 11:38:22.909726 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b821ea8d-9327-4623-b5be-3da7d3872cab-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2\" (UID: \"b821ea8d-9327-4623-b5be-3da7d3872cab\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2" Jan 21 11:38:22 crc kubenswrapper[4824]: I0121 11:38:22.909910 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b821ea8d-9327-4623-b5be-3da7d3872cab-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2\" (UID: \"b821ea8d-9327-4623-b5be-3da7d3872cab\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2" Jan 21 11:38:22 crc kubenswrapper[4824]: I0121 11:38:22.910088 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b821ea8d-9327-4623-b5be-3da7d3872cab-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2\" (UID: \"b821ea8d-9327-4623-b5be-3da7d3872cab\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2" Jan 21 11:38:22 crc kubenswrapper[4824]: I0121 11:38:22.910166 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b821ea8d-9327-4623-b5be-3da7d3872cab-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2\" (UID: \"b821ea8d-9327-4623-b5be-3da7d3872cab\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2" Jan 21 11:38:22 crc kubenswrapper[4824]: I0121 11:38:22.910267 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/b821ea8d-9327-4623-b5be-3da7d3872cab-ssh-key-openstack-edpm-ipam\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2\" (UID: \"b821ea8d-9327-4623-b5be-3da7d3872cab\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2" Jan 21 11:38:22 crc kubenswrapper[4824]: I0121 11:38:22.910375 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b821ea8d-9327-4623-b5be-3da7d3872cab-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2\" (UID: \"b821ea8d-9327-4623-b5be-3da7d3872cab\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2" Jan 21 11:38:22 crc kubenswrapper[4824]: I0121 11:38:22.910460 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b821ea8d-9327-4623-b5be-3da7d3872cab-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2\" (UID: \"b821ea8d-9327-4623-b5be-3da7d3872cab\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2" Jan 21 11:38:22 crc kubenswrapper[4824]: I0121 11:38:22.910535 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b821ea8d-9327-4623-b5be-3da7d3872cab-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2\" (UID: \"b821ea8d-9327-4623-b5be-3da7d3872cab\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2" Jan 21 11:38:22 crc kubenswrapper[4824]: I0121 11:38:22.910613 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b821ea8d-9327-4623-b5be-3da7d3872cab-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2\" (UID: \"b821ea8d-9327-4623-b5be-3da7d3872cab\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2" Jan 21 11:38:22 crc kubenswrapper[4824]: I0121 11:38:22.910697 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b821ea8d-9327-4623-b5be-3da7d3872cab-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2\" (UID: \"b821ea8d-9327-4623-b5be-3da7d3872cab\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2" Jan 21 11:38:22 crc kubenswrapper[4824]: I0121 11:38:22.910791 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b821ea8d-9327-4623-b5be-3da7d3872cab-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2\" (UID: \"b821ea8d-9327-4623-b5be-3da7d3872cab\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2" Jan 21 11:38:22 crc kubenswrapper[4824]: I0121 11:38:22.910864 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b821ea8d-9327-4623-b5be-3da7d3872cab-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2\" (UID: \"b821ea8d-9327-4623-b5be-3da7d3872cab\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2" Jan 21 11:38:22 crc kubenswrapper[4824]: I0121 11:38:22.910937 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dhxkm\" (UniqueName: \"kubernetes.io/projected/b821ea8d-9327-4623-b5be-3da7d3872cab-kube-api-access-dhxkm\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2\" (UID: \"b821ea8d-9327-4623-b5be-3da7d3872cab\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2" Jan 21 11:38:22 crc kubenswrapper[4824]: I0121 11:38:22.911035 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b821ea8d-9327-4623-b5be-3da7d3872cab-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2\" (UID: \"b821ea8d-9327-4623-b5be-3da7d3872cab\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2" Jan 21 11:38:22 crc kubenswrapper[4824]: I0121 11:38:22.913571 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b821ea8d-9327-4623-b5be-3da7d3872cab-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2\" (UID: \"b821ea8d-9327-4623-b5be-3da7d3872cab\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2" Jan 21 11:38:22 crc kubenswrapper[4824]: I0121 11:38:22.913586 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/b821ea8d-9327-4623-b5be-3da7d3872cab-ssh-key-openstack-edpm-ipam\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2\" (UID: \"b821ea8d-9327-4623-b5be-3da7d3872cab\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2" Jan 21 11:38:22 crc kubenswrapper[4824]: I0121 11:38:22.913900 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b821ea8d-9327-4623-b5be-3da7d3872cab-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2\" (UID: \"b821ea8d-9327-4623-b5be-3da7d3872cab\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2" Jan 21 11:38:22 crc kubenswrapper[4824]: I0121 11:38:22.914040 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b821ea8d-9327-4623-b5be-3da7d3872cab-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2\" (UID: \"b821ea8d-9327-4623-b5be-3da7d3872cab\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2" Jan 21 11:38:22 crc kubenswrapper[4824]: I0121 11:38:22.914196 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b821ea8d-9327-4623-b5be-3da7d3872cab-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2\" (UID: \"b821ea8d-9327-4623-b5be-3da7d3872cab\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2" Jan 21 11:38:22 crc kubenswrapper[4824]: I0121 11:38:22.914300 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b821ea8d-9327-4623-b5be-3da7d3872cab-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2\" (UID: \"b821ea8d-9327-4623-b5be-3da7d3872cab\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2" Jan 21 11:38:22 crc kubenswrapper[4824]: I0121 11:38:22.914591 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b821ea8d-9327-4623-b5be-3da7d3872cab-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2\" (UID: \"b821ea8d-9327-4623-b5be-3da7d3872cab\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2" Jan 21 11:38:22 crc kubenswrapper[4824]: I0121 11:38:22.914600 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b821ea8d-9327-4623-b5be-3da7d3872cab-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2\" (UID: \"b821ea8d-9327-4623-b5be-3da7d3872cab\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2" Jan 21 11:38:22 crc kubenswrapper[4824]: I0121 11:38:22.915599 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b821ea8d-9327-4623-b5be-3da7d3872cab-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2\" (UID: \"b821ea8d-9327-4623-b5be-3da7d3872cab\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2" Jan 21 11:38:22 crc kubenswrapper[4824]: I0121 11:38:22.915847 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b821ea8d-9327-4623-b5be-3da7d3872cab-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2\" (UID: \"b821ea8d-9327-4623-b5be-3da7d3872cab\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2" Jan 21 11:38:22 crc kubenswrapper[4824]: I0121 11:38:22.916582 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b821ea8d-9327-4623-b5be-3da7d3872cab-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2\" (UID: \"b821ea8d-9327-4623-b5be-3da7d3872cab\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2" Jan 21 11:38:22 crc kubenswrapper[4824]: I0121 11:38:22.917428 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b821ea8d-9327-4623-b5be-3da7d3872cab-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2\" (UID: \"b821ea8d-9327-4623-b5be-3da7d3872cab\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2" Jan 21 11:38:22 crc kubenswrapper[4824]: I0121 11:38:22.917538 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b821ea8d-9327-4623-b5be-3da7d3872cab-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2\" (UID: \"b821ea8d-9327-4623-b5be-3da7d3872cab\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2" Jan 21 11:38:22 crc kubenswrapper[4824]: I0121 11:38:22.926031 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dhxkm\" (UniqueName: \"kubernetes.io/projected/b821ea8d-9327-4623-b5be-3da7d3872cab-kube-api-access-dhxkm\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2\" (UID: \"b821ea8d-9327-4623-b5be-3da7d3872cab\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2" Jan 21 11:38:23 crc kubenswrapper[4824]: I0121 11:38:23.109050 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2" Jan 21 11:38:23 crc kubenswrapper[4824]: I0121 11:38:23.513654 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2"] Jan 21 11:38:23 crc kubenswrapper[4824]: I0121 11:38:23.748477 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2" event={"ID":"b821ea8d-9327-4623-b5be-3da7d3872cab","Type":"ContainerStarted","Data":"b432a7ecf664fae2f795f90960df62d71c3a7f5305cc1480692283386292bf1c"} Jan 21 11:38:24 crc kubenswrapper[4824]: I0121 11:38:24.757327 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2" event={"ID":"b821ea8d-9327-4623-b5be-3da7d3872cab","Type":"ContainerStarted","Data":"4463df9b5218899c21b0867a3bc9c0284f5c2175ab91fef1df2e89365b5d4cc1"} Jan 21 11:38:24 crc kubenswrapper[4824]: I0121 11:38:24.771632 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2" podStartSLOduration=2.2999615970000002 podStartE2EDuration="2.7716189s" podCreationTimestamp="2026-01-21 11:38:22 +0000 UTC" firstStartedPulling="2026-01-21 11:38:23.515542889 +0000 UTC m=+1645.808572181" lastFinishedPulling="2026-01-21 11:38:23.987200192 +0000 UTC m=+1646.280229484" observedRunningTime="2026-01-21 11:38:24.769013481 +0000 UTC m=+1647.062042774" watchObservedRunningTime="2026-01-21 11:38:24.7716189 +0000 UTC m=+1647.064648192" Jan 21 11:38:29 crc kubenswrapper[4824]: I0121 11:38:29.050006 4824 scope.go:117] "RemoveContainer" containerID="1cd835fe36ebb6541740e5ab5fd843470d46dfb78975fd968e5cc9aaef462499" Jan 21 11:38:29 crc kubenswrapper[4824]: E0121 11:38:29.051274 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 11:38:43 crc kubenswrapper[4824]: I0121 11:38:43.048982 4824 scope.go:117] "RemoveContainer" containerID="1cd835fe36ebb6541740e5ab5fd843470d46dfb78975fd968e5cc9aaef462499" Jan 21 11:38:43 crc kubenswrapper[4824]: E0121 11:38:43.049735 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 11:38:50 crc kubenswrapper[4824]: I0121 11:38:50.935288 4824 generic.go:334] "Generic (PLEG): container finished" podID="b821ea8d-9327-4623-b5be-3da7d3872cab" containerID="4463df9b5218899c21b0867a3bc9c0284f5c2175ab91fef1df2e89365b5d4cc1" exitCode=0 Jan 21 11:38:50 crc kubenswrapper[4824]: I0121 11:38:50.935369 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2" event={"ID":"b821ea8d-9327-4623-b5be-3da7d3872cab","Type":"ContainerDied","Data":"4463df9b5218899c21b0867a3bc9c0284f5c2175ab91fef1df2e89365b5d4cc1"} Jan 21 11:38:52 crc kubenswrapper[4824]: I0121 11:38:52.245125 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2" Jan 21 11:38:52 crc kubenswrapper[4824]: I0121 11:38:52.335216 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b821ea8d-9327-4623-b5be-3da7d3872cab-libvirt-combined-ca-bundle\") pod \"b821ea8d-9327-4623-b5be-3da7d3872cab\" (UID: \"b821ea8d-9327-4623-b5be-3da7d3872cab\") " Jan 21 11:38:52 crc kubenswrapper[4824]: I0121 11:38:52.335350 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b821ea8d-9327-4623-b5be-3da7d3872cab-ovn-combined-ca-bundle\") pod \"b821ea8d-9327-4623-b5be-3da7d3872cab\" (UID: \"b821ea8d-9327-4623-b5be-3da7d3872cab\") " Jan 21 11:38:52 crc kubenswrapper[4824]: I0121 11:38:52.335370 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b821ea8d-9327-4623-b5be-3da7d3872cab-telemetry-combined-ca-bundle\") pod \"b821ea8d-9327-4623-b5be-3da7d3872cab\" (UID: \"b821ea8d-9327-4623-b5be-3da7d3872cab\") " Jan 21 11:38:52 crc kubenswrapper[4824]: I0121 11:38:52.335385 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b821ea8d-9327-4623-b5be-3da7d3872cab-repo-setup-combined-ca-bundle\") pod \"b821ea8d-9327-4623-b5be-3da7d3872cab\" (UID: \"b821ea8d-9327-4623-b5be-3da7d3872cab\") " Jan 21 11:38:52 crc kubenswrapper[4824]: I0121 11:38:52.335401 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/b821ea8d-9327-4623-b5be-3da7d3872cab-ssh-key-openstack-edpm-ipam\") pod \"b821ea8d-9327-4623-b5be-3da7d3872cab\" (UID: \"b821ea8d-9327-4623-b5be-3da7d3872cab\") " Jan 21 11:38:52 crc kubenswrapper[4824]: I0121 11:38:52.335417 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b821ea8d-9327-4623-b5be-3da7d3872cab-openstack-edpm-ipam-ovn-default-certs-0\") pod \"b821ea8d-9327-4623-b5be-3da7d3872cab\" (UID: \"b821ea8d-9327-4623-b5be-3da7d3872cab\") " Jan 21 11:38:52 crc kubenswrapper[4824]: I0121 11:38:52.335447 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b821ea8d-9327-4623-b5be-3da7d3872cab-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"b821ea8d-9327-4623-b5be-3da7d3872cab\" (UID: \"b821ea8d-9327-4623-b5be-3da7d3872cab\") " Jan 21 11:38:52 crc kubenswrapper[4824]: I0121 11:38:52.335468 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b821ea8d-9327-4623-b5be-3da7d3872cab-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"b821ea8d-9327-4623-b5be-3da7d3872cab\" (UID: \"b821ea8d-9327-4623-b5be-3da7d3872cab\") " Jan 21 11:38:52 crc kubenswrapper[4824]: I0121 11:38:52.335525 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b821ea8d-9327-4623-b5be-3da7d3872cab-nova-combined-ca-bundle\") pod \"b821ea8d-9327-4623-b5be-3da7d3872cab\" (UID: \"b821ea8d-9327-4623-b5be-3da7d3872cab\") " Jan 21 11:38:52 crc kubenswrapper[4824]: I0121 11:38:52.335569 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b821ea8d-9327-4623-b5be-3da7d3872cab-neutron-metadata-combined-ca-bundle\") pod \"b821ea8d-9327-4623-b5be-3da7d3872cab\" (UID: \"b821ea8d-9327-4623-b5be-3da7d3872cab\") " Jan 21 11:38:52 crc kubenswrapper[4824]: I0121 11:38:52.335605 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b821ea8d-9327-4623-b5be-3da7d3872cab-bootstrap-combined-ca-bundle\") pod \"b821ea8d-9327-4623-b5be-3da7d3872cab\" (UID: \"b821ea8d-9327-4623-b5be-3da7d3872cab\") " Jan 21 11:38:52 crc kubenswrapper[4824]: I0121 11:38:52.335644 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dhxkm\" (UniqueName: \"kubernetes.io/projected/b821ea8d-9327-4623-b5be-3da7d3872cab-kube-api-access-dhxkm\") pod \"b821ea8d-9327-4623-b5be-3da7d3872cab\" (UID: \"b821ea8d-9327-4623-b5be-3da7d3872cab\") " Jan 21 11:38:52 crc kubenswrapper[4824]: I0121 11:38:52.335684 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b821ea8d-9327-4623-b5be-3da7d3872cab-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"b821ea8d-9327-4623-b5be-3da7d3872cab\" (UID: \"b821ea8d-9327-4623-b5be-3da7d3872cab\") " Jan 21 11:38:52 crc kubenswrapper[4824]: I0121 11:38:52.335704 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b821ea8d-9327-4623-b5be-3da7d3872cab-inventory\") pod \"b821ea8d-9327-4623-b5be-3da7d3872cab\" (UID: \"b821ea8d-9327-4623-b5be-3da7d3872cab\") " Jan 21 11:38:52 crc kubenswrapper[4824]: I0121 11:38:52.340455 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b821ea8d-9327-4623-b5be-3da7d3872cab-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "b821ea8d-9327-4623-b5be-3da7d3872cab" (UID: "b821ea8d-9327-4623-b5be-3da7d3872cab"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:38:52 crc kubenswrapper[4824]: I0121 11:38:52.340778 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b821ea8d-9327-4623-b5be-3da7d3872cab-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "b821ea8d-9327-4623-b5be-3da7d3872cab" (UID: "b821ea8d-9327-4623-b5be-3da7d3872cab"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:38:52 crc kubenswrapper[4824]: I0121 11:38:52.341229 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b821ea8d-9327-4623-b5be-3da7d3872cab-openstack-edpm-ipam-ovn-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-ovn-default-certs-0") pod "b821ea8d-9327-4623-b5be-3da7d3872cab" (UID: "b821ea8d-9327-4623-b5be-3da7d3872cab"). InnerVolumeSpecName "openstack-edpm-ipam-ovn-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:38:52 crc kubenswrapper[4824]: I0121 11:38:52.341354 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b821ea8d-9327-4623-b5be-3da7d3872cab-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "b821ea8d-9327-4623-b5be-3da7d3872cab" (UID: "b821ea8d-9327-4623-b5be-3da7d3872cab"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:38:52 crc kubenswrapper[4824]: I0121 11:38:52.341496 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b821ea8d-9327-4623-b5be-3da7d3872cab-openstack-edpm-ipam-telemetry-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-telemetry-default-certs-0") pod "b821ea8d-9327-4623-b5be-3da7d3872cab" (UID: "b821ea8d-9327-4623-b5be-3da7d3872cab"). InnerVolumeSpecName "openstack-edpm-ipam-telemetry-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:38:52 crc kubenswrapper[4824]: I0121 11:38:52.341517 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b821ea8d-9327-4623-b5be-3da7d3872cab-kube-api-access-dhxkm" (OuterVolumeSpecName: "kube-api-access-dhxkm") pod "b821ea8d-9327-4623-b5be-3da7d3872cab" (UID: "b821ea8d-9327-4623-b5be-3da7d3872cab"). InnerVolumeSpecName "kube-api-access-dhxkm". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:38:52 crc kubenswrapper[4824]: I0121 11:38:52.341720 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b821ea8d-9327-4623-b5be-3da7d3872cab-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "b821ea8d-9327-4623-b5be-3da7d3872cab" (UID: "b821ea8d-9327-4623-b5be-3da7d3872cab"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:38:52 crc kubenswrapper[4824]: I0121 11:38:52.341739 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b821ea8d-9327-4623-b5be-3da7d3872cab-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "b821ea8d-9327-4623-b5be-3da7d3872cab" (UID: "b821ea8d-9327-4623-b5be-3da7d3872cab"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:38:52 crc kubenswrapper[4824]: I0121 11:38:52.341755 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b821ea8d-9327-4623-b5be-3da7d3872cab-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "b821ea8d-9327-4623-b5be-3da7d3872cab" (UID: "b821ea8d-9327-4623-b5be-3da7d3872cab"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:38:52 crc kubenswrapper[4824]: I0121 11:38:52.342640 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b821ea8d-9327-4623-b5be-3da7d3872cab-openstack-edpm-ipam-libvirt-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-libvirt-default-certs-0") pod "b821ea8d-9327-4623-b5be-3da7d3872cab" (UID: "b821ea8d-9327-4623-b5be-3da7d3872cab"). InnerVolumeSpecName "openstack-edpm-ipam-libvirt-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:38:52 crc kubenswrapper[4824]: I0121 11:38:52.343289 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b821ea8d-9327-4623-b5be-3da7d3872cab-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "b821ea8d-9327-4623-b5be-3da7d3872cab" (UID: "b821ea8d-9327-4623-b5be-3da7d3872cab"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:38:52 crc kubenswrapper[4824]: I0121 11:38:52.343656 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b821ea8d-9327-4623-b5be-3da7d3872cab-openstack-edpm-ipam-neutron-metadata-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-neutron-metadata-default-certs-0") pod "b821ea8d-9327-4623-b5be-3da7d3872cab" (UID: "b821ea8d-9327-4623-b5be-3da7d3872cab"). InnerVolumeSpecName "openstack-edpm-ipam-neutron-metadata-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:38:52 crc kubenswrapper[4824]: E0121 11:38:52.355863 4824 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b821ea8d-9327-4623-b5be-3da7d3872cab-ssh-key-openstack-edpm-ipam podName:b821ea8d-9327-4623-b5be-3da7d3872cab nodeName:}" failed. No retries permitted until 2026-01-21 11:38:52.855840133 +0000 UTC m=+1675.148869425 (durationBeforeRetry 500ms). Error: error cleaning subPath mounts for volume "ssh-key-openstack-edpm-ipam" (UniqueName: "kubernetes.io/secret/b821ea8d-9327-4623-b5be-3da7d3872cab-ssh-key-openstack-edpm-ipam") pod "b821ea8d-9327-4623-b5be-3da7d3872cab" (UID: "b821ea8d-9327-4623-b5be-3da7d3872cab") : error deleting /var/lib/kubelet/pods/b821ea8d-9327-4623-b5be-3da7d3872cab/volume-subpaths: remove /var/lib/kubelet/pods/b821ea8d-9327-4623-b5be-3da7d3872cab/volume-subpaths: no such file or directory Jan 21 11:38:52 crc kubenswrapper[4824]: I0121 11:38:52.358470 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b821ea8d-9327-4623-b5be-3da7d3872cab-inventory" (OuterVolumeSpecName: "inventory") pod "b821ea8d-9327-4623-b5be-3da7d3872cab" (UID: "b821ea8d-9327-4623-b5be-3da7d3872cab"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:38:52 crc kubenswrapper[4824]: I0121 11:38:52.438088 4824 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b821ea8d-9327-4623-b5be-3da7d3872cab-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 21 11:38:52 crc kubenswrapper[4824]: I0121 11:38:52.438116 4824 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b821ea8d-9327-4623-b5be-3da7d3872cab-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 21 11:38:52 crc kubenswrapper[4824]: I0121 11:38:52.438129 4824 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b821ea8d-9327-4623-b5be-3da7d3872cab-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 21 11:38:52 crc kubenswrapper[4824]: I0121 11:38:52.438140 4824 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b821ea8d-9327-4623-b5be-3da7d3872cab-openstack-edpm-ipam-ovn-default-certs-0\") on node \"crc\" DevicePath \"\"" Jan 21 11:38:52 crc kubenswrapper[4824]: I0121 11:38:52.438151 4824 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b821ea8d-9327-4623-b5be-3da7d3872cab-openstack-edpm-ipam-telemetry-default-certs-0\") on node \"crc\" DevicePath \"\"" Jan 21 11:38:52 crc kubenswrapper[4824]: I0121 11:38:52.438161 4824 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b821ea8d-9327-4623-b5be-3da7d3872cab-openstack-edpm-ipam-libvirt-default-certs-0\") on node \"crc\" DevicePath \"\"" Jan 21 11:38:52 crc kubenswrapper[4824]: I0121 11:38:52.438169 4824 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b821ea8d-9327-4623-b5be-3da7d3872cab-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 21 11:38:52 crc kubenswrapper[4824]: I0121 11:38:52.438177 4824 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b821ea8d-9327-4623-b5be-3da7d3872cab-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 21 11:38:52 crc kubenswrapper[4824]: I0121 11:38:52.438185 4824 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b821ea8d-9327-4623-b5be-3da7d3872cab-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 21 11:38:52 crc kubenswrapper[4824]: I0121 11:38:52.438193 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dhxkm\" (UniqueName: \"kubernetes.io/projected/b821ea8d-9327-4623-b5be-3da7d3872cab-kube-api-access-dhxkm\") on node \"crc\" DevicePath \"\"" Jan 21 11:38:52 crc kubenswrapper[4824]: I0121 11:38:52.438201 4824 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b821ea8d-9327-4623-b5be-3da7d3872cab-openstack-edpm-ipam-neutron-metadata-default-certs-0\") on node \"crc\" DevicePath \"\"" Jan 21 11:38:52 crc kubenswrapper[4824]: I0121 11:38:52.438211 4824 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b821ea8d-9327-4623-b5be-3da7d3872cab-inventory\") on node \"crc\" DevicePath \"\"" Jan 21 11:38:52 crc kubenswrapper[4824]: I0121 11:38:52.438220 4824 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b821ea8d-9327-4623-b5be-3da7d3872cab-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 21 11:38:52 crc kubenswrapper[4824]: I0121 11:38:52.945864 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/b821ea8d-9327-4623-b5be-3da7d3872cab-ssh-key-openstack-edpm-ipam\") pod \"b821ea8d-9327-4623-b5be-3da7d3872cab\" (UID: \"b821ea8d-9327-4623-b5be-3da7d3872cab\") " Jan 21 11:38:52 crc kubenswrapper[4824]: I0121 11:38:52.950128 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b821ea8d-9327-4623-b5be-3da7d3872cab-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "b821ea8d-9327-4623-b5be-3da7d3872cab" (UID: "b821ea8d-9327-4623-b5be-3da7d3872cab"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:38:52 crc kubenswrapper[4824]: I0121 11:38:52.954384 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2" event={"ID":"b821ea8d-9327-4623-b5be-3da7d3872cab","Type":"ContainerDied","Data":"b432a7ecf664fae2f795f90960df62d71c3a7f5305cc1480692283386292bf1c"} Jan 21 11:38:52 crc kubenswrapper[4824]: I0121 11:38:52.954427 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b432a7ecf664fae2f795f90960df62d71c3a7f5305cc1480692283386292bf1c" Jan 21 11:38:52 crc kubenswrapper[4824]: I0121 11:38:52.954481 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2" Jan 21 11:38:53 crc kubenswrapper[4824]: I0121 11:38:53.031319 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-gf4zq"] Jan 21 11:38:53 crc kubenswrapper[4824]: E0121 11:38:53.031845 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b821ea8d-9327-4623-b5be-3da7d3872cab" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Jan 21 11:38:53 crc kubenswrapper[4824]: I0121 11:38:53.031863 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="b821ea8d-9327-4623-b5be-3da7d3872cab" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Jan 21 11:38:53 crc kubenswrapper[4824]: I0121 11:38:53.032075 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="b821ea8d-9327-4623-b5be-3da7d3872cab" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Jan 21 11:38:53 crc kubenswrapper[4824]: I0121 11:38:53.032635 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gf4zq" Jan 21 11:38:53 crc kubenswrapper[4824]: I0121 11:38:53.034205 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Jan 21 11:38:53 crc kubenswrapper[4824]: I0121 11:38:53.034434 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-tcqk6" Jan 21 11:38:53 crc kubenswrapper[4824]: I0121 11:38:53.035416 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Jan 21 11:38:53 crc kubenswrapper[4824]: I0121 11:38:53.035687 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Jan 21 11:38:53 crc kubenswrapper[4824]: I0121 11:38:53.037082 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Jan 21 11:38:53 crc kubenswrapper[4824]: I0121 11:38:53.038700 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-gf4zq"] Jan 21 11:38:53 crc kubenswrapper[4824]: I0121 11:38:53.052485 4824 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/b821ea8d-9327-4623-b5be-3da7d3872cab-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Jan 21 11:38:53 crc kubenswrapper[4824]: I0121 11:38:53.153664 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18e65eb9-3b96-48cd-88b2-9f7fb0e54b6f-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-gf4zq\" (UID: \"18e65eb9-3b96-48cd-88b2-9f7fb0e54b6f\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gf4zq" Jan 21 11:38:53 crc kubenswrapper[4824]: I0121 11:38:53.154002 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/18e65eb9-3b96-48cd-88b2-9f7fb0e54b6f-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-gf4zq\" (UID: \"18e65eb9-3b96-48cd-88b2-9f7fb0e54b6f\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gf4zq" Jan 21 11:38:53 crc kubenswrapper[4824]: I0121 11:38:53.154071 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-md7w4\" (UniqueName: \"kubernetes.io/projected/18e65eb9-3b96-48cd-88b2-9f7fb0e54b6f-kube-api-access-md7w4\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-gf4zq\" (UID: \"18e65eb9-3b96-48cd-88b2-9f7fb0e54b6f\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gf4zq" Jan 21 11:38:53 crc kubenswrapper[4824]: I0121 11:38:53.154147 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/18e65eb9-3b96-48cd-88b2-9f7fb0e54b6f-ssh-key-openstack-edpm-ipam\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-gf4zq\" (UID: \"18e65eb9-3b96-48cd-88b2-9f7fb0e54b6f\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gf4zq" Jan 21 11:38:53 crc kubenswrapper[4824]: I0121 11:38:53.154200 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/18e65eb9-3b96-48cd-88b2-9f7fb0e54b6f-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-gf4zq\" (UID: \"18e65eb9-3b96-48cd-88b2-9f7fb0e54b6f\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gf4zq" Jan 21 11:38:53 crc kubenswrapper[4824]: I0121 11:38:53.256112 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/18e65eb9-3b96-48cd-88b2-9f7fb0e54b6f-ssh-key-openstack-edpm-ipam\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-gf4zq\" (UID: \"18e65eb9-3b96-48cd-88b2-9f7fb0e54b6f\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gf4zq" Jan 21 11:38:53 crc kubenswrapper[4824]: I0121 11:38:53.256181 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/18e65eb9-3b96-48cd-88b2-9f7fb0e54b6f-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-gf4zq\" (UID: \"18e65eb9-3b96-48cd-88b2-9f7fb0e54b6f\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gf4zq" Jan 21 11:38:53 crc kubenswrapper[4824]: I0121 11:38:53.256228 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18e65eb9-3b96-48cd-88b2-9f7fb0e54b6f-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-gf4zq\" (UID: \"18e65eb9-3b96-48cd-88b2-9f7fb0e54b6f\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gf4zq" Jan 21 11:38:53 crc kubenswrapper[4824]: I0121 11:38:53.256455 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/18e65eb9-3b96-48cd-88b2-9f7fb0e54b6f-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-gf4zq\" (UID: \"18e65eb9-3b96-48cd-88b2-9f7fb0e54b6f\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gf4zq" Jan 21 11:38:53 crc kubenswrapper[4824]: I0121 11:38:53.256503 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-md7w4\" (UniqueName: \"kubernetes.io/projected/18e65eb9-3b96-48cd-88b2-9f7fb0e54b6f-kube-api-access-md7w4\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-gf4zq\" (UID: \"18e65eb9-3b96-48cd-88b2-9f7fb0e54b6f\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gf4zq" Jan 21 11:38:53 crc kubenswrapper[4824]: I0121 11:38:53.257329 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/18e65eb9-3b96-48cd-88b2-9f7fb0e54b6f-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-gf4zq\" (UID: \"18e65eb9-3b96-48cd-88b2-9f7fb0e54b6f\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gf4zq" Jan 21 11:38:53 crc kubenswrapper[4824]: I0121 11:38:53.259707 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/18e65eb9-3b96-48cd-88b2-9f7fb0e54b6f-ssh-key-openstack-edpm-ipam\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-gf4zq\" (UID: \"18e65eb9-3b96-48cd-88b2-9f7fb0e54b6f\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gf4zq" Jan 21 11:38:53 crc kubenswrapper[4824]: I0121 11:38:53.259918 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/18e65eb9-3b96-48cd-88b2-9f7fb0e54b6f-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-gf4zq\" (UID: \"18e65eb9-3b96-48cd-88b2-9f7fb0e54b6f\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gf4zq" Jan 21 11:38:53 crc kubenswrapper[4824]: I0121 11:38:53.260917 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18e65eb9-3b96-48cd-88b2-9f7fb0e54b6f-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-gf4zq\" (UID: \"18e65eb9-3b96-48cd-88b2-9f7fb0e54b6f\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gf4zq" Jan 21 11:38:53 crc kubenswrapper[4824]: I0121 11:38:53.270990 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-md7w4\" (UniqueName: \"kubernetes.io/projected/18e65eb9-3b96-48cd-88b2-9f7fb0e54b6f-kube-api-access-md7w4\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-gf4zq\" (UID: \"18e65eb9-3b96-48cd-88b2-9f7fb0e54b6f\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gf4zq" Jan 21 11:38:53 crc kubenswrapper[4824]: I0121 11:38:53.347829 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gf4zq" Jan 21 11:38:53 crc kubenswrapper[4824]: I0121 11:38:53.770789 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-gf4zq"] Jan 21 11:38:53 crc kubenswrapper[4824]: I0121 11:38:53.967011 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gf4zq" event={"ID":"18e65eb9-3b96-48cd-88b2-9f7fb0e54b6f","Type":"ContainerStarted","Data":"6d455d5215fcbfd8a47efb176dd7eb21ea2bd57fe6e56abd130fecb774e16899"} Jan 21 11:38:54 crc kubenswrapper[4824]: I0121 11:38:54.973602 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gf4zq" event={"ID":"18e65eb9-3b96-48cd-88b2-9f7fb0e54b6f","Type":"ContainerStarted","Data":"2ed717f75906d3a0f34ff5910051eb7ff1d604b1c193192f6557d69d614d1475"} Jan 21 11:38:54 crc kubenswrapper[4824]: I0121 11:38:54.990558 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gf4zq" podStartSLOduration=1.47454299 podStartE2EDuration="1.990539079s" podCreationTimestamp="2026-01-21 11:38:53 +0000 UTC" firstStartedPulling="2026-01-21 11:38:53.773767981 +0000 UTC m=+1676.066797273" lastFinishedPulling="2026-01-21 11:38:54.28976407 +0000 UTC m=+1676.582793362" observedRunningTime="2026-01-21 11:38:54.98405149 +0000 UTC m=+1677.277080783" watchObservedRunningTime="2026-01-21 11:38:54.990539079 +0000 UTC m=+1677.283568372" Jan 21 11:38:56 crc kubenswrapper[4824]: I0121 11:38:56.049375 4824 scope.go:117] "RemoveContainer" containerID="1cd835fe36ebb6541740e5ab5fd843470d46dfb78975fd968e5cc9aaef462499" Jan 21 11:38:56 crc kubenswrapper[4824]: E0121 11:38:56.050142 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 11:38:58 crc kubenswrapper[4824]: I0121 11:38:58.031508 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-6c642"] Jan 21 11:38:58 crc kubenswrapper[4824]: I0121 11:38:58.038230 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-6c642"] Jan 21 11:38:58 crc kubenswrapper[4824]: I0121 11:38:58.057162 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="12c162ff-82c2-4166-8a44-92e0455cae39" path="/var/lib/kubelet/pods/12c162ff-82c2-4166-8a44-92e0455cae39/volumes" Jan 21 11:39:00 crc kubenswrapper[4824]: I0121 11:39:00.512207 4824 scope.go:117] "RemoveContainer" containerID="e6d7bd7f99afa7be3d7eee7ba3147eedee78587b76d733c8cb546689d1b785fe" Jan 21 11:39:00 crc kubenswrapper[4824]: I0121 11:39:00.553468 4824 scope.go:117] "RemoveContainer" containerID="dd76c88e84c244299a1fbb85cad5296bd9ab946000d0d268303768f60e22c9c4" Jan 21 11:39:00 crc kubenswrapper[4824]: I0121 11:39:00.587831 4824 scope.go:117] "RemoveContainer" containerID="7468f8bf7c17d51a36aabd9a1f3d9c19a0cb1f2b474c4bfcd676598418dd86f7" Jan 21 11:39:07 crc kubenswrapper[4824]: I0121 11:39:07.048913 4824 scope.go:117] "RemoveContainer" containerID="1cd835fe36ebb6541740e5ab5fd843470d46dfb78975fd968e5cc9aaef462499" Jan 21 11:39:07 crc kubenswrapper[4824]: E0121 11:39:07.049754 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 11:39:18 crc kubenswrapper[4824]: I0121 11:39:18.053531 4824 scope.go:117] "RemoveContainer" containerID="1cd835fe36ebb6541740e5ab5fd843470d46dfb78975fd968e5cc9aaef462499" Jan 21 11:39:18 crc kubenswrapper[4824]: E0121 11:39:18.054245 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 11:39:29 crc kubenswrapper[4824]: I0121 11:39:29.049831 4824 scope.go:117] "RemoveContainer" containerID="1cd835fe36ebb6541740e5ab5fd843470d46dfb78975fd968e5cc9aaef462499" Jan 21 11:39:29 crc kubenswrapper[4824]: E0121 11:39:29.050990 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 11:39:38 crc kubenswrapper[4824]: I0121 11:39:38.244584 4824 generic.go:334] "Generic (PLEG): container finished" podID="18e65eb9-3b96-48cd-88b2-9f7fb0e54b6f" containerID="2ed717f75906d3a0f34ff5910051eb7ff1d604b1c193192f6557d69d614d1475" exitCode=0 Jan 21 11:39:38 crc kubenswrapper[4824]: I0121 11:39:38.244678 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gf4zq" event={"ID":"18e65eb9-3b96-48cd-88b2-9f7fb0e54b6f","Type":"ContainerDied","Data":"2ed717f75906d3a0f34ff5910051eb7ff1d604b1c193192f6557d69d614d1475"} Jan 21 11:39:39 crc kubenswrapper[4824]: I0121 11:39:39.566766 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gf4zq" Jan 21 11:39:39 crc kubenswrapper[4824]: I0121 11:39:39.661907 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/18e65eb9-3b96-48cd-88b2-9f7fb0e54b6f-inventory\") pod \"18e65eb9-3b96-48cd-88b2-9f7fb0e54b6f\" (UID: \"18e65eb9-3b96-48cd-88b2-9f7fb0e54b6f\") " Jan 21 11:39:39 crc kubenswrapper[4824]: I0121 11:39:39.661984 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18e65eb9-3b96-48cd-88b2-9f7fb0e54b6f-ovn-combined-ca-bundle\") pod \"18e65eb9-3b96-48cd-88b2-9f7fb0e54b6f\" (UID: \"18e65eb9-3b96-48cd-88b2-9f7fb0e54b6f\") " Jan 21 11:39:39 crc kubenswrapper[4824]: I0121 11:39:39.662142 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/18e65eb9-3b96-48cd-88b2-9f7fb0e54b6f-ovncontroller-config-0\") pod \"18e65eb9-3b96-48cd-88b2-9f7fb0e54b6f\" (UID: \"18e65eb9-3b96-48cd-88b2-9f7fb0e54b6f\") " Jan 21 11:39:39 crc kubenswrapper[4824]: I0121 11:39:39.662181 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-md7w4\" (UniqueName: \"kubernetes.io/projected/18e65eb9-3b96-48cd-88b2-9f7fb0e54b6f-kube-api-access-md7w4\") pod \"18e65eb9-3b96-48cd-88b2-9f7fb0e54b6f\" (UID: \"18e65eb9-3b96-48cd-88b2-9f7fb0e54b6f\") " Jan 21 11:39:39 crc kubenswrapper[4824]: I0121 11:39:39.662205 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/18e65eb9-3b96-48cd-88b2-9f7fb0e54b6f-ssh-key-openstack-edpm-ipam\") pod \"18e65eb9-3b96-48cd-88b2-9f7fb0e54b6f\" (UID: \"18e65eb9-3b96-48cd-88b2-9f7fb0e54b6f\") " Jan 21 11:39:39 crc kubenswrapper[4824]: I0121 11:39:39.666564 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/18e65eb9-3b96-48cd-88b2-9f7fb0e54b6f-kube-api-access-md7w4" (OuterVolumeSpecName: "kube-api-access-md7w4") pod "18e65eb9-3b96-48cd-88b2-9f7fb0e54b6f" (UID: "18e65eb9-3b96-48cd-88b2-9f7fb0e54b6f"). InnerVolumeSpecName "kube-api-access-md7w4". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:39:39 crc kubenswrapper[4824]: I0121 11:39:39.666600 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/18e65eb9-3b96-48cd-88b2-9f7fb0e54b6f-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "18e65eb9-3b96-48cd-88b2-9f7fb0e54b6f" (UID: "18e65eb9-3b96-48cd-88b2-9f7fb0e54b6f"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:39:39 crc kubenswrapper[4824]: I0121 11:39:39.681257 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/18e65eb9-3b96-48cd-88b2-9f7fb0e54b6f-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "18e65eb9-3b96-48cd-88b2-9f7fb0e54b6f" (UID: "18e65eb9-3b96-48cd-88b2-9f7fb0e54b6f"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:39:39 crc kubenswrapper[4824]: I0121 11:39:39.682373 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/18e65eb9-3b96-48cd-88b2-9f7fb0e54b6f-inventory" (OuterVolumeSpecName: "inventory") pod "18e65eb9-3b96-48cd-88b2-9f7fb0e54b6f" (UID: "18e65eb9-3b96-48cd-88b2-9f7fb0e54b6f"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:39:39 crc kubenswrapper[4824]: I0121 11:39:39.682932 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/18e65eb9-3b96-48cd-88b2-9f7fb0e54b6f-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "18e65eb9-3b96-48cd-88b2-9f7fb0e54b6f" (UID: "18e65eb9-3b96-48cd-88b2-9f7fb0e54b6f"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:39:39 crc kubenswrapper[4824]: I0121 11:39:39.764933 4824 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/18e65eb9-3b96-48cd-88b2-9f7fb0e54b6f-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Jan 21 11:39:39 crc kubenswrapper[4824]: I0121 11:39:39.764977 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-md7w4\" (UniqueName: \"kubernetes.io/projected/18e65eb9-3b96-48cd-88b2-9f7fb0e54b6f-kube-api-access-md7w4\") on node \"crc\" DevicePath \"\"" Jan 21 11:39:39 crc kubenswrapper[4824]: I0121 11:39:39.764988 4824 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/18e65eb9-3b96-48cd-88b2-9f7fb0e54b6f-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Jan 21 11:39:39 crc kubenswrapper[4824]: I0121 11:39:39.764998 4824 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/18e65eb9-3b96-48cd-88b2-9f7fb0e54b6f-inventory\") on node \"crc\" DevicePath \"\"" Jan 21 11:39:39 crc kubenswrapper[4824]: I0121 11:39:39.765006 4824 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18e65eb9-3b96-48cd-88b2-9f7fb0e54b6f-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 21 11:39:40 crc kubenswrapper[4824]: I0121 11:39:40.259762 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gf4zq" event={"ID":"18e65eb9-3b96-48cd-88b2-9f7fb0e54b6f","Type":"ContainerDied","Data":"6d455d5215fcbfd8a47efb176dd7eb21ea2bd57fe6e56abd130fecb774e16899"} Jan 21 11:39:40 crc kubenswrapper[4824]: I0121 11:39:40.259801 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6d455d5215fcbfd8a47efb176dd7eb21ea2bd57fe6e56abd130fecb774e16899" Jan 21 11:39:40 crc kubenswrapper[4824]: I0121 11:39:40.259845 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gf4zq" Jan 21 11:39:40 crc kubenswrapper[4824]: I0121 11:39:40.323204 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-rf47l"] Jan 21 11:39:40 crc kubenswrapper[4824]: E0121 11:39:40.323595 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="18e65eb9-3b96-48cd-88b2-9f7fb0e54b6f" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Jan 21 11:39:40 crc kubenswrapper[4824]: I0121 11:39:40.323614 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="18e65eb9-3b96-48cd-88b2-9f7fb0e54b6f" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Jan 21 11:39:40 crc kubenswrapper[4824]: I0121 11:39:40.323785 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="18e65eb9-3b96-48cd-88b2-9f7fb0e54b6f" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Jan 21 11:39:40 crc kubenswrapper[4824]: I0121 11:39:40.324353 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-rf47l" Jan 21 11:39:40 crc kubenswrapper[4824]: I0121 11:39:40.326357 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-tcqk6" Jan 21 11:39:40 crc kubenswrapper[4824]: I0121 11:39:40.326647 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-ovn-metadata-agent-neutron-config" Jan 21 11:39:40 crc kubenswrapper[4824]: I0121 11:39:40.326691 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Jan 21 11:39:40 crc kubenswrapper[4824]: I0121 11:39:40.326656 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Jan 21 11:39:40 crc kubenswrapper[4824]: I0121 11:39:40.327301 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-neutron-config" Jan 21 11:39:40 crc kubenswrapper[4824]: I0121 11:39:40.329664 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Jan 21 11:39:40 crc kubenswrapper[4824]: I0121 11:39:40.331307 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-rf47l"] Jan 21 11:39:40 crc kubenswrapper[4824]: I0121 11:39:40.375231 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/4f3c3a58-5d6b-4cc2-9707-d088b2fd463d-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-rf47l\" (UID: \"4f3c3a58-5d6b-4cc2-9707-d088b2fd463d\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-rf47l" Jan 21 11:39:40 crc kubenswrapper[4824]: I0121 11:39:40.375322 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/4f3c3a58-5d6b-4cc2-9707-d088b2fd463d-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-rf47l\" (UID: \"4f3c3a58-5d6b-4cc2-9707-d088b2fd463d\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-rf47l" Jan 21 11:39:40 crc kubenswrapper[4824]: I0121 11:39:40.375447 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f3c3a58-5d6b-4cc2-9707-d088b2fd463d-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-rf47l\" (UID: \"4f3c3a58-5d6b-4cc2-9707-d088b2fd463d\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-rf47l" Jan 21 11:39:40 crc kubenswrapper[4824]: I0121 11:39:40.375556 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zw25j\" (UniqueName: \"kubernetes.io/projected/4f3c3a58-5d6b-4cc2-9707-d088b2fd463d-kube-api-access-zw25j\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-rf47l\" (UID: \"4f3c3a58-5d6b-4cc2-9707-d088b2fd463d\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-rf47l" Jan 21 11:39:40 crc kubenswrapper[4824]: I0121 11:39:40.375631 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/4f3c3a58-5d6b-4cc2-9707-d088b2fd463d-ssh-key-openstack-edpm-ipam\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-rf47l\" (UID: \"4f3c3a58-5d6b-4cc2-9707-d088b2fd463d\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-rf47l" Jan 21 11:39:40 crc kubenswrapper[4824]: I0121 11:39:40.375742 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4f3c3a58-5d6b-4cc2-9707-d088b2fd463d-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-rf47l\" (UID: \"4f3c3a58-5d6b-4cc2-9707-d088b2fd463d\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-rf47l" Jan 21 11:39:40 crc kubenswrapper[4824]: I0121 11:39:40.476765 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zw25j\" (UniqueName: \"kubernetes.io/projected/4f3c3a58-5d6b-4cc2-9707-d088b2fd463d-kube-api-access-zw25j\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-rf47l\" (UID: \"4f3c3a58-5d6b-4cc2-9707-d088b2fd463d\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-rf47l" Jan 21 11:39:40 crc kubenswrapper[4824]: I0121 11:39:40.476835 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/4f3c3a58-5d6b-4cc2-9707-d088b2fd463d-ssh-key-openstack-edpm-ipam\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-rf47l\" (UID: \"4f3c3a58-5d6b-4cc2-9707-d088b2fd463d\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-rf47l" Jan 21 11:39:40 crc kubenswrapper[4824]: I0121 11:39:40.476932 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4f3c3a58-5d6b-4cc2-9707-d088b2fd463d-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-rf47l\" (UID: \"4f3c3a58-5d6b-4cc2-9707-d088b2fd463d\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-rf47l" Jan 21 11:39:40 crc kubenswrapper[4824]: I0121 11:39:40.477599 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/4f3c3a58-5d6b-4cc2-9707-d088b2fd463d-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-rf47l\" (UID: \"4f3c3a58-5d6b-4cc2-9707-d088b2fd463d\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-rf47l" Jan 21 11:39:40 crc kubenswrapper[4824]: I0121 11:39:40.477707 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/4f3c3a58-5d6b-4cc2-9707-d088b2fd463d-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-rf47l\" (UID: \"4f3c3a58-5d6b-4cc2-9707-d088b2fd463d\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-rf47l" Jan 21 11:39:40 crc kubenswrapper[4824]: I0121 11:39:40.477819 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f3c3a58-5d6b-4cc2-9707-d088b2fd463d-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-rf47l\" (UID: \"4f3c3a58-5d6b-4cc2-9707-d088b2fd463d\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-rf47l" Jan 21 11:39:40 crc kubenswrapper[4824]: I0121 11:39:40.481030 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/4f3c3a58-5d6b-4cc2-9707-d088b2fd463d-ssh-key-openstack-edpm-ipam\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-rf47l\" (UID: \"4f3c3a58-5d6b-4cc2-9707-d088b2fd463d\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-rf47l" Jan 21 11:39:40 crc kubenswrapper[4824]: I0121 11:39:40.481263 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f3c3a58-5d6b-4cc2-9707-d088b2fd463d-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-rf47l\" (UID: \"4f3c3a58-5d6b-4cc2-9707-d088b2fd463d\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-rf47l" Jan 21 11:39:40 crc kubenswrapper[4824]: I0121 11:39:40.482355 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/4f3c3a58-5d6b-4cc2-9707-d088b2fd463d-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-rf47l\" (UID: \"4f3c3a58-5d6b-4cc2-9707-d088b2fd463d\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-rf47l" Jan 21 11:39:40 crc kubenswrapper[4824]: I0121 11:39:40.484082 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/4f3c3a58-5d6b-4cc2-9707-d088b2fd463d-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-rf47l\" (UID: \"4f3c3a58-5d6b-4cc2-9707-d088b2fd463d\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-rf47l" Jan 21 11:39:40 crc kubenswrapper[4824]: I0121 11:39:40.484156 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4f3c3a58-5d6b-4cc2-9707-d088b2fd463d-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-rf47l\" (UID: \"4f3c3a58-5d6b-4cc2-9707-d088b2fd463d\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-rf47l" Jan 21 11:39:40 crc kubenswrapper[4824]: I0121 11:39:40.491536 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zw25j\" (UniqueName: \"kubernetes.io/projected/4f3c3a58-5d6b-4cc2-9707-d088b2fd463d-kube-api-access-zw25j\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-rf47l\" (UID: \"4f3c3a58-5d6b-4cc2-9707-d088b2fd463d\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-rf47l" Jan 21 11:39:40 crc kubenswrapper[4824]: I0121 11:39:40.637313 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-rf47l" Jan 21 11:39:41 crc kubenswrapper[4824]: I0121 11:39:41.048912 4824 scope.go:117] "RemoveContainer" containerID="1cd835fe36ebb6541740e5ab5fd843470d46dfb78975fd968e5cc9aaef462499" Jan 21 11:39:41 crc kubenswrapper[4824]: E0121 11:39:41.049527 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 11:39:41 crc kubenswrapper[4824]: I0121 11:39:41.082811 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-rf47l"] Jan 21 11:39:41 crc kubenswrapper[4824]: I0121 11:39:41.267198 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-rf47l" event={"ID":"4f3c3a58-5d6b-4cc2-9707-d088b2fd463d","Type":"ContainerStarted","Data":"ca1530bd513a46b192d4e6e7ec5ad68d0b40713a10661b674e5682caf2459e0c"} Jan 21 11:39:42 crc kubenswrapper[4824]: I0121 11:39:42.275126 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-rf47l" event={"ID":"4f3c3a58-5d6b-4cc2-9707-d088b2fd463d","Type":"ContainerStarted","Data":"6d5206d4d635a0ba105e072dcc9351ce51a4045b38ec07cb4ebf2776319165b3"} Jan 21 11:39:42 crc kubenswrapper[4824]: I0121 11:39:42.289212 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-rf47l" podStartSLOduration=1.729008087 podStartE2EDuration="2.289198968s" podCreationTimestamp="2026-01-21 11:39:40 +0000 UTC" firstStartedPulling="2026-01-21 11:39:41.086063237 +0000 UTC m=+1723.379092529" lastFinishedPulling="2026-01-21 11:39:41.646254118 +0000 UTC m=+1723.939283410" observedRunningTime="2026-01-21 11:39:42.285723546 +0000 UTC m=+1724.578752837" watchObservedRunningTime="2026-01-21 11:39:42.289198968 +0000 UTC m=+1724.582228260" Jan 21 11:39:55 crc kubenswrapper[4824]: I0121 11:39:55.049939 4824 scope.go:117] "RemoveContainer" containerID="1cd835fe36ebb6541740e5ab5fd843470d46dfb78975fd968e5cc9aaef462499" Jan 21 11:39:55 crc kubenswrapper[4824]: E0121 11:39:55.050658 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 11:40:09 crc kubenswrapper[4824]: I0121 11:40:09.049137 4824 scope.go:117] "RemoveContainer" containerID="1cd835fe36ebb6541740e5ab5fd843470d46dfb78975fd968e5cc9aaef462499" Jan 21 11:40:09 crc kubenswrapper[4824]: E0121 11:40:09.050084 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 11:40:16 crc kubenswrapper[4824]: I0121 11:40:16.481908 4824 generic.go:334] "Generic (PLEG): container finished" podID="4f3c3a58-5d6b-4cc2-9707-d088b2fd463d" containerID="6d5206d4d635a0ba105e072dcc9351ce51a4045b38ec07cb4ebf2776319165b3" exitCode=0 Jan 21 11:40:16 crc kubenswrapper[4824]: I0121 11:40:16.481988 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-rf47l" event={"ID":"4f3c3a58-5d6b-4cc2-9707-d088b2fd463d","Type":"ContainerDied","Data":"6d5206d4d635a0ba105e072dcc9351ce51a4045b38ec07cb4ebf2776319165b3"} Jan 21 11:40:17 crc kubenswrapper[4824]: I0121 11:40:17.789128 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-rf47l" Jan 21 11:40:17 crc kubenswrapper[4824]: I0121 11:40:17.910150 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/4f3c3a58-5d6b-4cc2-9707-d088b2fd463d-ssh-key-openstack-edpm-ipam\") pod \"4f3c3a58-5d6b-4cc2-9707-d088b2fd463d\" (UID: \"4f3c3a58-5d6b-4cc2-9707-d088b2fd463d\") " Jan 21 11:40:17 crc kubenswrapper[4824]: I0121 11:40:17.910212 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/4f3c3a58-5d6b-4cc2-9707-d088b2fd463d-nova-metadata-neutron-config-0\") pod \"4f3c3a58-5d6b-4cc2-9707-d088b2fd463d\" (UID: \"4f3c3a58-5d6b-4cc2-9707-d088b2fd463d\") " Jan 21 11:40:17 crc kubenswrapper[4824]: I0121 11:40:17.910284 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4f3c3a58-5d6b-4cc2-9707-d088b2fd463d-inventory\") pod \"4f3c3a58-5d6b-4cc2-9707-d088b2fd463d\" (UID: \"4f3c3a58-5d6b-4cc2-9707-d088b2fd463d\") " Jan 21 11:40:17 crc kubenswrapper[4824]: I0121 11:40:17.910318 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f3c3a58-5d6b-4cc2-9707-d088b2fd463d-neutron-metadata-combined-ca-bundle\") pod \"4f3c3a58-5d6b-4cc2-9707-d088b2fd463d\" (UID: \"4f3c3a58-5d6b-4cc2-9707-d088b2fd463d\") " Jan 21 11:40:17 crc kubenswrapper[4824]: I0121 11:40:17.910396 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zw25j\" (UniqueName: \"kubernetes.io/projected/4f3c3a58-5d6b-4cc2-9707-d088b2fd463d-kube-api-access-zw25j\") pod \"4f3c3a58-5d6b-4cc2-9707-d088b2fd463d\" (UID: \"4f3c3a58-5d6b-4cc2-9707-d088b2fd463d\") " Jan 21 11:40:17 crc kubenswrapper[4824]: I0121 11:40:17.910442 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/4f3c3a58-5d6b-4cc2-9707-d088b2fd463d-neutron-ovn-metadata-agent-neutron-config-0\") pod \"4f3c3a58-5d6b-4cc2-9707-d088b2fd463d\" (UID: \"4f3c3a58-5d6b-4cc2-9707-d088b2fd463d\") " Jan 21 11:40:17 crc kubenswrapper[4824]: I0121 11:40:17.914619 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4f3c3a58-5d6b-4cc2-9707-d088b2fd463d-kube-api-access-zw25j" (OuterVolumeSpecName: "kube-api-access-zw25j") pod "4f3c3a58-5d6b-4cc2-9707-d088b2fd463d" (UID: "4f3c3a58-5d6b-4cc2-9707-d088b2fd463d"). InnerVolumeSpecName "kube-api-access-zw25j". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:40:17 crc kubenswrapper[4824]: I0121 11:40:17.918916 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4f3c3a58-5d6b-4cc2-9707-d088b2fd463d-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "4f3c3a58-5d6b-4cc2-9707-d088b2fd463d" (UID: "4f3c3a58-5d6b-4cc2-9707-d088b2fd463d"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:40:17 crc kubenswrapper[4824]: I0121 11:40:17.930810 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4f3c3a58-5d6b-4cc2-9707-d088b2fd463d-neutron-ovn-metadata-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-ovn-metadata-agent-neutron-config-0") pod "4f3c3a58-5d6b-4cc2-9707-d088b2fd463d" (UID: "4f3c3a58-5d6b-4cc2-9707-d088b2fd463d"). InnerVolumeSpecName "neutron-ovn-metadata-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:40:17 crc kubenswrapper[4824]: I0121 11:40:17.931034 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4f3c3a58-5d6b-4cc2-9707-d088b2fd463d-nova-metadata-neutron-config-0" (OuterVolumeSpecName: "nova-metadata-neutron-config-0") pod "4f3c3a58-5d6b-4cc2-9707-d088b2fd463d" (UID: "4f3c3a58-5d6b-4cc2-9707-d088b2fd463d"). InnerVolumeSpecName "nova-metadata-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:40:17 crc kubenswrapper[4824]: I0121 11:40:17.932111 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4f3c3a58-5d6b-4cc2-9707-d088b2fd463d-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "4f3c3a58-5d6b-4cc2-9707-d088b2fd463d" (UID: "4f3c3a58-5d6b-4cc2-9707-d088b2fd463d"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:40:17 crc kubenswrapper[4824]: I0121 11:40:17.932685 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4f3c3a58-5d6b-4cc2-9707-d088b2fd463d-inventory" (OuterVolumeSpecName: "inventory") pod "4f3c3a58-5d6b-4cc2-9707-d088b2fd463d" (UID: "4f3c3a58-5d6b-4cc2-9707-d088b2fd463d"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:40:18 crc kubenswrapper[4824]: I0121 11:40:18.013136 4824 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/4f3c3a58-5d6b-4cc2-9707-d088b2fd463d-nova-metadata-neutron-config-0\") on node \"crc\" DevicePath \"\"" Jan 21 11:40:18 crc kubenswrapper[4824]: I0121 11:40:18.013165 4824 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4f3c3a58-5d6b-4cc2-9707-d088b2fd463d-inventory\") on node \"crc\" DevicePath \"\"" Jan 21 11:40:18 crc kubenswrapper[4824]: I0121 11:40:18.013175 4824 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f3c3a58-5d6b-4cc2-9707-d088b2fd463d-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 21 11:40:18 crc kubenswrapper[4824]: I0121 11:40:18.013185 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zw25j\" (UniqueName: \"kubernetes.io/projected/4f3c3a58-5d6b-4cc2-9707-d088b2fd463d-kube-api-access-zw25j\") on node \"crc\" DevicePath \"\"" Jan 21 11:40:18 crc kubenswrapper[4824]: I0121 11:40:18.013194 4824 reconciler_common.go:293] "Volume detached for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/4f3c3a58-5d6b-4cc2-9707-d088b2fd463d-neutron-ovn-metadata-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Jan 21 11:40:18 crc kubenswrapper[4824]: I0121 11:40:18.013205 4824 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/4f3c3a58-5d6b-4cc2-9707-d088b2fd463d-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Jan 21 11:40:18 crc kubenswrapper[4824]: I0121 11:40:18.496409 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-rf47l" event={"ID":"4f3c3a58-5d6b-4cc2-9707-d088b2fd463d","Type":"ContainerDied","Data":"ca1530bd513a46b192d4e6e7ec5ad68d0b40713a10661b674e5682caf2459e0c"} Jan 21 11:40:18 crc kubenswrapper[4824]: I0121 11:40:18.496444 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ca1530bd513a46b192d4e6e7ec5ad68d0b40713a10661b674e5682caf2459e0c" Jan 21 11:40:18 crc kubenswrapper[4824]: I0121 11:40:18.496663 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-rf47l" Jan 21 11:40:18 crc kubenswrapper[4824]: I0121 11:40:18.562483 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-pgjpn"] Jan 21 11:40:18 crc kubenswrapper[4824]: E0121 11:40:18.562819 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4f3c3a58-5d6b-4cc2-9707-d088b2fd463d" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Jan 21 11:40:18 crc kubenswrapper[4824]: I0121 11:40:18.562836 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="4f3c3a58-5d6b-4cc2-9707-d088b2fd463d" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Jan 21 11:40:18 crc kubenswrapper[4824]: I0121 11:40:18.563035 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="4f3c3a58-5d6b-4cc2-9707-d088b2fd463d" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Jan 21 11:40:18 crc kubenswrapper[4824]: I0121 11:40:18.563529 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-pgjpn" Jan 21 11:40:18 crc kubenswrapper[4824]: I0121 11:40:18.565289 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Jan 21 11:40:18 crc kubenswrapper[4824]: I0121 11:40:18.565307 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-tcqk6" Jan 21 11:40:18 crc kubenswrapper[4824]: I0121 11:40:18.566526 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"libvirt-secret" Jan 21 11:40:18 crc kubenswrapper[4824]: I0121 11:40:18.567144 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Jan 21 11:40:18 crc kubenswrapper[4824]: I0121 11:40:18.567689 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Jan 21 11:40:18 crc kubenswrapper[4824]: I0121 11:40:18.577432 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-pgjpn"] Jan 21 11:40:18 crc kubenswrapper[4824]: I0121 11:40:18.725795 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qnnln\" (UniqueName: \"kubernetes.io/projected/fdde8991-c204-45a0-b344-dcdc41a9a275-kube-api-access-qnnln\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-pgjpn\" (UID: \"fdde8991-c204-45a0-b344-dcdc41a9a275\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-pgjpn" Jan 21 11:40:18 crc kubenswrapper[4824]: I0121 11:40:18.725848 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fdde8991-c204-45a0-b344-dcdc41a9a275-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-pgjpn\" (UID: \"fdde8991-c204-45a0-b344-dcdc41a9a275\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-pgjpn" Jan 21 11:40:18 crc kubenswrapper[4824]: I0121 11:40:18.726025 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fdde8991-c204-45a0-b344-dcdc41a9a275-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-pgjpn\" (UID: \"fdde8991-c204-45a0-b344-dcdc41a9a275\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-pgjpn" Jan 21 11:40:18 crc kubenswrapper[4824]: I0121 11:40:18.726371 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/fdde8991-c204-45a0-b344-dcdc41a9a275-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-pgjpn\" (UID: \"fdde8991-c204-45a0-b344-dcdc41a9a275\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-pgjpn" Jan 21 11:40:18 crc kubenswrapper[4824]: I0121 11:40:18.726540 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/fdde8991-c204-45a0-b344-dcdc41a9a275-ssh-key-openstack-edpm-ipam\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-pgjpn\" (UID: \"fdde8991-c204-45a0-b344-dcdc41a9a275\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-pgjpn" Jan 21 11:40:18 crc kubenswrapper[4824]: I0121 11:40:18.828434 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/fdde8991-c204-45a0-b344-dcdc41a9a275-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-pgjpn\" (UID: \"fdde8991-c204-45a0-b344-dcdc41a9a275\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-pgjpn" Jan 21 11:40:18 crc kubenswrapper[4824]: I0121 11:40:18.828551 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/fdde8991-c204-45a0-b344-dcdc41a9a275-ssh-key-openstack-edpm-ipam\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-pgjpn\" (UID: \"fdde8991-c204-45a0-b344-dcdc41a9a275\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-pgjpn" Jan 21 11:40:18 crc kubenswrapper[4824]: I0121 11:40:18.828691 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qnnln\" (UniqueName: \"kubernetes.io/projected/fdde8991-c204-45a0-b344-dcdc41a9a275-kube-api-access-qnnln\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-pgjpn\" (UID: \"fdde8991-c204-45a0-b344-dcdc41a9a275\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-pgjpn" Jan 21 11:40:18 crc kubenswrapper[4824]: I0121 11:40:18.828721 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fdde8991-c204-45a0-b344-dcdc41a9a275-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-pgjpn\" (UID: \"fdde8991-c204-45a0-b344-dcdc41a9a275\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-pgjpn" Jan 21 11:40:18 crc kubenswrapper[4824]: I0121 11:40:18.828789 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fdde8991-c204-45a0-b344-dcdc41a9a275-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-pgjpn\" (UID: \"fdde8991-c204-45a0-b344-dcdc41a9a275\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-pgjpn" Jan 21 11:40:18 crc kubenswrapper[4824]: I0121 11:40:18.831580 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/fdde8991-c204-45a0-b344-dcdc41a9a275-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-pgjpn\" (UID: \"fdde8991-c204-45a0-b344-dcdc41a9a275\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-pgjpn" Jan 21 11:40:18 crc kubenswrapper[4824]: I0121 11:40:18.831907 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fdde8991-c204-45a0-b344-dcdc41a9a275-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-pgjpn\" (UID: \"fdde8991-c204-45a0-b344-dcdc41a9a275\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-pgjpn" Jan 21 11:40:18 crc kubenswrapper[4824]: I0121 11:40:18.832485 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fdde8991-c204-45a0-b344-dcdc41a9a275-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-pgjpn\" (UID: \"fdde8991-c204-45a0-b344-dcdc41a9a275\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-pgjpn" Jan 21 11:40:18 crc kubenswrapper[4824]: I0121 11:40:18.833129 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/fdde8991-c204-45a0-b344-dcdc41a9a275-ssh-key-openstack-edpm-ipam\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-pgjpn\" (UID: \"fdde8991-c204-45a0-b344-dcdc41a9a275\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-pgjpn" Jan 21 11:40:18 crc kubenswrapper[4824]: I0121 11:40:18.841643 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qnnln\" (UniqueName: \"kubernetes.io/projected/fdde8991-c204-45a0-b344-dcdc41a9a275-kube-api-access-qnnln\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-pgjpn\" (UID: \"fdde8991-c204-45a0-b344-dcdc41a9a275\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-pgjpn" Jan 21 11:40:18 crc kubenswrapper[4824]: I0121 11:40:18.877485 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-pgjpn" Jan 21 11:40:19 crc kubenswrapper[4824]: I0121 11:40:19.305813 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-pgjpn"] Jan 21 11:40:19 crc kubenswrapper[4824]: W0121 11:40:19.306374 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfdde8991_c204_45a0_b344_dcdc41a9a275.slice/crio-bf1417745a4dbeb59dc25ccbfc9986f4a9ff1ab7739e8b004c84cda87ad44202 WatchSource:0}: Error finding container bf1417745a4dbeb59dc25ccbfc9986f4a9ff1ab7739e8b004c84cda87ad44202: Status 404 returned error can't find the container with id bf1417745a4dbeb59dc25ccbfc9986f4a9ff1ab7739e8b004c84cda87ad44202 Jan 21 11:40:19 crc kubenswrapper[4824]: I0121 11:40:19.503811 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-pgjpn" event={"ID":"fdde8991-c204-45a0-b344-dcdc41a9a275","Type":"ContainerStarted","Data":"bf1417745a4dbeb59dc25ccbfc9986f4a9ff1ab7739e8b004c84cda87ad44202"} Jan 21 11:40:20 crc kubenswrapper[4824]: I0121 11:40:20.511203 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-pgjpn" event={"ID":"fdde8991-c204-45a0-b344-dcdc41a9a275","Type":"ContainerStarted","Data":"1f9e141178592276bca51a00e34332a99986e41de0caf5feaec25186ff7878f9"} Jan 21 11:40:20 crc kubenswrapper[4824]: I0121 11:40:20.522800 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-pgjpn" podStartSLOduration=2.044140647 podStartE2EDuration="2.522788581s" podCreationTimestamp="2026-01-21 11:40:18 +0000 UTC" firstStartedPulling="2026-01-21 11:40:19.308019365 +0000 UTC m=+1761.601048657" lastFinishedPulling="2026-01-21 11:40:19.7866673 +0000 UTC m=+1762.079696591" observedRunningTime="2026-01-21 11:40:20.52187017 +0000 UTC m=+1762.814899462" watchObservedRunningTime="2026-01-21 11:40:20.522788581 +0000 UTC m=+1762.815817874" Jan 21 11:40:21 crc kubenswrapper[4824]: I0121 11:40:21.049780 4824 scope.go:117] "RemoveContainer" containerID="1cd835fe36ebb6541740e5ab5fd843470d46dfb78975fd968e5cc9aaef462499" Jan 21 11:40:21 crc kubenswrapper[4824]: E0121 11:40:21.050223 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 11:40:34 crc kubenswrapper[4824]: I0121 11:40:34.050587 4824 scope.go:117] "RemoveContainer" containerID="1cd835fe36ebb6541740e5ab5fd843470d46dfb78975fd968e5cc9aaef462499" Jan 21 11:40:34 crc kubenswrapper[4824]: E0121 11:40:34.051348 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 11:40:47 crc kubenswrapper[4824]: I0121 11:40:47.048827 4824 scope.go:117] "RemoveContainer" containerID="1cd835fe36ebb6541740e5ab5fd843470d46dfb78975fd968e5cc9aaef462499" Jan 21 11:40:47 crc kubenswrapper[4824]: E0121 11:40:47.049393 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 11:41:00 crc kubenswrapper[4824]: I0121 11:41:00.049806 4824 scope.go:117] "RemoveContainer" containerID="1cd835fe36ebb6541740e5ab5fd843470d46dfb78975fd968e5cc9aaef462499" Jan 21 11:41:00 crc kubenswrapper[4824]: E0121 11:41:00.050423 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 11:41:13 crc kubenswrapper[4824]: I0121 11:41:13.049334 4824 scope.go:117] "RemoveContainer" containerID="1cd835fe36ebb6541740e5ab5fd843470d46dfb78975fd968e5cc9aaef462499" Jan 21 11:41:13 crc kubenswrapper[4824]: E0121 11:41:13.050110 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 11:41:26 crc kubenswrapper[4824]: I0121 11:41:26.049872 4824 scope.go:117] "RemoveContainer" containerID="1cd835fe36ebb6541740e5ab5fd843470d46dfb78975fd968e5cc9aaef462499" Jan 21 11:41:26 crc kubenswrapper[4824]: I0121 11:41:26.928833 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" event={"ID":"33f3d922-4ffe-409b-a49a-d88c85898260","Type":"ContainerStarted","Data":"1283af2f214718d7c6e80663aef792afcaa88b12e9e64dbc3ab357a33b0df07b"} Jan 21 11:43:10 crc kubenswrapper[4824]: I0121 11:43:10.589800 4824 generic.go:334] "Generic (PLEG): container finished" podID="fdde8991-c204-45a0-b344-dcdc41a9a275" containerID="1f9e141178592276bca51a00e34332a99986e41de0caf5feaec25186ff7878f9" exitCode=0 Jan 21 11:43:10 crc kubenswrapper[4824]: I0121 11:43:10.590529 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-pgjpn" event={"ID":"fdde8991-c204-45a0-b344-dcdc41a9a275","Type":"ContainerDied","Data":"1f9e141178592276bca51a00e34332a99986e41de0caf5feaec25186ff7878f9"} Jan 21 11:43:11 crc kubenswrapper[4824]: I0121 11:43:11.930472 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-pgjpn" Jan 21 11:43:11 crc kubenswrapper[4824]: I0121 11:43:11.969389 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fdde8991-c204-45a0-b344-dcdc41a9a275-inventory\") pod \"fdde8991-c204-45a0-b344-dcdc41a9a275\" (UID: \"fdde8991-c204-45a0-b344-dcdc41a9a275\") " Jan 21 11:43:11 crc kubenswrapper[4824]: I0121 11:43:11.969568 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/fdde8991-c204-45a0-b344-dcdc41a9a275-ssh-key-openstack-edpm-ipam\") pod \"fdde8991-c204-45a0-b344-dcdc41a9a275\" (UID: \"fdde8991-c204-45a0-b344-dcdc41a9a275\") " Jan 21 11:43:11 crc kubenswrapper[4824]: I0121 11:43:11.969654 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qnnln\" (UniqueName: \"kubernetes.io/projected/fdde8991-c204-45a0-b344-dcdc41a9a275-kube-api-access-qnnln\") pod \"fdde8991-c204-45a0-b344-dcdc41a9a275\" (UID: \"fdde8991-c204-45a0-b344-dcdc41a9a275\") " Jan 21 11:43:11 crc kubenswrapper[4824]: I0121 11:43:11.969737 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/fdde8991-c204-45a0-b344-dcdc41a9a275-libvirt-secret-0\") pod \"fdde8991-c204-45a0-b344-dcdc41a9a275\" (UID: \"fdde8991-c204-45a0-b344-dcdc41a9a275\") " Jan 21 11:43:11 crc kubenswrapper[4824]: I0121 11:43:11.969802 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fdde8991-c204-45a0-b344-dcdc41a9a275-libvirt-combined-ca-bundle\") pod \"fdde8991-c204-45a0-b344-dcdc41a9a275\" (UID: \"fdde8991-c204-45a0-b344-dcdc41a9a275\") " Jan 21 11:43:11 crc kubenswrapper[4824]: I0121 11:43:11.976119 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fdde8991-c204-45a0-b344-dcdc41a9a275-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "fdde8991-c204-45a0-b344-dcdc41a9a275" (UID: "fdde8991-c204-45a0-b344-dcdc41a9a275"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:43:11 crc kubenswrapper[4824]: I0121 11:43:11.979510 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fdde8991-c204-45a0-b344-dcdc41a9a275-kube-api-access-qnnln" (OuterVolumeSpecName: "kube-api-access-qnnln") pod "fdde8991-c204-45a0-b344-dcdc41a9a275" (UID: "fdde8991-c204-45a0-b344-dcdc41a9a275"). InnerVolumeSpecName "kube-api-access-qnnln". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:43:11 crc kubenswrapper[4824]: I0121 11:43:11.994168 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fdde8991-c204-45a0-b344-dcdc41a9a275-inventory" (OuterVolumeSpecName: "inventory") pod "fdde8991-c204-45a0-b344-dcdc41a9a275" (UID: "fdde8991-c204-45a0-b344-dcdc41a9a275"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:43:11 crc kubenswrapper[4824]: I0121 11:43:11.994217 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fdde8991-c204-45a0-b344-dcdc41a9a275-libvirt-secret-0" (OuterVolumeSpecName: "libvirt-secret-0") pod "fdde8991-c204-45a0-b344-dcdc41a9a275" (UID: "fdde8991-c204-45a0-b344-dcdc41a9a275"). InnerVolumeSpecName "libvirt-secret-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:43:12 crc kubenswrapper[4824]: I0121 11:43:12.000377 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fdde8991-c204-45a0-b344-dcdc41a9a275-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "fdde8991-c204-45a0-b344-dcdc41a9a275" (UID: "fdde8991-c204-45a0-b344-dcdc41a9a275"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:43:12 crc kubenswrapper[4824]: I0121 11:43:12.072977 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qnnln\" (UniqueName: \"kubernetes.io/projected/fdde8991-c204-45a0-b344-dcdc41a9a275-kube-api-access-qnnln\") on node \"crc\" DevicePath \"\"" Jan 21 11:43:12 crc kubenswrapper[4824]: I0121 11:43:12.073010 4824 reconciler_common.go:293] "Volume detached for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/fdde8991-c204-45a0-b344-dcdc41a9a275-libvirt-secret-0\") on node \"crc\" DevicePath \"\"" Jan 21 11:43:12 crc kubenswrapper[4824]: I0121 11:43:12.073019 4824 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fdde8991-c204-45a0-b344-dcdc41a9a275-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 21 11:43:12 crc kubenswrapper[4824]: I0121 11:43:12.073029 4824 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fdde8991-c204-45a0-b344-dcdc41a9a275-inventory\") on node \"crc\" DevicePath \"\"" Jan 21 11:43:12 crc kubenswrapper[4824]: I0121 11:43:12.073040 4824 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/fdde8991-c204-45a0-b344-dcdc41a9a275-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Jan 21 11:43:12 crc kubenswrapper[4824]: I0121 11:43:12.605701 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-pgjpn" event={"ID":"fdde8991-c204-45a0-b344-dcdc41a9a275","Type":"ContainerDied","Data":"bf1417745a4dbeb59dc25ccbfc9986f4a9ff1ab7739e8b004c84cda87ad44202"} Jan 21 11:43:12 crc kubenswrapper[4824]: I0121 11:43:12.605751 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-pgjpn" Jan 21 11:43:12 crc kubenswrapper[4824]: I0121 11:43:12.605757 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bf1417745a4dbeb59dc25ccbfc9986f4a9ff1ab7739e8b004c84cda87ad44202" Jan 21 11:43:12 crc kubenswrapper[4824]: I0121 11:43:12.674199 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-7crw8"] Jan 21 11:43:12 crc kubenswrapper[4824]: E0121 11:43:12.674596 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fdde8991-c204-45a0-b344-dcdc41a9a275" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Jan 21 11:43:12 crc kubenswrapper[4824]: I0121 11:43:12.674619 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="fdde8991-c204-45a0-b344-dcdc41a9a275" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Jan 21 11:43:12 crc kubenswrapper[4824]: I0121 11:43:12.674816 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="fdde8991-c204-45a0-b344-dcdc41a9a275" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Jan 21 11:43:12 crc kubenswrapper[4824]: I0121 11:43:12.675406 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7crw8" Jan 21 11:43:12 crc kubenswrapper[4824]: I0121 11:43:12.678465 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Jan 21 11:43:12 crc kubenswrapper[4824]: I0121 11:43:12.678767 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-extra-config" Jan 21 11:43:12 crc kubenswrapper[4824]: I0121 11:43:12.678982 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Jan 21 11:43:12 crc kubenswrapper[4824]: I0121 11:43:12.679047 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Jan 21 11:43:12 crc kubenswrapper[4824]: I0121 11:43:12.679174 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-tcqk6" Jan 21 11:43:12 crc kubenswrapper[4824]: I0121 11:43:12.679311 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Jan 21 11:43:12 crc kubenswrapper[4824]: I0121 11:43:12.679435 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Jan 21 11:43:12 crc kubenswrapper[4824]: I0121 11:43:12.683224 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/7355a8fc-8c18-4de1-8baf-f6294ea9e11c-ssh-key-openstack-edpm-ipam\") pod \"nova-edpm-deployment-openstack-edpm-ipam-7crw8\" (UID: \"7355a8fc-8c18-4de1-8baf-f6294ea9e11c\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7crw8" Jan 21 11:43:12 crc kubenswrapper[4824]: I0121 11:43:12.683521 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7355a8fc-8c18-4de1-8baf-f6294ea9e11c-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-7crw8\" (UID: \"7355a8fc-8c18-4de1-8baf-f6294ea9e11c\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7crw8" Jan 21 11:43:12 crc kubenswrapper[4824]: I0121 11:43:12.683687 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/7355a8fc-8c18-4de1-8baf-f6294ea9e11c-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-7crw8\" (UID: \"7355a8fc-8c18-4de1-8baf-f6294ea9e11c\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7crw8" Jan 21 11:43:12 crc kubenswrapper[4824]: I0121 11:43:12.683793 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/7355a8fc-8c18-4de1-8baf-f6294ea9e11c-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-7crw8\" (UID: \"7355a8fc-8c18-4de1-8baf-f6294ea9e11c\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7crw8" Jan 21 11:43:12 crc kubenswrapper[4824]: I0121 11:43:12.683862 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/7355a8fc-8c18-4de1-8baf-f6294ea9e11c-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-7crw8\" (UID: \"7355a8fc-8c18-4de1-8baf-f6294ea9e11c\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7crw8" Jan 21 11:43:12 crc kubenswrapper[4824]: I0121 11:43:12.684172 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7355a8fc-8c18-4de1-8baf-f6294ea9e11c-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-7crw8\" (UID: \"7355a8fc-8c18-4de1-8baf-f6294ea9e11c\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7crw8" Jan 21 11:43:12 crc kubenswrapper[4824]: I0121 11:43:12.684316 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/7355a8fc-8c18-4de1-8baf-f6294ea9e11c-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-7crw8\" (UID: \"7355a8fc-8c18-4de1-8baf-f6294ea9e11c\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7crw8" Jan 21 11:43:12 crc kubenswrapper[4824]: I0121 11:43:12.684530 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/7355a8fc-8c18-4de1-8baf-f6294ea9e11c-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-7crw8\" (UID: \"7355a8fc-8c18-4de1-8baf-f6294ea9e11c\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7crw8" Jan 21 11:43:12 crc kubenswrapper[4824]: I0121 11:43:12.684570 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t2hm6\" (UniqueName: \"kubernetes.io/projected/7355a8fc-8c18-4de1-8baf-f6294ea9e11c-kube-api-access-t2hm6\") pod \"nova-edpm-deployment-openstack-edpm-ipam-7crw8\" (UID: \"7355a8fc-8c18-4de1-8baf-f6294ea9e11c\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7crw8" Jan 21 11:43:12 crc kubenswrapper[4824]: I0121 11:43:12.687819 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-7crw8"] Jan 21 11:43:12 crc kubenswrapper[4824]: I0121 11:43:12.787797 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/7355a8fc-8c18-4de1-8baf-f6294ea9e11c-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-7crw8\" (UID: \"7355a8fc-8c18-4de1-8baf-f6294ea9e11c\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7crw8" Jan 21 11:43:12 crc kubenswrapper[4824]: I0121 11:43:12.788229 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/7355a8fc-8c18-4de1-8baf-f6294ea9e11c-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-7crw8\" (UID: \"7355a8fc-8c18-4de1-8baf-f6294ea9e11c\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7crw8" Jan 21 11:43:12 crc kubenswrapper[4824]: I0121 11:43:12.788265 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t2hm6\" (UniqueName: \"kubernetes.io/projected/7355a8fc-8c18-4de1-8baf-f6294ea9e11c-kube-api-access-t2hm6\") pod \"nova-edpm-deployment-openstack-edpm-ipam-7crw8\" (UID: \"7355a8fc-8c18-4de1-8baf-f6294ea9e11c\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7crw8" Jan 21 11:43:12 crc kubenswrapper[4824]: I0121 11:43:12.788422 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/7355a8fc-8c18-4de1-8baf-f6294ea9e11c-ssh-key-openstack-edpm-ipam\") pod \"nova-edpm-deployment-openstack-edpm-ipam-7crw8\" (UID: \"7355a8fc-8c18-4de1-8baf-f6294ea9e11c\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7crw8" Jan 21 11:43:12 crc kubenswrapper[4824]: I0121 11:43:12.788484 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7355a8fc-8c18-4de1-8baf-f6294ea9e11c-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-7crw8\" (UID: \"7355a8fc-8c18-4de1-8baf-f6294ea9e11c\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7crw8" Jan 21 11:43:12 crc kubenswrapper[4824]: I0121 11:43:12.788545 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/7355a8fc-8c18-4de1-8baf-f6294ea9e11c-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-7crw8\" (UID: \"7355a8fc-8c18-4de1-8baf-f6294ea9e11c\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7crw8" Jan 21 11:43:12 crc kubenswrapper[4824]: I0121 11:43:12.788611 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/7355a8fc-8c18-4de1-8baf-f6294ea9e11c-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-7crw8\" (UID: \"7355a8fc-8c18-4de1-8baf-f6294ea9e11c\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7crw8" Jan 21 11:43:12 crc kubenswrapper[4824]: I0121 11:43:12.788650 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/7355a8fc-8c18-4de1-8baf-f6294ea9e11c-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-7crw8\" (UID: \"7355a8fc-8c18-4de1-8baf-f6294ea9e11c\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7crw8" Jan 21 11:43:12 crc kubenswrapper[4824]: I0121 11:43:12.788797 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7355a8fc-8c18-4de1-8baf-f6294ea9e11c-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-7crw8\" (UID: \"7355a8fc-8c18-4de1-8baf-f6294ea9e11c\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7crw8" Jan 21 11:43:12 crc kubenswrapper[4824]: I0121 11:43:12.789530 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/7355a8fc-8c18-4de1-8baf-f6294ea9e11c-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-7crw8\" (UID: \"7355a8fc-8c18-4de1-8baf-f6294ea9e11c\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7crw8" Jan 21 11:43:12 crc kubenswrapper[4824]: I0121 11:43:12.791898 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7355a8fc-8c18-4de1-8baf-f6294ea9e11c-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-7crw8\" (UID: \"7355a8fc-8c18-4de1-8baf-f6294ea9e11c\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7crw8" Jan 21 11:43:12 crc kubenswrapper[4824]: I0121 11:43:12.792184 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/7355a8fc-8c18-4de1-8baf-f6294ea9e11c-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-7crw8\" (UID: \"7355a8fc-8c18-4de1-8baf-f6294ea9e11c\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7crw8" Jan 21 11:43:12 crc kubenswrapper[4824]: I0121 11:43:12.792444 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/7355a8fc-8c18-4de1-8baf-f6294ea9e11c-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-7crw8\" (UID: \"7355a8fc-8c18-4de1-8baf-f6294ea9e11c\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7crw8" Jan 21 11:43:12 crc kubenswrapper[4824]: I0121 11:43:12.793450 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7355a8fc-8c18-4de1-8baf-f6294ea9e11c-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-7crw8\" (UID: \"7355a8fc-8c18-4de1-8baf-f6294ea9e11c\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7crw8" Jan 21 11:43:12 crc kubenswrapper[4824]: I0121 11:43:12.794545 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/7355a8fc-8c18-4de1-8baf-f6294ea9e11c-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-7crw8\" (UID: \"7355a8fc-8c18-4de1-8baf-f6294ea9e11c\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7crw8" Jan 21 11:43:12 crc kubenswrapper[4824]: I0121 11:43:12.794946 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/7355a8fc-8c18-4de1-8baf-f6294ea9e11c-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-7crw8\" (UID: \"7355a8fc-8c18-4de1-8baf-f6294ea9e11c\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7crw8" Jan 21 11:43:12 crc kubenswrapper[4824]: I0121 11:43:12.795399 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/7355a8fc-8c18-4de1-8baf-f6294ea9e11c-ssh-key-openstack-edpm-ipam\") pod \"nova-edpm-deployment-openstack-edpm-ipam-7crw8\" (UID: \"7355a8fc-8c18-4de1-8baf-f6294ea9e11c\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7crw8" Jan 21 11:43:12 crc kubenswrapper[4824]: I0121 11:43:12.801415 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t2hm6\" (UniqueName: \"kubernetes.io/projected/7355a8fc-8c18-4de1-8baf-f6294ea9e11c-kube-api-access-t2hm6\") pod \"nova-edpm-deployment-openstack-edpm-ipam-7crw8\" (UID: \"7355a8fc-8c18-4de1-8baf-f6294ea9e11c\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7crw8" Jan 21 11:43:12 crc kubenswrapper[4824]: I0121 11:43:12.993250 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7crw8" Jan 21 11:43:13 crc kubenswrapper[4824]: I0121 11:43:13.436716 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-7crw8"] Jan 21 11:43:13 crc kubenswrapper[4824]: I0121 11:43:13.450251 4824 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 21 11:43:13 crc kubenswrapper[4824]: I0121 11:43:13.614756 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7crw8" event={"ID":"7355a8fc-8c18-4de1-8baf-f6294ea9e11c","Type":"ContainerStarted","Data":"800a2f48f970dd7813b9b6aa1e55ec8212739dafcfa774a9c1c44f5bcec499fb"} Jan 21 11:43:14 crc kubenswrapper[4824]: I0121 11:43:14.621845 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7crw8" event={"ID":"7355a8fc-8c18-4de1-8baf-f6294ea9e11c","Type":"ContainerStarted","Data":"ffa277a7594795b5b30433a2fa1d40a85f0bccc5efe3c5c45b34f69bd8dc1a6d"} Jan 21 11:43:14 crc kubenswrapper[4824]: I0121 11:43:14.639371 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7crw8" podStartSLOduration=2.191622291 podStartE2EDuration="2.639357851s" podCreationTimestamp="2026-01-21 11:43:12 +0000 UTC" firstStartedPulling="2026-01-21 11:43:13.450026727 +0000 UTC m=+1935.743056020" lastFinishedPulling="2026-01-21 11:43:13.897762298 +0000 UTC m=+1936.190791580" observedRunningTime="2026-01-21 11:43:14.632796099 +0000 UTC m=+1936.925825391" watchObservedRunningTime="2026-01-21 11:43:14.639357851 +0000 UTC m=+1936.932387143" Jan 21 11:43:46 crc kubenswrapper[4824]: I0121 11:43:46.065216 4824 patch_prober.go:28] interesting pod/machine-config-daemon-zqjg9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 21 11:43:46 crc kubenswrapper[4824]: I0121 11:43:46.065609 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 21 11:44:16 crc kubenswrapper[4824]: I0121 11:44:16.065326 4824 patch_prober.go:28] interesting pod/machine-config-daemon-zqjg9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 21 11:44:16 crc kubenswrapper[4824]: I0121 11:44:16.065735 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 21 11:44:46 crc kubenswrapper[4824]: I0121 11:44:46.065753 4824 patch_prober.go:28] interesting pod/machine-config-daemon-zqjg9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 21 11:44:46 crc kubenswrapper[4824]: I0121 11:44:46.066334 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 21 11:44:46 crc kubenswrapper[4824]: I0121 11:44:46.066370 4824 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" Jan 21 11:44:46 crc kubenswrapper[4824]: I0121 11:44:46.067007 4824 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"1283af2f214718d7c6e80663aef792afcaa88b12e9e64dbc3ab357a33b0df07b"} pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 21 11:44:46 crc kubenswrapper[4824]: I0121 11:44:46.067054 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" containerName="machine-config-daemon" containerID="cri-o://1283af2f214718d7c6e80663aef792afcaa88b12e9e64dbc3ab357a33b0df07b" gracePeriod=600 Jan 21 11:44:46 crc kubenswrapper[4824]: I0121 11:44:46.209887 4824 generic.go:334] "Generic (PLEG): container finished" podID="33f3d922-4ffe-409b-a49a-d88c85898260" containerID="1283af2f214718d7c6e80663aef792afcaa88b12e9e64dbc3ab357a33b0df07b" exitCode=0 Jan 21 11:44:46 crc kubenswrapper[4824]: I0121 11:44:46.209974 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" event={"ID":"33f3d922-4ffe-409b-a49a-d88c85898260","Type":"ContainerDied","Data":"1283af2f214718d7c6e80663aef792afcaa88b12e9e64dbc3ab357a33b0df07b"} Jan 21 11:44:46 crc kubenswrapper[4824]: I0121 11:44:46.210166 4824 scope.go:117] "RemoveContainer" containerID="1cd835fe36ebb6541740e5ab5fd843470d46dfb78975fd968e5cc9aaef462499" Jan 21 11:44:47 crc kubenswrapper[4824]: I0121 11:44:47.217757 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" event={"ID":"33f3d922-4ffe-409b-a49a-d88c85898260","Type":"ContainerStarted","Data":"9e4a3d10f6183509a66349b3f877d2ea1f0b26ebbfb7ecfc617c73f022a56788"} Jan 21 11:44:57 crc kubenswrapper[4824]: I0121 11:44:57.287685 4824 generic.go:334] "Generic (PLEG): container finished" podID="7355a8fc-8c18-4de1-8baf-f6294ea9e11c" containerID="ffa277a7594795b5b30433a2fa1d40a85f0bccc5efe3c5c45b34f69bd8dc1a6d" exitCode=0 Jan 21 11:44:57 crc kubenswrapper[4824]: I0121 11:44:57.287773 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7crw8" event={"ID":"7355a8fc-8c18-4de1-8baf-f6294ea9e11c","Type":"ContainerDied","Data":"ffa277a7594795b5b30433a2fa1d40a85f0bccc5efe3c5c45b34f69bd8dc1a6d"} Jan 21 11:44:58 crc kubenswrapper[4824]: I0121 11:44:58.597288 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7crw8" Jan 21 11:44:58 crc kubenswrapper[4824]: I0121 11:44:58.755048 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/7355a8fc-8c18-4de1-8baf-f6294ea9e11c-nova-migration-ssh-key-1\") pod \"7355a8fc-8c18-4de1-8baf-f6294ea9e11c\" (UID: \"7355a8fc-8c18-4de1-8baf-f6294ea9e11c\") " Jan 21 11:44:58 crc kubenswrapper[4824]: I0121 11:44:58.755083 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7355a8fc-8c18-4de1-8baf-f6294ea9e11c-nova-combined-ca-bundle\") pod \"7355a8fc-8c18-4de1-8baf-f6294ea9e11c\" (UID: \"7355a8fc-8c18-4de1-8baf-f6294ea9e11c\") " Jan 21 11:44:58 crc kubenswrapper[4824]: I0121 11:44:58.755101 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/7355a8fc-8c18-4de1-8baf-f6294ea9e11c-nova-extra-config-0\") pod \"7355a8fc-8c18-4de1-8baf-f6294ea9e11c\" (UID: \"7355a8fc-8c18-4de1-8baf-f6294ea9e11c\") " Jan 21 11:44:58 crc kubenswrapper[4824]: I0121 11:44:58.755165 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/7355a8fc-8c18-4de1-8baf-f6294ea9e11c-ssh-key-openstack-edpm-ipam\") pod \"7355a8fc-8c18-4de1-8baf-f6294ea9e11c\" (UID: \"7355a8fc-8c18-4de1-8baf-f6294ea9e11c\") " Jan 21 11:44:58 crc kubenswrapper[4824]: I0121 11:44:58.755182 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7355a8fc-8c18-4de1-8baf-f6294ea9e11c-inventory\") pod \"7355a8fc-8c18-4de1-8baf-f6294ea9e11c\" (UID: \"7355a8fc-8c18-4de1-8baf-f6294ea9e11c\") " Jan 21 11:44:58 crc kubenswrapper[4824]: I0121 11:44:58.755265 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/7355a8fc-8c18-4de1-8baf-f6294ea9e11c-nova-cell1-compute-config-1\") pod \"7355a8fc-8c18-4de1-8baf-f6294ea9e11c\" (UID: \"7355a8fc-8c18-4de1-8baf-f6294ea9e11c\") " Jan 21 11:44:58 crc kubenswrapper[4824]: I0121 11:44:58.755295 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/7355a8fc-8c18-4de1-8baf-f6294ea9e11c-nova-migration-ssh-key-0\") pod \"7355a8fc-8c18-4de1-8baf-f6294ea9e11c\" (UID: \"7355a8fc-8c18-4de1-8baf-f6294ea9e11c\") " Jan 21 11:44:58 crc kubenswrapper[4824]: I0121 11:44:58.755372 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t2hm6\" (UniqueName: \"kubernetes.io/projected/7355a8fc-8c18-4de1-8baf-f6294ea9e11c-kube-api-access-t2hm6\") pod \"7355a8fc-8c18-4de1-8baf-f6294ea9e11c\" (UID: \"7355a8fc-8c18-4de1-8baf-f6294ea9e11c\") " Jan 21 11:44:58 crc kubenswrapper[4824]: I0121 11:44:58.755408 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/7355a8fc-8c18-4de1-8baf-f6294ea9e11c-nova-cell1-compute-config-0\") pod \"7355a8fc-8c18-4de1-8baf-f6294ea9e11c\" (UID: \"7355a8fc-8c18-4de1-8baf-f6294ea9e11c\") " Jan 21 11:44:58 crc kubenswrapper[4824]: I0121 11:44:58.760738 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7355a8fc-8c18-4de1-8baf-f6294ea9e11c-kube-api-access-t2hm6" (OuterVolumeSpecName: "kube-api-access-t2hm6") pod "7355a8fc-8c18-4de1-8baf-f6294ea9e11c" (UID: "7355a8fc-8c18-4de1-8baf-f6294ea9e11c"). InnerVolumeSpecName "kube-api-access-t2hm6". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:44:58 crc kubenswrapper[4824]: I0121 11:44:58.760341 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7355a8fc-8c18-4de1-8baf-f6294ea9e11c-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "7355a8fc-8c18-4de1-8baf-f6294ea9e11c" (UID: "7355a8fc-8c18-4de1-8baf-f6294ea9e11c"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:44:58 crc kubenswrapper[4824]: I0121 11:44:58.776080 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7355a8fc-8c18-4de1-8baf-f6294ea9e11c-nova-extra-config-0" (OuterVolumeSpecName: "nova-extra-config-0") pod "7355a8fc-8c18-4de1-8baf-f6294ea9e11c" (UID: "7355a8fc-8c18-4de1-8baf-f6294ea9e11c"). InnerVolumeSpecName "nova-extra-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:44:58 crc kubenswrapper[4824]: I0121 11:44:58.777408 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7355a8fc-8c18-4de1-8baf-f6294ea9e11c-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "7355a8fc-8c18-4de1-8baf-f6294ea9e11c" (UID: "7355a8fc-8c18-4de1-8baf-f6294ea9e11c"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:44:58 crc kubenswrapper[4824]: I0121 11:44:58.778018 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7355a8fc-8c18-4de1-8baf-f6294ea9e11c-inventory" (OuterVolumeSpecName: "inventory") pod "7355a8fc-8c18-4de1-8baf-f6294ea9e11c" (UID: "7355a8fc-8c18-4de1-8baf-f6294ea9e11c"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:44:58 crc kubenswrapper[4824]: I0121 11:44:58.778426 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7355a8fc-8c18-4de1-8baf-f6294ea9e11c-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "7355a8fc-8c18-4de1-8baf-f6294ea9e11c" (UID: "7355a8fc-8c18-4de1-8baf-f6294ea9e11c"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:44:58 crc kubenswrapper[4824]: I0121 11:44:58.779007 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7355a8fc-8c18-4de1-8baf-f6294ea9e11c-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "7355a8fc-8c18-4de1-8baf-f6294ea9e11c" (UID: "7355a8fc-8c18-4de1-8baf-f6294ea9e11c"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:44:58 crc kubenswrapper[4824]: I0121 11:44:58.779029 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7355a8fc-8c18-4de1-8baf-f6294ea9e11c-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "7355a8fc-8c18-4de1-8baf-f6294ea9e11c" (UID: "7355a8fc-8c18-4de1-8baf-f6294ea9e11c"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:44:58 crc kubenswrapper[4824]: I0121 11:44:58.781609 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7355a8fc-8c18-4de1-8baf-f6294ea9e11c-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "7355a8fc-8c18-4de1-8baf-f6294ea9e11c" (UID: "7355a8fc-8c18-4de1-8baf-f6294ea9e11c"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:44:58 crc kubenswrapper[4824]: I0121 11:44:58.857433 4824 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/7355a8fc-8c18-4de1-8baf-f6294ea9e11c-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Jan 21 11:44:58 crc kubenswrapper[4824]: I0121 11:44:58.857650 4824 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/7355a8fc-8c18-4de1-8baf-f6294ea9e11c-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Jan 21 11:44:58 crc kubenswrapper[4824]: I0121 11:44:58.857661 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t2hm6\" (UniqueName: \"kubernetes.io/projected/7355a8fc-8c18-4de1-8baf-f6294ea9e11c-kube-api-access-t2hm6\") on node \"crc\" DevicePath \"\"" Jan 21 11:44:58 crc kubenswrapper[4824]: I0121 11:44:58.857671 4824 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/7355a8fc-8c18-4de1-8baf-f6294ea9e11c-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Jan 21 11:44:58 crc kubenswrapper[4824]: I0121 11:44:58.857679 4824 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/7355a8fc-8c18-4de1-8baf-f6294ea9e11c-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Jan 21 11:44:58 crc kubenswrapper[4824]: I0121 11:44:58.857686 4824 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7355a8fc-8c18-4de1-8baf-f6294ea9e11c-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 21 11:44:58 crc kubenswrapper[4824]: I0121 11:44:58.857694 4824 reconciler_common.go:293] "Volume detached for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/7355a8fc-8c18-4de1-8baf-f6294ea9e11c-nova-extra-config-0\") on node \"crc\" DevicePath \"\"" Jan 21 11:44:58 crc kubenswrapper[4824]: I0121 11:44:58.857701 4824 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/7355a8fc-8c18-4de1-8baf-f6294ea9e11c-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Jan 21 11:44:58 crc kubenswrapper[4824]: I0121 11:44:58.857709 4824 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7355a8fc-8c18-4de1-8baf-f6294ea9e11c-inventory\") on node \"crc\" DevicePath \"\"" Jan 21 11:44:59 crc kubenswrapper[4824]: I0121 11:44:59.300785 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7crw8" event={"ID":"7355a8fc-8c18-4de1-8baf-f6294ea9e11c","Type":"ContainerDied","Data":"800a2f48f970dd7813b9b6aa1e55ec8212739dafcfa774a9c1c44f5bcec499fb"} Jan 21 11:44:59 crc kubenswrapper[4824]: I0121 11:44:59.300815 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-7crw8" Jan 21 11:44:59 crc kubenswrapper[4824]: I0121 11:44:59.300822 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="800a2f48f970dd7813b9b6aa1e55ec8212739dafcfa774a9c1c44f5bcec499fb" Jan 21 11:44:59 crc kubenswrapper[4824]: I0121 11:44:59.412339 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-g8gx6"] Jan 21 11:44:59 crc kubenswrapper[4824]: E0121 11:44:59.412852 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7355a8fc-8c18-4de1-8baf-f6294ea9e11c" containerName="nova-edpm-deployment-openstack-edpm-ipam" Jan 21 11:44:59 crc kubenswrapper[4824]: I0121 11:44:59.412938 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="7355a8fc-8c18-4de1-8baf-f6294ea9e11c" containerName="nova-edpm-deployment-openstack-edpm-ipam" Jan 21 11:44:59 crc kubenswrapper[4824]: I0121 11:44:59.413219 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="7355a8fc-8c18-4de1-8baf-f6294ea9e11c" containerName="nova-edpm-deployment-openstack-edpm-ipam" Jan 21 11:44:59 crc kubenswrapper[4824]: I0121 11:44:59.413771 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-g8gx6" Jan 21 11:44:59 crc kubenswrapper[4824]: I0121 11:44:59.417016 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-tcqk6" Jan 21 11:44:59 crc kubenswrapper[4824]: I0121 11:44:59.417136 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Jan 21 11:44:59 crc kubenswrapper[4824]: I0121 11:44:59.417525 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Jan 21 11:44:59 crc kubenswrapper[4824]: I0121 11:44:59.417777 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Jan 21 11:44:59 crc kubenswrapper[4824]: I0121 11:44:59.417939 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-compute-config-data" Jan 21 11:44:59 crc kubenswrapper[4824]: I0121 11:44:59.425029 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-g8gx6"] Jan 21 11:44:59 crc kubenswrapper[4824]: I0121 11:44:59.569159 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/49b2a136-e5c3-4741-80c6-4c545d4a82a0-ssh-key-openstack-edpm-ipam\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-g8gx6\" (UID: \"49b2a136-e5c3-4741-80c6-4c545d4a82a0\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-g8gx6" Jan 21 11:44:59 crc kubenswrapper[4824]: I0121 11:44:59.569258 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/49b2a136-e5c3-4741-80c6-4c545d4a82a0-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-g8gx6\" (UID: \"49b2a136-e5c3-4741-80c6-4c545d4a82a0\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-g8gx6" Jan 21 11:44:59 crc kubenswrapper[4824]: I0121 11:44:59.569370 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/49b2a136-e5c3-4741-80c6-4c545d4a82a0-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-g8gx6\" (UID: \"49b2a136-e5c3-4741-80c6-4c545d4a82a0\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-g8gx6" Jan 21 11:44:59 crc kubenswrapper[4824]: I0121 11:44:59.569405 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/49b2a136-e5c3-4741-80c6-4c545d4a82a0-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-g8gx6\" (UID: \"49b2a136-e5c3-4741-80c6-4c545d4a82a0\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-g8gx6" Jan 21 11:44:59 crc kubenswrapper[4824]: I0121 11:44:59.569444 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sgqg9\" (UniqueName: \"kubernetes.io/projected/49b2a136-e5c3-4741-80c6-4c545d4a82a0-kube-api-access-sgqg9\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-g8gx6\" (UID: \"49b2a136-e5c3-4741-80c6-4c545d4a82a0\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-g8gx6" Jan 21 11:44:59 crc kubenswrapper[4824]: I0121 11:44:59.569468 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/49b2a136-e5c3-4741-80c6-4c545d4a82a0-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-g8gx6\" (UID: \"49b2a136-e5c3-4741-80c6-4c545d4a82a0\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-g8gx6" Jan 21 11:44:59 crc kubenswrapper[4824]: I0121 11:44:59.569488 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/49b2a136-e5c3-4741-80c6-4c545d4a82a0-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-g8gx6\" (UID: \"49b2a136-e5c3-4741-80c6-4c545d4a82a0\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-g8gx6" Jan 21 11:44:59 crc kubenswrapper[4824]: I0121 11:44:59.671281 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/49b2a136-e5c3-4741-80c6-4c545d4a82a0-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-g8gx6\" (UID: \"49b2a136-e5c3-4741-80c6-4c545d4a82a0\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-g8gx6" Jan 21 11:44:59 crc kubenswrapper[4824]: I0121 11:44:59.671349 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/49b2a136-e5c3-4741-80c6-4c545d4a82a0-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-g8gx6\" (UID: \"49b2a136-e5c3-4741-80c6-4c545d4a82a0\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-g8gx6" Jan 21 11:44:59 crc kubenswrapper[4824]: I0121 11:44:59.671396 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sgqg9\" (UniqueName: \"kubernetes.io/projected/49b2a136-e5c3-4741-80c6-4c545d4a82a0-kube-api-access-sgqg9\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-g8gx6\" (UID: \"49b2a136-e5c3-4741-80c6-4c545d4a82a0\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-g8gx6" Jan 21 11:44:59 crc kubenswrapper[4824]: I0121 11:44:59.671423 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/49b2a136-e5c3-4741-80c6-4c545d4a82a0-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-g8gx6\" (UID: \"49b2a136-e5c3-4741-80c6-4c545d4a82a0\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-g8gx6" Jan 21 11:44:59 crc kubenswrapper[4824]: I0121 11:44:59.671444 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/49b2a136-e5c3-4741-80c6-4c545d4a82a0-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-g8gx6\" (UID: \"49b2a136-e5c3-4741-80c6-4c545d4a82a0\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-g8gx6" Jan 21 11:44:59 crc kubenswrapper[4824]: I0121 11:44:59.671520 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/49b2a136-e5c3-4741-80c6-4c545d4a82a0-ssh-key-openstack-edpm-ipam\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-g8gx6\" (UID: \"49b2a136-e5c3-4741-80c6-4c545d4a82a0\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-g8gx6" Jan 21 11:44:59 crc kubenswrapper[4824]: I0121 11:44:59.671644 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/49b2a136-e5c3-4741-80c6-4c545d4a82a0-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-g8gx6\" (UID: \"49b2a136-e5c3-4741-80c6-4c545d4a82a0\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-g8gx6" Jan 21 11:44:59 crc kubenswrapper[4824]: I0121 11:44:59.674626 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/49b2a136-e5c3-4741-80c6-4c545d4a82a0-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-g8gx6\" (UID: \"49b2a136-e5c3-4741-80c6-4c545d4a82a0\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-g8gx6" Jan 21 11:44:59 crc kubenswrapper[4824]: I0121 11:44:59.674718 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/49b2a136-e5c3-4741-80c6-4c545d4a82a0-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-g8gx6\" (UID: \"49b2a136-e5c3-4741-80c6-4c545d4a82a0\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-g8gx6" Jan 21 11:44:59 crc kubenswrapper[4824]: I0121 11:44:59.677455 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/49b2a136-e5c3-4741-80c6-4c545d4a82a0-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-g8gx6\" (UID: \"49b2a136-e5c3-4741-80c6-4c545d4a82a0\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-g8gx6" Jan 21 11:44:59 crc kubenswrapper[4824]: I0121 11:44:59.677551 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/49b2a136-e5c3-4741-80c6-4c545d4a82a0-ssh-key-openstack-edpm-ipam\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-g8gx6\" (UID: \"49b2a136-e5c3-4741-80c6-4c545d4a82a0\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-g8gx6" Jan 21 11:44:59 crc kubenswrapper[4824]: I0121 11:44:59.677578 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/49b2a136-e5c3-4741-80c6-4c545d4a82a0-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-g8gx6\" (UID: \"49b2a136-e5c3-4741-80c6-4c545d4a82a0\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-g8gx6" Jan 21 11:44:59 crc kubenswrapper[4824]: I0121 11:44:59.677838 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/49b2a136-e5c3-4741-80c6-4c545d4a82a0-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-g8gx6\" (UID: \"49b2a136-e5c3-4741-80c6-4c545d4a82a0\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-g8gx6" Jan 21 11:44:59 crc kubenswrapper[4824]: I0121 11:44:59.684773 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sgqg9\" (UniqueName: \"kubernetes.io/projected/49b2a136-e5c3-4741-80c6-4c545d4a82a0-kube-api-access-sgqg9\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-g8gx6\" (UID: \"49b2a136-e5c3-4741-80c6-4c545d4a82a0\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-g8gx6" Jan 21 11:44:59 crc kubenswrapper[4824]: I0121 11:44:59.729565 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-g8gx6" Jan 21 11:45:00 crc kubenswrapper[4824]: I0121 11:45:00.129684 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29483265-s59xh"] Jan 21 11:45:00 crc kubenswrapper[4824]: I0121 11:45:00.130974 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29483265-s59xh" Jan 21 11:45:00 crc kubenswrapper[4824]: I0121 11:45:00.132755 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Jan 21 11:45:00 crc kubenswrapper[4824]: I0121 11:45:00.133188 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Jan 21 11:45:00 crc kubenswrapper[4824]: I0121 11:45:00.139645 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29483265-s59xh"] Jan 21 11:45:00 crc kubenswrapper[4824]: I0121 11:45:00.233818 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-g8gx6"] Jan 21 11:45:00 crc kubenswrapper[4824]: I0121 11:45:00.282759 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/35e40551-5ca0-42ef-8305-66fd69d36a9f-secret-volume\") pod \"collect-profiles-29483265-s59xh\" (UID: \"35e40551-5ca0-42ef-8305-66fd69d36a9f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29483265-s59xh" Jan 21 11:45:00 crc kubenswrapper[4824]: I0121 11:45:00.283050 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/35e40551-5ca0-42ef-8305-66fd69d36a9f-config-volume\") pod \"collect-profiles-29483265-s59xh\" (UID: \"35e40551-5ca0-42ef-8305-66fd69d36a9f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29483265-s59xh" Jan 21 11:45:00 crc kubenswrapper[4824]: I0121 11:45:00.283329 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jgsr7\" (UniqueName: \"kubernetes.io/projected/35e40551-5ca0-42ef-8305-66fd69d36a9f-kube-api-access-jgsr7\") pod \"collect-profiles-29483265-s59xh\" (UID: \"35e40551-5ca0-42ef-8305-66fd69d36a9f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29483265-s59xh" Jan 21 11:45:00 crc kubenswrapper[4824]: I0121 11:45:00.307976 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-g8gx6" event={"ID":"49b2a136-e5c3-4741-80c6-4c545d4a82a0","Type":"ContainerStarted","Data":"b8c227627e16ce045a124bfbe36c3ecb09476a5cdfcf39963427bafa05e52cee"} Jan 21 11:45:00 crc kubenswrapper[4824]: I0121 11:45:00.384894 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jgsr7\" (UniqueName: \"kubernetes.io/projected/35e40551-5ca0-42ef-8305-66fd69d36a9f-kube-api-access-jgsr7\") pod \"collect-profiles-29483265-s59xh\" (UID: \"35e40551-5ca0-42ef-8305-66fd69d36a9f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29483265-s59xh" Jan 21 11:45:00 crc kubenswrapper[4824]: I0121 11:45:00.384986 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/35e40551-5ca0-42ef-8305-66fd69d36a9f-secret-volume\") pod \"collect-profiles-29483265-s59xh\" (UID: \"35e40551-5ca0-42ef-8305-66fd69d36a9f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29483265-s59xh" Jan 21 11:45:00 crc kubenswrapper[4824]: I0121 11:45:00.385082 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/35e40551-5ca0-42ef-8305-66fd69d36a9f-config-volume\") pod \"collect-profiles-29483265-s59xh\" (UID: \"35e40551-5ca0-42ef-8305-66fd69d36a9f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29483265-s59xh" Jan 21 11:45:00 crc kubenswrapper[4824]: I0121 11:45:00.385913 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/35e40551-5ca0-42ef-8305-66fd69d36a9f-config-volume\") pod \"collect-profiles-29483265-s59xh\" (UID: \"35e40551-5ca0-42ef-8305-66fd69d36a9f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29483265-s59xh" Jan 21 11:45:00 crc kubenswrapper[4824]: I0121 11:45:00.389803 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/35e40551-5ca0-42ef-8305-66fd69d36a9f-secret-volume\") pod \"collect-profiles-29483265-s59xh\" (UID: \"35e40551-5ca0-42ef-8305-66fd69d36a9f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29483265-s59xh" Jan 21 11:45:00 crc kubenswrapper[4824]: I0121 11:45:00.398041 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jgsr7\" (UniqueName: \"kubernetes.io/projected/35e40551-5ca0-42ef-8305-66fd69d36a9f-kube-api-access-jgsr7\") pod \"collect-profiles-29483265-s59xh\" (UID: \"35e40551-5ca0-42ef-8305-66fd69d36a9f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29483265-s59xh" Jan 21 11:45:00 crc kubenswrapper[4824]: I0121 11:45:00.450311 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29483265-s59xh" Jan 21 11:45:00 crc kubenswrapper[4824]: I0121 11:45:00.817730 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29483265-s59xh"] Jan 21 11:45:00 crc kubenswrapper[4824]: W0121 11:45:00.819586 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod35e40551_5ca0_42ef_8305_66fd69d36a9f.slice/crio-77b8a29c451e3dc008e9eba1be68bd3ab5f2ce3556851124403bf6c37f3ce310 WatchSource:0}: Error finding container 77b8a29c451e3dc008e9eba1be68bd3ab5f2ce3556851124403bf6c37f3ce310: Status 404 returned error can't find the container with id 77b8a29c451e3dc008e9eba1be68bd3ab5f2ce3556851124403bf6c37f3ce310 Jan 21 11:45:01 crc kubenswrapper[4824]: I0121 11:45:01.316128 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-g8gx6" event={"ID":"49b2a136-e5c3-4741-80c6-4c545d4a82a0","Type":"ContainerStarted","Data":"6cc48beb53c571878821f1833c1a75164e55e0ffea7eb2292d2abd90321cb03f"} Jan 21 11:45:01 crc kubenswrapper[4824]: I0121 11:45:01.317649 4824 generic.go:334] "Generic (PLEG): container finished" podID="35e40551-5ca0-42ef-8305-66fd69d36a9f" containerID="34d65e26d2a1896dfa268eb4b192173adbc67325a4a1d10c57a0833b00d02595" exitCode=0 Jan 21 11:45:01 crc kubenswrapper[4824]: I0121 11:45:01.317682 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29483265-s59xh" event={"ID":"35e40551-5ca0-42ef-8305-66fd69d36a9f","Type":"ContainerDied","Data":"34d65e26d2a1896dfa268eb4b192173adbc67325a4a1d10c57a0833b00d02595"} Jan 21 11:45:01 crc kubenswrapper[4824]: I0121 11:45:01.317711 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29483265-s59xh" event={"ID":"35e40551-5ca0-42ef-8305-66fd69d36a9f","Type":"ContainerStarted","Data":"77b8a29c451e3dc008e9eba1be68bd3ab5f2ce3556851124403bf6c37f3ce310"} Jan 21 11:45:01 crc kubenswrapper[4824]: I0121 11:45:01.330536 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-g8gx6" podStartSLOduration=1.820143318 podStartE2EDuration="2.330522137s" podCreationTimestamp="2026-01-21 11:44:59 +0000 UTC" firstStartedPulling="2026-01-21 11:45:00.24527848 +0000 UTC m=+2042.538307773" lastFinishedPulling="2026-01-21 11:45:00.75565731 +0000 UTC m=+2043.048686592" observedRunningTime="2026-01-21 11:45:01.327297198 +0000 UTC m=+2043.620326490" watchObservedRunningTime="2026-01-21 11:45:01.330522137 +0000 UTC m=+2043.623551429" Jan 21 11:45:02 crc kubenswrapper[4824]: I0121 11:45:02.574264 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29483265-s59xh" Jan 21 11:45:02 crc kubenswrapper[4824]: I0121 11:45:02.720652 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/35e40551-5ca0-42ef-8305-66fd69d36a9f-secret-volume\") pod \"35e40551-5ca0-42ef-8305-66fd69d36a9f\" (UID: \"35e40551-5ca0-42ef-8305-66fd69d36a9f\") " Jan 21 11:45:02 crc kubenswrapper[4824]: I0121 11:45:02.720781 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/35e40551-5ca0-42ef-8305-66fd69d36a9f-config-volume\") pod \"35e40551-5ca0-42ef-8305-66fd69d36a9f\" (UID: \"35e40551-5ca0-42ef-8305-66fd69d36a9f\") " Jan 21 11:45:02 crc kubenswrapper[4824]: I0121 11:45:02.720817 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jgsr7\" (UniqueName: \"kubernetes.io/projected/35e40551-5ca0-42ef-8305-66fd69d36a9f-kube-api-access-jgsr7\") pod \"35e40551-5ca0-42ef-8305-66fd69d36a9f\" (UID: \"35e40551-5ca0-42ef-8305-66fd69d36a9f\") " Jan 21 11:45:02 crc kubenswrapper[4824]: I0121 11:45:02.721342 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/35e40551-5ca0-42ef-8305-66fd69d36a9f-config-volume" (OuterVolumeSpecName: "config-volume") pod "35e40551-5ca0-42ef-8305-66fd69d36a9f" (UID: "35e40551-5ca0-42ef-8305-66fd69d36a9f"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:45:02 crc kubenswrapper[4824]: I0121 11:45:02.721489 4824 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/35e40551-5ca0-42ef-8305-66fd69d36a9f-config-volume\") on node \"crc\" DevicePath \"\"" Jan 21 11:45:02 crc kubenswrapper[4824]: I0121 11:45:02.725728 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/35e40551-5ca0-42ef-8305-66fd69d36a9f-kube-api-access-jgsr7" (OuterVolumeSpecName: "kube-api-access-jgsr7") pod "35e40551-5ca0-42ef-8305-66fd69d36a9f" (UID: "35e40551-5ca0-42ef-8305-66fd69d36a9f"). InnerVolumeSpecName "kube-api-access-jgsr7". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:45:02 crc kubenswrapper[4824]: I0121 11:45:02.725744 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/35e40551-5ca0-42ef-8305-66fd69d36a9f-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "35e40551-5ca0-42ef-8305-66fd69d36a9f" (UID: "35e40551-5ca0-42ef-8305-66fd69d36a9f"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:45:02 crc kubenswrapper[4824]: I0121 11:45:02.823456 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jgsr7\" (UniqueName: \"kubernetes.io/projected/35e40551-5ca0-42ef-8305-66fd69d36a9f-kube-api-access-jgsr7\") on node \"crc\" DevicePath \"\"" Jan 21 11:45:02 crc kubenswrapper[4824]: I0121 11:45:02.823500 4824 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/35e40551-5ca0-42ef-8305-66fd69d36a9f-secret-volume\") on node \"crc\" DevicePath \"\"" Jan 21 11:45:03 crc kubenswrapper[4824]: I0121 11:45:03.332883 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29483265-s59xh" event={"ID":"35e40551-5ca0-42ef-8305-66fd69d36a9f","Type":"ContainerDied","Data":"77b8a29c451e3dc008e9eba1be68bd3ab5f2ce3556851124403bf6c37f3ce310"} Jan 21 11:45:03 crc kubenswrapper[4824]: I0121 11:45:03.332919 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29483265-s59xh" Jan 21 11:45:03 crc kubenswrapper[4824]: I0121 11:45:03.332932 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="77b8a29c451e3dc008e9eba1be68bd3ab5f2ce3556851124403bf6c37f3ce310" Jan 21 11:45:03 crc kubenswrapper[4824]: I0121 11:45:03.624426 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29483220-6r4fz"] Jan 21 11:45:03 crc kubenswrapper[4824]: I0121 11:45:03.630563 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29483220-6r4fz"] Jan 21 11:45:04 crc kubenswrapper[4824]: I0121 11:45:04.056722 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09bd0b9c-46fb-433f-bd30-014c69c90d9f" path="/var/lib/kubelet/pods/09bd0b9c-46fb-433f-bd30-014c69c90d9f/volumes" Jan 21 11:46:00 crc kubenswrapper[4824]: I0121 11:46:00.758616 4824 scope.go:117] "RemoveContainer" containerID="277011d8fed4b1333d53f5a4a084181f52dada1e6d85b31750d9bab9937b3678" Jan 21 11:46:46 crc kubenswrapper[4824]: I0121 11:46:46.065075 4824 patch_prober.go:28] interesting pod/machine-config-daemon-zqjg9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 21 11:46:46 crc kubenswrapper[4824]: I0121 11:46:46.065468 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 21 11:46:47 crc kubenswrapper[4824]: I0121 11:46:47.974023 4824 generic.go:334] "Generic (PLEG): container finished" podID="49b2a136-e5c3-4741-80c6-4c545d4a82a0" containerID="6cc48beb53c571878821f1833c1a75164e55e0ffea7eb2292d2abd90321cb03f" exitCode=0 Jan 21 11:46:47 crc kubenswrapper[4824]: I0121 11:46:47.974112 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-g8gx6" event={"ID":"49b2a136-e5c3-4741-80c6-4c545d4a82a0","Type":"ContainerDied","Data":"6cc48beb53c571878821f1833c1a75164e55e0ffea7eb2292d2abd90321cb03f"} Jan 21 11:46:49 crc kubenswrapper[4824]: I0121 11:46:49.297347 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-g8gx6" Jan 21 11:46:49 crc kubenswrapper[4824]: I0121 11:46:49.420017 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/49b2a136-e5c3-4741-80c6-4c545d4a82a0-ceilometer-compute-config-data-0\") pod \"49b2a136-e5c3-4741-80c6-4c545d4a82a0\" (UID: \"49b2a136-e5c3-4741-80c6-4c545d4a82a0\") " Jan 21 11:46:49 crc kubenswrapper[4824]: I0121 11:46:49.420072 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sgqg9\" (UniqueName: \"kubernetes.io/projected/49b2a136-e5c3-4741-80c6-4c545d4a82a0-kube-api-access-sgqg9\") pod \"49b2a136-e5c3-4741-80c6-4c545d4a82a0\" (UID: \"49b2a136-e5c3-4741-80c6-4c545d4a82a0\") " Jan 21 11:46:49 crc kubenswrapper[4824]: I0121 11:46:49.420150 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/49b2a136-e5c3-4741-80c6-4c545d4a82a0-ssh-key-openstack-edpm-ipam\") pod \"49b2a136-e5c3-4741-80c6-4c545d4a82a0\" (UID: \"49b2a136-e5c3-4741-80c6-4c545d4a82a0\") " Jan 21 11:46:49 crc kubenswrapper[4824]: I0121 11:46:49.420193 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/49b2a136-e5c3-4741-80c6-4c545d4a82a0-inventory\") pod \"49b2a136-e5c3-4741-80c6-4c545d4a82a0\" (UID: \"49b2a136-e5c3-4741-80c6-4c545d4a82a0\") " Jan 21 11:46:49 crc kubenswrapper[4824]: I0121 11:46:49.420246 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/49b2a136-e5c3-4741-80c6-4c545d4a82a0-ceilometer-compute-config-data-2\") pod \"49b2a136-e5c3-4741-80c6-4c545d4a82a0\" (UID: \"49b2a136-e5c3-4741-80c6-4c545d4a82a0\") " Jan 21 11:46:49 crc kubenswrapper[4824]: I0121 11:46:49.420269 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/49b2a136-e5c3-4741-80c6-4c545d4a82a0-telemetry-combined-ca-bundle\") pod \"49b2a136-e5c3-4741-80c6-4c545d4a82a0\" (UID: \"49b2a136-e5c3-4741-80c6-4c545d4a82a0\") " Jan 21 11:46:49 crc kubenswrapper[4824]: I0121 11:46:49.420322 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/49b2a136-e5c3-4741-80c6-4c545d4a82a0-ceilometer-compute-config-data-1\") pod \"49b2a136-e5c3-4741-80c6-4c545d4a82a0\" (UID: \"49b2a136-e5c3-4741-80c6-4c545d4a82a0\") " Jan 21 11:46:49 crc kubenswrapper[4824]: I0121 11:46:49.424468 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49b2a136-e5c3-4741-80c6-4c545d4a82a0-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "49b2a136-e5c3-4741-80c6-4c545d4a82a0" (UID: "49b2a136-e5c3-4741-80c6-4c545d4a82a0"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:46:49 crc kubenswrapper[4824]: I0121 11:46:49.424564 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49b2a136-e5c3-4741-80c6-4c545d4a82a0-kube-api-access-sgqg9" (OuterVolumeSpecName: "kube-api-access-sgqg9") pod "49b2a136-e5c3-4741-80c6-4c545d4a82a0" (UID: "49b2a136-e5c3-4741-80c6-4c545d4a82a0"). InnerVolumeSpecName "kube-api-access-sgqg9". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:46:49 crc kubenswrapper[4824]: I0121 11:46:49.440567 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49b2a136-e5c3-4741-80c6-4c545d4a82a0-inventory" (OuterVolumeSpecName: "inventory") pod "49b2a136-e5c3-4741-80c6-4c545d4a82a0" (UID: "49b2a136-e5c3-4741-80c6-4c545d4a82a0"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:46:49 crc kubenswrapper[4824]: I0121 11:46:49.440877 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49b2a136-e5c3-4741-80c6-4c545d4a82a0-ceilometer-compute-config-data-1" (OuterVolumeSpecName: "ceilometer-compute-config-data-1") pod "49b2a136-e5c3-4741-80c6-4c545d4a82a0" (UID: "49b2a136-e5c3-4741-80c6-4c545d4a82a0"). InnerVolumeSpecName "ceilometer-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:46:49 crc kubenswrapper[4824]: I0121 11:46:49.441805 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49b2a136-e5c3-4741-80c6-4c545d4a82a0-ceilometer-compute-config-data-2" (OuterVolumeSpecName: "ceilometer-compute-config-data-2") pod "49b2a136-e5c3-4741-80c6-4c545d4a82a0" (UID: "49b2a136-e5c3-4741-80c6-4c545d4a82a0"). InnerVolumeSpecName "ceilometer-compute-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:46:49 crc kubenswrapper[4824]: I0121 11:46:49.441889 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49b2a136-e5c3-4741-80c6-4c545d4a82a0-ceilometer-compute-config-data-0" (OuterVolumeSpecName: "ceilometer-compute-config-data-0") pod "49b2a136-e5c3-4741-80c6-4c545d4a82a0" (UID: "49b2a136-e5c3-4741-80c6-4c545d4a82a0"). InnerVolumeSpecName "ceilometer-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:46:49 crc kubenswrapper[4824]: I0121 11:46:49.442924 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49b2a136-e5c3-4741-80c6-4c545d4a82a0-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "49b2a136-e5c3-4741-80c6-4c545d4a82a0" (UID: "49b2a136-e5c3-4741-80c6-4c545d4a82a0"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:46:49 crc kubenswrapper[4824]: I0121 11:46:49.522332 4824 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/49b2a136-e5c3-4741-80c6-4c545d4a82a0-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Jan 21 11:46:49 crc kubenswrapper[4824]: I0121 11:46:49.522728 4824 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/49b2a136-e5c3-4741-80c6-4c545d4a82a0-inventory\") on node \"crc\" DevicePath \"\"" Jan 21 11:46:49 crc kubenswrapper[4824]: I0121 11:46:49.522797 4824 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/49b2a136-e5c3-4741-80c6-4c545d4a82a0-ceilometer-compute-config-data-2\") on node \"crc\" DevicePath \"\"" Jan 21 11:46:49 crc kubenswrapper[4824]: I0121 11:46:49.522863 4824 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/49b2a136-e5c3-4741-80c6-4c545d4a82a0-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 21 11:46:49 crc kubenswrapper[4824]: I0121 11:46:49.522921 4824 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/49b2a136-e5c3-4741-80c6-4c545d4a82a0-ceilometer-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Jan 21 11:46:49 crc kubenswrapper[4824]: I0121 11:46:49.523010 4824 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/49b2a136-e5c3-4741-80c6-4c545d4a82a0-ceilometer-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Jan 21 11:46:49 crc kubenswrapper[4824]: I0121 11:46:49.523069 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sgqg9\" (UniqueName: \"kubernetes.io/projected/49b2a136-e5c3-4741-80c6-4c545d4a82a0-kube-api-access-sgqg9\") on node \"crc\" DevicePath \"\"" Jan 21 11:46:49 crc kubenswrapper[4824]: I0121 11:46:49.990056 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-g8gx6" event={"ID":"49b2a136-e5c3-4741-80c6-4c545d4a82a0","Type":"ContainerDied","Data":"b8c227627e16ce045a124bfbe36c3ecb09476a5cdfcf39963427bafa05e52cee"} Jan 21 11:46:49 crc kubenswrapper[4824]: I0121 11:46:49.990290 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b8c227627e16ce045a124bfbe36c3ecb09476a5cdfcf39963427bafa05e52cee" Jan 21 11:46:49 crc kubenswrapper[4824]: I0121 11:46:49.990115 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-g8gx6" Jan 21 11:47:09 crc kubenswrapper[4824]: I0121 11:47:09.032567 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-kwnpj"] Jan 21 11:47:09 crc kubenswrapper[4824]: E0121 11:47:09.033295 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35e40551-5ca0-42ef-8305-66fd69d36a9f" containerName="collect-profiles" Jan 21 11:47:09 crc kubenswrapper[4824]: I0121 11:47:09.033309 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="35e40551-5ca0-42ef-8305-66fd69d36a9f" containerName="collect-profiles" Jan 21 11:47:09 crc kubenswrapper[4824]: E0121 11:47:09.033325 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="49b2a136-e5c3-4741-80c6-4c545d4a82a0" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Jan 21 11:47:09 crc kubenswrapper[4824]: I0121 11:47:09.033331 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="49b2a136-e5c3-4741-80c6-4c545d4a82a0" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Jan 21 11:47:09 crc kubenswrapper[4824]: I0121 11:47:09.033498 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="35e40551-5ca0-42ef-8305-66fd69d36a9f" containerName="collect-profiles" Jan 21 11:47:09 crc kubenswrapper[4824]: I0121 11:47:09.033517 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="49b2a136-e5c3-4741-80c6-4c545d4a82a0" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Jan 21 11:47:09 crc kubenswrapper[4824]: I0121 11:47:09.034669 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kwnpj" Jan 21 11:47:09 crc kubenswrapper[4824]: I0121 11:47:09.046642 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-kwnpj"] Jan 21 11:47:09 crc kubenswrapper[4824]: I0121 11:47:09.147752 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1ffd9d0f-1714-416f-84a9-4f3ab0305ee2-catalog-content\") pod \"redhat-marketplace-kwnpj\" (UID: \"1ffd9d0f-1714-416f-84a9-4f3ab0305ee2\") " pod="openshift-marketplace/redhat-marketplace-kwnpj" Jan 21 11:47:09 crc kubenswrapper[4824]: I0121 11:47:09.147799 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1ffd9d0f-1714-416f-84a9-4f3ab0305ee2-utilities\") pod \"redhat-marketplace-kwnpj\" (UID: \"1ffd9d0f-1714-416f-84a9-4f3ab0305ee2\") " pod="openshift-marketplace/redhat-marketplace-kwnpj" Jan 21 11:47:09 crc kubenswrapper[4824]: I0121 11:47:09.148043 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-spzf5\" (UniqueName: \"kubernetes.io/projected/1ffd9d0f-1714-416f-84a9-4f3ab0305ee2-kube-api-access-spzf5\") pod \"redhat-marketplace-kwnpj\" (UID: \"1ffd9d0f-1714-416f-84a9-4f3ab0305ee2\") " pod="openshift-marketplace/redhat-marketplace-kwnpj" Jan 21 11:47:09 crc kubenswrapper[4824]: I0121 11:47:09.249919 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1ffd9d0f-1714-416f-84a9-4f3ab0305ee2-catalog-content\") pod \"redhat-marketplace-kwnpj\" (UID: \"1ffd9d0f-1714-416f-84a9-4f3ab0305ee2\") " pod="openshift-marketplace/redhat-marketplace-kwnpj" Jan 21 11:47:09 crc kubenswrapper[4824]: I0121 11:47:09.249975 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1ffd9d0f-1714-416f-84a9-4f3ab0305ee2-utilities\") pod \"redhat-marketplace-kwnpj\" (UID: \"1ffd9d0f-1714-416f-84a9-4f3ab0305ee2\") " pod="openshift-marketplace/redhat-marketplace-kwnpj" Jan 21 11:47:09 crc kubenswrapper[4824]: I0121 11:47:09.250030 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-spzf5\" (UniqueName: \"kubernetes.io/projected/1ffd9d0f-1714-416f-84a9-4f3ab0305ee2-kube-api-access-spzf5\") pod \"redhat-marketplace-kwnpj\" (UID: \"1ffd9d0f-1714-416f-84a9-4f3ab0305ee2\") " pod="openshift-marketplace/redhat-marketplace-kwnpj" Jan 21 11:47:09 crc kubenswrapper[4824]: I0121 11:47:09.250421 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1ffd9d0f-1714-416f-84a9-4f3ab0305ee2-utilities\") pod \"redhat-marketplace-kwnpj\" (UID: \"1ffd9d0f-1714-416f-84a9-4f3ab0305ee2\") " pod="openshift-marketplace/redhat-marketplace-kwnpj" Jan 21 11:47:09 crc kubenswrapper[4824]: I0121 11:47:09.250709 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1ffd9d0f-1714-416f-84a9-4f3ab0305ee2-catalog-content\") pod \"redhat-marketplace-kwnpj\" (UID: \"1ffd9d0f-1714-416f-84a9-4f3ab0305ee2\") " pod="openshift-marketplace/redhat-marketplace-kwnpj" Jan 21 11:47:09 crc kubenswrapper[4824]: I0121 11:47:09.266598 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-spzf5\" (UniqueName: \"kubernetes.io/projected/1ffd9d0f-1714-416f-84a9-4f3ab0305ee2-kube-api-access-spzf5\") pod \"redhat-marketplace-kwnpj\" (UID: \"1ffd9d0f-1714-416f-84a9-4f3ab0305ee2\") " pod="openshift-marketplace/redhat-marketplace-kwnpj" Jan 21 11:47:09 crc kubenswrapper[4824]: I0121 11:47:09.355160 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kwnpj" Jan 21 11:47:09 crc kubenswrapper[4824]: I0121 11:47:09.753836 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-kwnpj"] Jan 21 11:47:09 crc kubenswrapper[4824]: W0121 11:47:09.755296 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1ffd9d0f_1714_416f_84a9_4f3ab0305ee2.slice/crio-e039f89cd27aa709b422bb17b8c143529a0e6be5ab4250e2f360bbd333a60f34 WatchSource:0}: Error finding container e039f89cd27aa709b422bb17b8c143529a0e6be5ab4250e2f360bbd333a60f34: Status 404 returned error can't find the container with id e039f89cd27aa709b422bb17b8c143529a0e6be5ab4250e2f360bbd333a60f34 Jan 21 11:47:10 crc kubenswrapper[4824]: E0121 11:47:10.004531 4824 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1ffd9d0f_1714_416f_84a9_4f3ab0305ee2.slice/crio-conmon-ae2cc6f1b5d1a60a527caf8bd123c011c61068de03cd0c91932d6344ae5bd4d2.scope\": RecentStats: unable to find data in memory cache]" Jan 21 11:47:10 crc kubenswrapper[4824]: I0121 11:47:10.105035 4824 generic.go:334] "Generic (PLEG): container finished" podID="1ffd9d0f-1714-416f-84a9-4f3ab0305ee2" containerID="ae2cc6f1b5d1a60a527caf8bd123c011c61068de03cd0c91932d6344ae5bd4d2" exitCode=0 Jan 21 11:47:10 crc kubenswrapper[4824]: I0121 11:47:10.105075 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kwnpj" event={"ID":"1ffd9d0f-1714-416f-84a9-4f3ab0305ee2","Type":"ContainerDied","Data":"ae2cc6f1b5d1a60a527caf8bd123c011c61068de03cd0c91932d6344ae5bd4d2"} Jan 21 11:47:10 crc kubenswrapper[4824]: I0121 11:47:10.105101 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kwnpj" event={"ID":"1ffd9d0f-1714-416f-84a9-4f3ab0305ee2","Type":"ContainerStarted","Data":"e039f89cd27aa709b422bb17b8c143529a0e6be5ab4250e2f360bbd333a60f34"} Jan 21 11:47:11 crc kubenswrapper[4824]: I0121 11:47:11.133594 4824 generic.go:334] "Generic (PLEG): container finished" podID="1ffd9d0f-1714-416f-84a9-4f3ab0305ee2" containerID="0c05986916a800695e5243a7d1386a3c32ad00677d5d71b1532637f623250b42" exitCode=0 Jan 21 11:47:11 crc kubenswrapper[4824]: I0121 11:47:11.133635 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kwnpj" event={"ID":"1ffd9d0f-1714-416f-84a9-4f3ab0305ee2","Type":"ContainerDied","Data":"0c05986916a800695e5243a7d1386a3c32ad00677d5d71b1532637f623250b42"} Jan 21 11:47:12 crc kubenswrapper[4824]: I0121 11:47:12.140782 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kwnpj" event={"ID":"1ffd9d0f-1714-416f-84a9-4f3ab0305ee2","Type":"ContainerStarted","Data":"27aa2132a2fd020a218b2edf1fe443e3b9b798fe73cbf970c789812ce59a94f0"} Jan 21 11:47:12 crc kubenswrapper[4824]: I0121 11:47:12.166208 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-kwnpj" podStartSLOduration=1.657175521 podStartE2EDuration="3.166195009s" podCreationTimestamp="2026-01-21 11:47:09 +0000 UTC" firstStartedPulling="2026-01-21 11:47:10.106950281 +0000 UTC m=+2172.399979572" lastFinishedPulling="2026-01-21 11:47:11.615969768 +0000 UTC m=+2173.908999060" observedRunningTime="2026-01-21 11:47:12.152665756 +0000 UTC m=+2174.445695048" watchObservedRunningTime="2026-01-21 11:47:12.166195009 +0000 UTC m=+2174.459224301" Jan 21 11:47:12 crc kubenswrapper[4824]: I0121 11:47:12.817245 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-kkvsf"] Jan 21 11:47:12 crc kubenswrapper[4824]: I0121 11:47:12.820967 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kkvsf" Jan 21 11:47:12 crc kubenswrapper[4824]: I0121 11:47:12.840872 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-kkvsf"] Jan 21 11:47:12 crc kubenswrapper[4824]: I0121 11:47:12.911936 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/06d23466-53be-467a-85af-c4c342b5c86f-utilities\") pod \"community-operators-kkvsf\" (UID: \"06d23466-53be-467a-85af-c4c342b5c86f\") " pod="openshift-marketplace/community-operators-kkvsf" Jan 21 11:47:12 crc kubenswrapper[4824]: I0121 11:47:12.912068 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/06d23466-53be-467a-85af-c4c342b5c86f-catalog-content\") pod \"community-operators-kkvsf\" (UID: \"06d23466-53be-467a-85af-c4c342b5c86f\") " pod="openshift-marketplace/community-operators-kkvsf" Jan 21 11:47:12 crc kubenswrapper[4824]: I0121 11:47:12.912113 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d8cq8\" (UniqueName: \"kubernetes.io/projected/06d23466-53be-467a-85af-c4c342b5c86f-kube-api-access-d8cq8\") pod \"community-operators-kkvsf\" (UID: \"06d23466-53be-467a-85af-c4c342b5c86f\") " pod="openshift-marketplace/community-operators-kkvsf" Jan 21 11:47:13 crc kubenswrapper[4824]: I0121 11:47:13.013616 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/06d23466-53be-467a-85af-c4c342b5c86f-utilities\") pod \"community-operators-kkvsf\" (UID: \"06d23466-53be-467a-85af-c4c342b5c86f\") " pod="openshift-marketplace/community-operators-kkvsf" Jan 21 11:47:13 crc kubenswrapper[4824]: I0121 11:47:13.013702 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/06d23466-53be-467a-85af-c4c342b5c86f-catalog-content\") pod \"community-operators-kkvsf\" (UID: \"06d23466-53be-467a-85af-c4c342b5c86f\") " pod="openshift-marketplace/community-operators-kkvsf" Jan 21 11:47:13 crc kubenswrapper[4824]: I0121 11:47:13.013740 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d8cq8\" (UniqueName: \"kubernetes.io/projected/06d23466-53be-467a-85af-c4c342b5c86f-kube-api-access-d8cq8\") pod \"community-operators-kkvsf\" (UID: \"06d23466-53be-467a-85af-c4c342b5c86f\") " pod="openshift-marketplace/community-operators-kkvsf" Jan 21 11:47:13 crc kubenswrapper[4824]: I0121 11:47:13.014114 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/06d23466-53be-467a-85af-c4c342b5c86f-utilities\") pod \"community-operators-kkvsf\" (UID: \"06d23466-53be-467a-85af-c4c342b5c86f\") " pod="openshift-marketplace/community-operators-kkvsf" Jan 21 11:47:13 crc kubenswrapper[4824]: I0121 11:47:13.014168 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/06d23466-53be-467a-85af-c4c342b5c86f-catalog-content\") pod \"community-operators-kkvsf\" (UID: \"06d23466-53be-467a-85af-c4c342b5c86f\") " pod="openshift-marketplace/community-operators-kkvsf" Jan 21 11:47:13 crc kubenswrapper[4824]: I0121 11:47:13.033646 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d8cq8\" (UniqueName: \"kubernetes.io/projected/06d23466-53be-467a-85af-c4c342b5c86f-kube-api-access-d8cq8\") pod \"community-operators-kkvsf\" (UID: \"06d23466-53be-467a-85af-c4c342b5c86f\") " pod="openshift-marketplace/community-operators-kkvsf" Jan 21 11:47:13 crc kubenswrapper[4824]: I0121 11:47:13.135583 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kkvsf" Jan 21 11:47:13 crc kubenswrapper[4824]: I0121 11:47:13.551888 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-kkvsf"] Jan 21 11:47:14 crc kubenswrapper[4824]: I0121 11:47:14.164768 4824 generic.go:334] "Generic (PLEG): container finished" podID="06d23466-53be-467a-85af-c4c342b5c86f" containerID="062312422cc40765db6b2f1f9a59b726f2a31e61b86f99f4823a9e0e0068bd7b" exitCode=0 Jan 21 11:47:14 crc kubenswrapper[4824]: I0121 11:47:14.164918 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kkvsf" event={"ID":"06d23466-53be-467a-85af-c4c342b5c86f","Type":"ContainerDied","Data":"062312422cc40765db6b2f1f9a59b726f2a31e61b86f99f4823a9e0e0068bd7b"} Jan 21 11:47:14 crc kubenswrapper[4824]: I0121 11:47:14.164999 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kkvsf" event={"ID":"06d23466-53be-467a-85af-c4c342b5c86f","Type":"ContainerStarted","Data":"baa513cc43814d362a686ad77f5a0ae6ae18c0565e4cb439ccbea8853383114f"} Jan 21 11:47:15 crc kubenswrapper[4824]: I0121 11:47:15.173193 4824 generic.go:334] "Generic (PLEG): container finished" podID="06d23466-53be-467a-85af-c4c342b5c86f" containerID="f847d2fa1581eb16fe80b07ba3f7303b1c9d6a7a96d50358553f3ec90e0721b8" exitCode=0 Jan 21 11:47:15 crc kubenswrapper[4824]: I0121 11:47:15.173426 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kkvsf" event={"ID":"06d23466-53be-467a-85af-c4c342b5c86f","Type":"ContainerDied","Data":"f847d2fa1581eb16fe80b07ba3f7303b1c9d6a7a96d50358553f3ec90e0721b8"} Jan 21 11:47:16 crc kubenswrapper[4824]: I0121 11:47:16.065760 4824 patch_prober.go:28] interesting pod/machine-config-daemon-zqjg9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 21 11:47:16 crc kubenswrapper[4824]: I0121 11:47:16.066178 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 21 11:47:16 crc kubenswrapper[4824]: I0121 11:47:16.181730 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kkvsf" event={"ID":"06d23466-53be-467a-85af-c4c342b5c86f","Type":"ContainerStarted","Data":"105d6de379be849d13603aea31763ebdb2481a6f07307d78a44ce1113f54c57a"} Jan 21 11:47:16 crc kubenswrapper[4824]: I0121 11:47:16.194707 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-kkvsf" podStartSLOduration=2.721735921 podStartE2EDuration="4.194693702s" podCreationTimestamp="2026-01-21 11:47:12 +0000 UTC" firstStartedPulling="2026-01-21 11:47:14.166241608 +0000 UTC m=+2176.459270900" lastFinishedPulling="2026-01-21 11:47:15.639199399 +0000 UTC m=+2177.932228681" observedRunningTime="2026-01-21 11:47:16.19405051 +0000 UTC m=+2178.487079802" watchObservedRunningTime="2026-01-21 11:47:16.194693702 +0000 UTC m=+2178.487722994" Jan 21 11:47:19 crc kubenswrapper[4824]: I0121 11:47:19.355713 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-kwnpj" Jan 21 11:47:19 crc kubenswrapper[4824]: I0121 11:47:19.356049 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-kwnpj" Jan 21 11:47:19 crc kubenswrapper[4824]: I0121 11:47:19.389968 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-kwnpj" Jan 21 11:47:20 crc kubenswrapper[4824]: I0121 11:47:20.241184 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-kwnpj" Jan 21 11:47:20 crc kubenswrapper[4824]: I0121 11:47:20.605560 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-kwnpj"] Jan 21 11:47:22 crc kubenswrapper[4824]: I0121 11:47:22.218566 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-kwnpj" podUID="1ffd9d0f-1714-416f-84a9-4f3ab0305ee2" containerName="registry-server" containerID="cri-o://27aa2132a2fd020a218b2edf1fe443e3b9b798fe73cbf970c789812ce59a94f0" gracePeriod=2 Jan 21 11:47:22 crc kubenswrapper[4824]: I0121 11:47:22.567357 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kwnpj" Jan 21 11:47:22 crc kubenswrapper[4824]: I0121 11:47:22.686613 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1ffd9d0f-1714-416f-84a9-4f3ab0305ee2-catalog-content\") pod \"1ffd9d0f-1714-416f-84a9-4f3ab0305ee2\" (UID: \"1ffd9d0f-1714-416f-84a9-4f3ab0305ee2\") " Jan 21 11:47:22 crc kubenswrapper[4824]: I0121 11:47:22.686898 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1ffd9d0f-1714-416f-84a9-4f3ab0305ee2-utilities\") pod \"1ffd9d0f-1714-416f-84a9-4f3ab0305ee2\" (UID: \"1ffd9d0f-1714-416f-84a9-4f3ab0305ee2\") " Jan 21 11:47:22 crc kubenswrapper[4824]: I0121 11:47:22.686935 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-spzf5\" (UniqueName: \"kubernetes.io/projected/1ffd9d0f-1714-416f-84a9-4f3ab0305ee2-kube-api-access-spzf5\") pod \"1ffd9d0f-1714-416f-84a9-4f3ab0305ee2\" (UID: \"1ffd9d0f-1714-416f-84a9-4f3ab0305ee2\") " Jan 21 11:47:22 crc kubenswrapper[4824]: I0121 11:47:22.691781 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1ffd9d0f-1714-416f-84a9-4f3ab0305ee2-utilities" (OuterVolumeSpecName: "utilities") pod "1ffd9d0f-1714-416f-84a9-4f3ab0305ee2" (UID: "1ffd9d0f-1714-416f-84a9-4f3ab0305ee2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:47:22 crc kubenswrapper[4824]: I0121 11:47:22.704361 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1ffd9d0f-1714-416f-84a9-4f3ab0305ee2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1ffd9d0f-1714-416f-84a9-4f3ab0305ee2" (UID: "1ffd9d0f-1714-416f-84a9-4f3ab0305ee2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:47:22 crc kubenswrapper[4824]: I0121 11:47:22.710560 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1ffd9d0f-1714-416f-84a9-4f3ab0305ee2-kube-api-access-spzf5" (OuterVolumeSpecName: "kube-api-access-spzf5") pod "1ffd9d0f-1714-416f-84a9-4f3ab0305ee2" (UID: "1ffd9d0f-1714-416f-84a9-4f3ab0305ee2"). InnerVolumeSpecName "kube-api-access-spzf5". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:47:22 crc kubenswrapper[4824]: I0121 11:47:22.788731 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1ffd9d0f-1714-416f-84a9-4f3ab0305ee2-utilities\") on node \"crc\" DevicePath \"\"" Jan 21 11:47:22 crc kubenswrapper[4824]: I0121 11:47:22.788773 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-spzf5\" (UniqueName: \"kubernetes.io/projected/1ffd9d0f-1714-416f-84a9-4f3ab0305ee2-kube-api-access-spzf5\") on node \"crc\" DevicePath \"\"" Jan 21 11:47:22 crc kubenswrapper[4824]: I0121 11:47:22.788783 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1ffd9d0f-1714-416f-84a9-4f3ab0305ee2-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 21 11:47:23 crc kubenswrapper[4824]: I0121 11:47:23.135850 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-kkvsf" Jan 21 11:47:23 crc kubenswrapper[4824]: I0121 11:47:23.136546 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-kkvsf" Jan 21 11:47:23 crc kubenswrapper[4824]: I0121 11:47:23.169528 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-kkvsf" Jan 21 11:47:23 crc kubenswrapper[4824]: I0121 11:47:23.225524 4824 generic.go:334] "Generic (PLEG): container finished" podID="1ffd9d0f-1714-416f-84a9-4f3ab0305ee2" containerID="27aa2132a2fd020a218b2edf1fe443e3b9b798fe73cbf970c789812ce59a94f0" exitCode=0 Jan 21 11:47:23 crc kubenswrapper[4824]: I0121 11:47:23.225589 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kwnpj" Jan 21 11:47:23 crc kubenswrapper[4824]: I0121 11:47:23.225606 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kwnpj" event={"ID":"1ffd9d0f-1714-416f-84a9-4f3ab0305ee2","Type":"ContainerDied","Data":"27aa2132a2fd020a218b2edf1fe443e3b9b798fe73cbf970c789812ce59a94f0"} Jan 21 11:47:23 crc kubenswrapper[4824]: I0121 11:47:23.225633 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kwnpj" event={"ID":"1ffd9d0f-1714-416f-84a9-4f3ab0305ee2","Type":"ContainerDied","Data":"e039f89cd27aa709b422bb17b8c143529a0e6be5ab4250e2f360bbd333a60f34"} Jan 21 11:47:23 crc kubenswrapper[4824]: I0121 11:47:23.225648 4824 scope.go:117] "RemoveContainer" containerID="27aa2132a2fd020a218b2edf1fe443e3b9b798fe73cbf970c789812ce59a94f0" Jan 21 11:47:23 crc kubenswrapper[4824]: I0121 11:47:23.251085 4824 scope.go:117] "RemoveContainer" containerID="0c05986916a800695e5243a7d1386a3c32ad00677d5d71b1532637f623250b42" Jan 21 11:47:23 crc kubenswrapper[4824]: I0121 11:47:23.252901 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-kwnpj"] Jan 21 11:47:23 crc kubenswrapper[4824]: I0121 11:47:23.260365 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-kwnpj"] Jan 21 11:47:23 crc kubenswrapper[4824]: I0121 11:47:23.265650 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-kkvsf" Jan 21 11:47:23 crc kubenswrapper[4824]: I0121 11:47:23.276797 4824 scope.go:117] "RemoveContainer" containerID="ae2cc6f1b5d1a60a527caf8bd123c011c61068de03cd0c91932d6344ae5bd4d2" Jan 21 11:47:23 crc kubenswrapper[4824]: I0121 11:47:23.303909 4824 scope.go:117] "RemoveContainer" containerID="27aa2132a2fd020a218b2edf1fe443e3b9b798fe73cbf970c789812ce59a94f0" Jan 21 11:47:23 crc kubenswrapper[4824]: E0121 11:47:23.304434 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"27aa2132a2fd020a218b2edf1fe443e3b9b798fe73cbf970c789812ce59a94f0\": container with ID starting with 27aa2132a2fd020a218b2edf1fe443e3b9b798fe73cbf970c789812ce59a94f0 not found: ID does not exist" containerID="27aa2132a2fd020a218b2edf1fe443e3b9b798fe73cbf970c789812ce59a94f0" Jan 21 11:47:23 crc kubenswrapper[4824]: I0121 11:47:23.304460 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"27aa2132a2fd020a218b2edf1fe443e3b9b798fe73cbf970c789812ce59a94f0"} err="failed to get container status \"27aa2132a2fd020a218b2edf1fe443e3b9b798fe73cbf970c789812ce59a94f0\": rpc error: code = NotFound desc = could not find container \"27aa2132a2fd020a218b2edf1fe443e3b9b798fe73cbf970c789812ce59a94f0\": container with ID starting with 27aa2132a2fd020a218b2edf1fe443e3b9b798fe73cbf970c789812ce59a94f0 not found: ID does not exist" Jan 21 11:47:23 crc kubenswrapper[4824]: I0121 11:47:23.304479 4824 scope.go:117] "RemoveContainer" containerID="0c05986916a800695e5243a7d1386a3c32ad00677d5d71b1532637f623250b42" Jan 21 11:47:23 crc kubenswrapper[4824]: E0121 11:47:23.304723 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0c05986916a800695e5243a7d1386a3c32ad00677d5d71b1532637f623250b42\": container with ID starting with 0c05986916a800695e5243a7d1386a3c32ad00677d5d71b1532637f623250b42 not found: ID does not exist" containerID="0c05986916a800695e5243a7d1386a3c32ad00677d5d71b1532637f623250b42" Jan 21 11:47:23 crc kubenswrapper[4824]: I0121 11:47:23.304742 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0c05986916a800695e5243a7d1386a3c32ad00677d5d71b1532637f623250b42"} err="failed to get container status \"0c05986916a800695e5243a7d1386a3c32ad00677d5d71b1532637f623250b42\": rpc error: code = NotFound desc = could not find container \"0c05986916a800695e5243a7d1386a3c32ad00677d5d71b1532637f623250b42\": container with ID starting with 0c05986916a800695e5243a7d1386a3c32ad00677d5d71b1532637f623250b42 not found: ID does not exist" Jan 21 11:47:23 crc kubenswrapper[4824]: I0121 11:47:23.304755 4824 scope.go:117] "RemoveContainer" containerID="ae2cc6f1b5d1a60a527caf8bd123c011c61068de03cd0c91932d6344ae5bd4d2" Jan 21 11:47:23 crc kubenswrapper[4824]: E0121 11:47:23.305065 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ae2cc6f1b5d1a60a527caf8bd123c011c61068de03cd0c91932d6344ae5bd4d2\": container with ID starting with ae2cc6f1b5d1a60a527caf8bd123c011c61068de03cd0c91932d6344ae5bd4d2 not found: ID does not exist" containerID="ae2cc6f1b5d1a60a527caf8bd123c011c61068de03cd0c91932d6344ae5bd4d2" Jan 21 11:47:23 crc kubenswrapper[4824]: I0121 11:47:23.305087 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ae2cc6f1b5d1a60a527caf8bd123c011c61068de03cd0c91932d6344ae5bd4d2"} err="failed to get container status \"ae2cc6f1b5d1a60a527caf8bd123c011c61068de03cd0c91932d6344ae5bd4d2\": rpc error: code = NotFound desc = could not find container \"ae2cc6f1b5d1a60a527caf8bd123c011c61068de03cd0c91932d6344ae5bd4d2\": container with ID starting with ae2cc6f1b5d1a60a527caf8bd123c011c61068de03cd0c91932d6344ae5bd4d2 not found: ID does not exist" Jan 21 11:47:24 crc kubenswrapper[4824]: I0121 11:47:24.058336 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1ffd9d0f-1714-416f-84a9-4f3ab0305ee2" path="/var/lib/kubelet/pods/1ffd9d0f-1714-416f-84a9-4f3ab0305ee2/volumes" Jan 21 11:47:24 crc kubenswrapper[4824]: I0121 11:47:24.608237 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-kkvsf"] Jan 21 11:47:26 crc kubenswrapper[4824]: I0121 11:47:26.244086 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-kkvsf" podUID="06d23466-53be-467a-85af-c4c342b5c86f" containerName="registry-server" containerID="cri-o://105d6de379be849d13603aea31763ebdb2481a6f07307d78a44ce1113f54c57a" gracePeriod=2 Jan 21 11:47:26 crc kubenswrapper[4824]: I0121 11:47:26.596592 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kkvsf" Jan 21 11:47:26 crc kubenswrapper[4824]: I0121 11:47:26.758214 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d8cq8\" (UniqueName: \"kubernetes.io/projected/06d23466-53be-467a-85af-c4c342b5c86f-kube-api-access-d8cq8\") pod \"06d23466-53be-467a-85af-c4c342b5c86f\" (UID: \"06d23466-53be-467a-85af-c4c342b5c86f\") " Jan 21 11:47:26 crc kubenswrapper[4824]: I0121 11:47:26.758341 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/06d23466-53be-467a-85af-c4c342b5c86f-catalog-content\") pod \"06d23466-53be-467a-85af-c4c342b5c86f\" (UID: \"06d23466-53be-467a-85af-c4c342b5c86f\") " Jan 21 11:47:26 crc kubenswrapper[4824]: I0121 11:47:26.758446 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/06d23466-53be-467a-85af-c4c342b5c86f-utilities\") pod \"06d23466-53be-467a-85af-c4c342b5c86f\" (UID: \"06d23466-53be-467a-85af-c4c342b5c86f\") " Jan 21 11:47:26 crc kubenswrapper[4824]: I0121 11:47:26.758997 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/06d23466-53be-467a-85af-c4c342b5c86f-utilities" (OuterVolumeSpecName: "utilities") pod "06d23466-53be-467a-85af-c4c342b5c86f" (UID: "06d23466-53be-467a-85af-c4c342b5c86f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:47:26 crc kubenswrapper[4824]: I0121 11:47:26.763153 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/06d23466-53be-467a-85af-c4c342b5c86f-kube-api-access-d8cq8" (OuterVolumeSpecName: "kube-api-access-d8cq8") pod "06d23466-53be-467a-85af-c4c342b5c86f" (UID: "06d23466-53be-467a-85af-c4c342b5c86f"). InnerVolumeSpecName "kube-api-access-d8cq8". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:47:26 crc kubenswrapper[4824]: I0121 11:47:26.794527 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/06d23466-53be-467a-85af-c4c342b5c86f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "06d23466-53be-467a-85af-c4c342b5c86f" (UID: "06d23466-53be-467a-85af-c4c342b5c86f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:47:26 crc kubenswrapper[4824]: I0121 11:47:26.860380 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/06d23466-53be-467a-85af-c4c342b5c86f-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 21 11:47:26 crc kubenswrapper[4824]: I0121 11:47:26.860404 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/06d23466-53be-467a-85af-c4c342b5c86f-utilities\") on node \"crc\" DevicePath \"\"" Jan 21 11:47:26 crc kubenswrapper[4824]: I0121 11:47:26.860414 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d8cq8\" (UniqueName: \"kubernetes.io/projected/06d23466-53be-467a-85af-c4c342b5c86f-kube-api-access-d8cq8\") on node \"crc\" DevicePath \"\"" Jan 21 11:47:27 crc kubenswrapper[4824]: I0121 11:47:27.252618 4824 generic.go:334] "Generic (PLEG): container finished" podID="06d23466-53be-467a-85af-c4c342b5c86f" containerID="105d6de379be849d13603aea31763ebdb2481a6f07307d78a44ce1113f54c57a" exitCode=0 Jan 21 11:47:27 crc kubenswrapper[4824]: I0121 11:47:27.252658 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kkvsf" Jan 21 11:47:27 crc kubenswrapper[4824]: I0121 11:47:27.252675 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kkvsf" event={"ID":"06d23466-53be-467a-85af-c4c342b5c86f","Type":"ContainerDied","Data":"105d6de379be849d13603aea31763ebdb2481a6f07307d78a44ce1113f54c57a"} Jan 21 11:47:27 crc kubenswrapper[4824]: I0121 11:47:27.252915 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kkvsf" event={"ID":"06d23466-53be-467a-85af-c4c342b5c86f","Type":"ContainerDied","Data":"baa513cc43814d362a686ad77f5a0ae6ae18c0565e4cb439ccbea8853383114f"} Jan 21 11:47:27 crc kubenswrapper[4824]: I0121 11:47:27.252932 4824 scope.go:117] "RemoveContainer" containerID="105d6de379be849d13603aea31763ebdb2481a6f07307d78a44ce1113f54c57a" Jan 21 11:47:27 crc kubenswrapper[4824]: I0121 11:47:27.267388 4824 scope.go:117] "RemoveContainer" containerID="f847d2fa1581eb16fe80b07ba3f7303b1c9d6a7a96d50358553f3ec90e0721b8" Jan 21 11:47:27 crc kubenswrapper[4824]: I0121 11:47:27.278614 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-kkvsf"] Jan 21 11:47:27 crc kubenswrapper[4824]: I0121 11:47:27.284843 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-kkvsf"] Jan 21 11:47:27 crc kubenswrapper[4824]: I0121 11:47:27.296320 4824 scope.go:117] "RemoveContainer" containerID="062312422cc40765db6b2f1f9a59b726f2a31e61b86f99f4823a9e0e0068bd7b" Jan 21 11:47:27 crc kubenswrapper[4824]: I0121 11:47:27.316257 4824 scope.go:117] "RemoveContainer" containerID="105d6de379be849d13603aea31763ebdb2481a6f07307d78a44ce1113f54c57a" Jan 21 11:47:27 crc kubenswrapper[4824]: E0121 11:47:27.316601 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"105d6de379be849d13603aea31763ebdb2481a6f07307d78a44ce1113f54c57a\": container with ID starting with 105d6de379be849d13603aea31763ebdb2481a6f07307d78a44ce1113f54c57a not found: ID does not exist" containerID="105d6de379be849d13603aea31763ebdb2481a6f07307d78a44ce1113f54c57a" Jan 21 11:47:27 crc kubenswrapper[4824]: I0121 11:47:27.316637 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"105d6de379be849d13603aea31763ebdb2481a6f07307d78a44ce1113f54c57a"} err="failed to get container status \"105d6de379be849d13603aea31763ebdb2481a6f07307d78a44ce1113f54c57a\": rpc error: code = NotFound desc = could not find container \"105d6de379be849d13603aea31763ebdb2481a6f07307d78a44ce1113f54c57a\": container with ID starting with 105d6de379be849d13603aea31763ebdb2481a6f07307d78a44ce1113f54c57a not found: ID does not exist" Jan 21 11:47:27 crc kubenswrapper[4824]: I0121 11:47:27.316658 4824 scope.go:117] "RemoveContainer" containerID="f847d2fa1581eb16fe80b07ba3f7303b1c9d6a7a96d50358553f3ec90e0721b8" Jan 21 11:47:27 crc kubenswrapper[4824]: E0121 11:47:27.316988 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f847d2fa1581eb16fe80b07ba3f7303b1c9d6a7a96d50358553f3ec90e0721b8\": container with ID starting with f847d2fa1581eb16fe80b07ba3f7303b1c9d6a7a96d50358553f3ec90e0721b8 not found: ID does not exist" containerID="f847d2fa1581eb16fe80b07ba3f7303b1c9d6a7a96d50358553f3ec90e0721b8" Jan 21 11:47:27 crc kubenswrapper[4824]: I0121 11:47:27.317104 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f847d2fa1581eb16fe80b07ba3f7303b1c9d6a7a96d50358553f3ec90e0721b8"} err="failed to get container status \"f847d2fa1581eb16fe80b07ba3f7303b1c9d6a7a96d50358553f3ec90e0721b8\": rpc error: code = NotFound desc = could not find container \"f847d2fa1581eb16fe80b07ba3f7303b1c9d6a7a96d50358553f3ec90e0721b8\": container with ID starting with f847d2fa1581eb16fe80b07ba3f7303b1c9d6a7a96d50358553f3ec90e0721b8 not found: ID does not exist" Jan 21 11:47:27 crc kubenswrapper[4824]: I0121 11:47:27.317200 4824 scope.go:117] "RemoveContainer" containerID="062312422cc40765db6b2f1f9a59b726f2a31e61b86f99f4823a9e0e0068bd7b" Jan 21 11:47:27 crc kubenswrapper[4824]: E0121 11:47:27.317552 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"062312422cc40765db6b2f1f9a59b726f2a31e61b86f99f4823a9e0e0068bd7b\": container with ID starting with 062312422cc40765db6b2f1f9a59b726f2a31e61b86f99f4823a9e0e0068bd7b not found: ID does not exist" containerID="062312422cc40765db6b2f1f9a59b726f2a31e61b86f99f4823a9e0e0068bd7b" Jan 21 11:47:27 crc kubenswrapper[4824]: I0121 11:47:27.317577 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"062312422cc40765db6b2f1f9a59b726f2a31e61b86f99f4823a9e0e0068bd7b"} err="failed to get container status \"062312422cc40765db6b2f1f9a59b726f2a31e61b86f99f4823a9e0e0068bd7b\": rpc error: code = NotFound desc = could not find container \"062312422cc40765db6b2f1f9a59b726f2a31e61b86f99f4823a9e0e0068bd7b\": container with ID starting with 062312422cc40765db6b2f1f9a59b726f2a31e61b86f99f4823a9e0e0068bd7b not found: ID does not exist" Jan 21 11:47:28 crc kubenswrapper[4824]: I0121 11:47:28.056299 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="06d23466-53be-467a-85af-c4c342b5c86f" path="/var/lib/kubelet/pods/06d23466-53be-467a-85af-c4c342b5c86f/volumes" Jan 21 11:47:33 crc kubenswrapper[4824]: I0121 11:47:33.103059 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tempest-tests-tempest"] Jan 21 11:47:33 crc kubenswrapper[4824]: E0121 11:47:33.103575 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="06d23466-53be-467a-85af-c4c342b5c86f" containerName="extract-content" Jan 21 11:47:33 crc kubenswrapper[4824]: I0121 11:47:33.103588 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="06d23466-53be-467a-85af-c4c342b5c86f" containerName="extract-content" Jan 21 11:47:33 crc kubenswrapper[4824]: E0121 11:47:33.103604 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ffd9d0f-1714-416f-84a9-4f3ab0305ee2" containerName="extract-content" Jan 21 11:47:33 crc kubenswrapper[4824]: I0121 11:47:33.103610 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ffd9d0f-1714-416f-84a9-4f3ab0305ee2" containerName="extract-content" Jan 21 11:47:33 crc kubenswrapper[4824]: E0121 11:47:33.103625 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ffd9d0f-1714-416f-84a9-4f3ab0305ee2" containerName="registry-server" Jan 21 11:47:33 crc kubenswrapper[4824]: I0121 11:47:33.103631 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ffd9d0f-1714-416f-84a9-4f3ab0305ee2" containerName="registry-server" Jan 21 11:47:33 crc kubenswrapper[4824]: E0121 11:47:33.103639 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="06d23466-53be-467a-85af-c4c342b5c86f" containerName="extract-utilities" Jan 21 11:47:33 crc kubenswrapper[4824]: I0121 11:47:33.103645 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="06d23466-53be-467a-85af-c4c342b5c86f" containerName="extract-utilities" Jan 21 11:47:33 crc kubenswrapper[4824]: E0121 11:47:33.103668 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ffd9d0f-1714-416f-84a9-4f3ab0305ee2" containerName="extract-utilities" Jan 21 11:47:33 crc kubenswrapper[4824]: I0121 11:47:33.103673 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ffd9d0f-1714-416f-84a9-4f3ab0305ee2" containerName="extract-utilities" Jan 21 11:47:33 crc kubenswrapper[4824]: E0121 11:47:33.103682 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="06d23466-53be-467a-85af-c4c342b5c86f" containerName="registry-server" Jan 21 11:47:33 crc kubenswrapper[4824]: I0121 11:47:33.103687 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="06d23466-53be-467a-85af-c4c342b5c86f" containerName="registry-server" Jan 21 11:47:33 crc kubenswrapper[4824]: I0121 11:47:33.103848 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="06d23466-53be-467a-85af-c4c342b5c86f" containerName="registry-server" Jan 21 11:47:33 crc kubenswrapper[4824]: I0121 11:47:33.103868 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="1ffd9d0f-1714-416f-84a9-4f3ab0305ee2" containerName="registry-server" Jan 21 11:47:33 crc kubenswrapper[4824]: I0121 11:47:33.104421 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Jan 21 11:47:33 crc kubenswrapper[4824]: I0121 11:47:33.105790 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-qs54z" Jan 21 11:47:33 crc kubenswrapper[4824]: I0121 11:47:33.106017 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"test-operator-controller-priv-key" Jan 21 11:47:33 crc kubenswrapper[4824]: I0121 11:47:33.106035 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-custom-data-s0" Jan 21 11:47:33 crc kubenswrapper[4824]: I0121 11:47:33.106055 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Jan 21 11:47:33 crc kubenswrapper[4824]: I0121 11:47:33.109575 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Jan 21 11:47:33 crc kubenswrapper[4824]: I0121 11:47:33.263743 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dr6pd\" (UniqueName: \"kubernetes.io/projected/58d30951-ee56-4217-a614-a04951369c8b-kube-api-access-dr6pd\") pod \"tempest-tests-tempest\" (UID: \"58d30951-ee56-4217-a614-a04951369c8b\") " pod="openstack/tempest-tests-tempest" Jan 21 11:47:33 crc kubenswrapper[4824]: I0121 11:47:33.263816 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/58d30951-ee56-4217-a614-a04951369c8b-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"58d30951-ee56-4217-a614-a04951369c8b\") " pod="openstack/tempest-tests-tempest" Jan 21 11:47:33 crc kubenswrapper[4824]: I0121 11:47:33.264096 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/58d30951-ee56-4217-a614-a04951369c8b-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"58d30951-ee56-4217-a614-a04951369c8b\") " pod="openstack/tempest-tests-tempest" Jan 21 11:47:33 crc kubenswrapper[4824]: I0121 11:47:33.264165 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"tempest-tests-tempest\" (UID: \"58d30951-ee56-4217-a614-a04951369c8b\") " pod="openstack/tempest-tests-tempest" Jan 21 11:47:33 crc kubenswrapper[4824]: I0121 11:47:33.264229 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/58d30951-ee56-4217-a614-a04951369c8b-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"58d30951-ee56-4217-a614-a04951369c8b\") " pod="openstack/tempest-tests-tempest" Jan 21 11:47:33 crc kubenswrapper[4824]: I0121 11:47:33.264259 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/58d30951-ee56-4217-a614-a04951369c8b-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"58d30951-ee56-4217-a614-a04951369c8b\") " pod="openstack/tempest-tests-tempest" Jan 21 11:47:33 crc kubenswrapper[4824]: I0121 11:47:33.264303 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/58d30951-ee56-4217-a614-a04951369c8b-config-data\") pod \"tempest-tests-tempest\" (UID: \"58d30951-ee56-4217-a614-a04951369c8b\") " pod="openstack/tempest-tests-tempest" Jan 21 11:47:33 crc kubenswrapper[4824]: I0121 11:47:33.264336 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/58d30951-ee56-4217-a614-a04951369c8b-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"58d30951-ee56-4217-a614-a04951369c8b\") " pod="openstack/tempest-tests-tempest" Jan 21 11:47:33 crc kubenswrapper[4824]: I0121 11:47:33.264385 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/58d30951-ee56-4217-a614-a04951369c8b-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"58d30951-ee56-4217-a614-a04951369c8b\") " pod="openstack/tempest-tests-tempest" Jan 21 11:47:33 crc kubenswrapper[4824]: I0121 11:47:33.365975 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/58d30951-ee56-4217-a614-a04951369c8b-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"58d30951-ee56-4217-a614-a04951369c8b\") " pod="openstack/tempest-tests-tempest" Jan 21 11:47:33 crc kubenswrapper[4824]: I0121 11:47:33.366018 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"tempest-tests-tempest\" (UID: \"58d30951-ee56-4217-a614-a04951369c8b\") " pod="openstack/tempest-tests-tempest" Jan 21 11:47:33 crc kubenswrapper[4824]: I0121 11:47:33.366061 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/58d30951-ee56-4217-a614-a04951369c8b-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"58d30951-ee56-4217-a614-a04951369c8b\") " pod="openstack/tempest-tests-tempest" Jan 21 11:47:33 crc kubenswrapper[4824]: I0121 11:47:33.366079 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/58d30951-ee56-4217-a614-a04951369c8b-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"58d30951-ee56-4217-a614-a04951369c8b\") " pod="openstack/tempest-tests-tempest" Jan 21 11:47:33 crc kubenswrapper[4824]: I0121 11:47:33.366123 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/58d30951-ee56-4217-a614-a04951369c8b-config-data\") pod \"tempest-tests-tempest\" (UID: \"58d30951-ee56-4217-a614-a04951369c8b\") " pod="openstack/tempest-tests-tempest" Jan 21 11:47:33 crc kubenswrapper[4824]: I0121 11:47:33.366148 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/58d30951-ee56-4217-a614-a04951369c8b-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"58d30951-ee56-4217-a614-a04951369c8b\") " pod="openstack/tempest-tests-tempest" Jan 21 11:47:33 crc kubenswrapper[4824]: I0121 11:47:33.366887 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/58d30951-ee56-4217-a614-a04951369c8b-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"58d30951-ee56-4217-a614-a04951369c8b\") " pod="openstack/tempest-tests-tempest" Jan 21 11:47:33 crc kubenswrapper[4824]: I0121 11:47:33.366344 4824 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"tempest-tests-tempest\" (UID: \"58d30951-ee56-4217-a614-a04951369c8b\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/tempest-tests-tempest" Jan 21 11:47:33 crc kubenswrapper[4824]: I0121 11:47:33.366551 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/58d30951-ee56-4217-a614-a04951369c8b-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"58d30951-ee56-4217-a614-a04951369c8b\") " pod="openstack/tempest-tests-tempest" Jan 21 11:47:33 crc kubenswrapper[4824]: I0121 11:47:33.366461 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/58d30951-ee56-4217-a614-a04951369c8b-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"58d30951-ee56-4217-a614-a04951369c8b\") " pod="openstack/tempest-tests-tempest" Jan 21 11:47:33 crc kubenswrapper[4824]: I0121 11:47:33.367197 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dr6pd\" (UniqueName: \"kubernetes.io/projected/58d30951-ee56-4217-a614-a04951369c8b-kube-api-access-dr6pd\") pod \"tempest-tests-tempest\" (UID: \"58d30951-ee56-4217-a614-a04951369c8b\") " pod="openstack/tempest-tests-tempest" Jan 21 11:47:33 crc kubenswrapper[4824]: I0121 11:47:33.367265 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/58d30951-ee56-4217-a614-a04951369c8b-config-data\") pod \"tempest-tests-tempest\" (UID: \"58d30951-ee56-4217-a614-a04951369c8b\") " pod="openstack/tempest-tests-tempest" Jan 21 11:47:33 crc kubenswrapper[4824]: I0121 11:47:33.367309 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/58d30951-ee56-4217-a614-a04951369c8b-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"58d30951-ee56-4217-a614-a04951369c8b\") " pod="openstack/tempest-tests-tempest" Jan 21 11:47:33 crc kubenswrapper[4824]: I0121 11:47:33.367839 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/58d30951-ee56-4217-a614-a04951369c8b-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"58d30951-ee56-4217-a614-a04951369c8b\") " pod="openstack/tempest-tests-tempest" Jan 21 11:47:33 crc kubenswrapper[4824]: I0121 11:47:33.371821 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/58d30951-ee56-4217-a614-a04951369c8b-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"58d30951-ee56-4217-a614-a04951369c8b\") " pod="openstack/tempest-tests-tempest" Jan 21 11:47:33 crc kubenswrapper[4824]: I0121 11:47:33.372033 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/58d30951-ee56-4217-a614-a04951369c8b-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"58d30951-ee56-4217-a614-a04951369c8b\") " pod="openstack/tempest-tests-tempest" Jan 21 11:47:33 crc kubenswrapper[4824]: I0121 11:47:33.372150 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/58d30951-ee56-4217-a614-a04951369c8b-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"58d30951-ee56-4217-a614-a04951369c8b\") " pod="openstack/tempest-tests-tempest" Jan 21 11:47:33 crc kubenswrapper[4824]: I0121 11:47:33.383921 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dr6pd\" (UniqueName: \"kubernetes.io/projected/58d30951-ee56-4217-a614-a04951369c8b-kube-api-access-dr6pd\") pod \"tempest-tests-tempest\" (UID: \"58d30951-ee56-4217-a614-a04951369c8b\") " pod="openstack/tempest-tests-tempest" Jan 21 11:47:33 crc kubenswrapper[4824]: I0121 11:47:33.385945 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"tempest-tests-tempest\" (UID: \"58d30951-ee56-4217-a614-a04951369c8b\") " pod="openstack/tempest-tests-tempest" Jan 21 11:47:33 crc kubenswrapper[4824]: I0121 11:47:33.426556 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Jan 21 11:47:33 crc kubenswrapper[4824]: I0121 11:47:33.846055 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Jan 21 11:47:34 crc kubenswrapper[4824]: I0121 11:47:34.296532 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"58d30951-ee56-4217-a614-a04951369c8b","Type":"ContainerStarted","Data":"5bc5adacec79286d2293a6c562788392bcdca8f811b129442be3e2105c5568bf"} Jan 21 11:47:46 crc kubenswrapper[4824]: I0121 11:47:46.065544 4824 patch_prober.go:28] interesting pod/machine-config-daemon-zqjg9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 21 11:47:46 crc kubenswrapper[4824]: I0121 11:47:46.065905 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 21 11:47:46 crc kubenswrapper[4824]: I0121 11:47:46.065940 4824 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" Jan 21 11:47:46 crc kubenswrapper[4824]: I0121 11:47:46.066375 4824 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"9e4a3d10f6183509a66349b3f877d2ea1f0b26ebbfb7ecfc617c73f022a56788"} pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 21 11:47:46 crc kubenswrapper[4824]: I0121 11:47:46.066422 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" containerName="machine-config-daemon" containerID="cri-o://9e4a3d10f6183509a66349b3f877d2ea1f0b26ebbfb7ecfc617c73f022a56788" gracePeriod=600 Jan 21 11:47:46 crc kubenswrapper[4824]: I0121 11:47:46.383245 4824 generic.go:334] "Generic (PLEG): container finished" podID="33f3d922-4ffe-409b-a49a-d88c85898260" containerID="9e4a3d10f6183509a66349b3f877d2ea1f0b26ebbfb7ecfc617c73f022a56788" exitCode=0 Jan 21 11:47:46 crc kubenswrapper[4824]: I0121 11:47:46.383290 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" event={"ID":"33f3d922-4ffe-409b-a49a-d88c85898260","Type":"ContainerDied","Data":"9e4a3d10f6183509a66349b3f877d2ea1f0b26ebbfb7ecfc617c73f022a56788"} Jan 21 11:47:46 crc kubenswrapper[4824]: I0121 11:47:46.383336 4824 scope.go:117] "RemoveContainer" containerID="1283af2f214718d7c6e80663aef792afcaa88b12e9e64dbc3ab357a33b0df07b" Jan 21 11:47:54 crc kubenswrapper[4824]: E0121 11:47:54.241843 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 11:47:54 crc kubenswrapper[4824]: E0121 11:47:54.321064 4824 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified" Jan 21 11:47:54 crc kubenswrapper[4824]: E0121 11:47:54.321435 4824 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:tempest-tests-tempest-tests-runner,Image:quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/test_operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-workdir,ReadOnly:false,MountPath:/var/lib/tempest,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-temporary,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-logs,ReadOnly:false,MountPath:/var/lib/tempest/external_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/etc/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/var/lib/tempest/.config/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/etc/openstack/secure.yaml,SubPath:secure.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ca-certs,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ssh-key,ReadOnly:false,MountPath:/var/lib/tempest/id_ecdsa,SubPath:ssh_key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-dr6pd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42480,RunAsNonRoot:*false,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:*true,RunAsGroup:*42480,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-custom-data-s0,},Optional:nil,},SecretRef:nil,},EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-env-vars-s0,},Optional:nil,},SecretRef:nil,},},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod tempest-tests-tempest_openstack(58d30951-ee56-4217-a614-a04951369c8b): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Jan 21 11:47:54 crc kubenswrapper[4824]: E0121 11:47:54.323100 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/tempest-tests-tempest" podUID="58d30951-ee56-4217-a614-a04951369c8b" Jan 21 11:47:54 crc kubenswrapper[4824]: I0121 11:47:54.444025 4824 scope.go:117] "RemoveContainer" containerID="9e4a3d10f6183509a66349b3f877d2ea1f0b26ebbfb7ecfc617c73f022a56788" Jan 21 11:47:54 crc kubenswrapper[4824]: E0121 11:47:54.444362 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 11:47:54 crc kubenswrapper[4824]: E0121 11:47:54.444641 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified\\\"\"" pod="openstack/tempest-tests-tempest" podUID="58d30951-ee56-4217-a614-a04951369c8b" Jan 21 11:48:08 crc kubenswrapper[4824]: I0121 11:48:08.053365 4824 scope.go:117] "RemoveContainer" containerID="9e4a3d10f6183509a66349b3f877d2ea1f0b26ebbfb7ecfc617c73f022a56788" Jan 21 11:48:08 crc kubenswrapper[4824]: E0121 11:48:08.054056 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 11:48:09 crc kubenswrapper[4824]: I0121 11:48:09.569255 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Jan 21 11:48:10 crc kubenswrapper[4824]: I0121 11:48:10.553519 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"58d30951-ee56-4217-a614-a04951369c8b","Type":"ContainerStarted","Data":"f56ac3b5212e3537571012922916244691168864ad9d36d1f4cde9480e97ed7a"} Jan 21 11:48:10 crc kubenswrapper[4824]: I0121 11:48:10.575737 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tempest-tests-tempest" podStartSLOduration=2.8575966729999998 podStartE2EDuration="38.575717158s" podCreationTimestamp="2026-01-21 11:47:32 +0000 UTC" firstStartedPulling="2026-01-21 11:47:33.848737983 +0000 UTC m=+2196.141767275" lastFinishedPulling="2026-01-21 11:48:09.566858467 +0000 UTC m=+2231.859887760" observedRunningTime="2026-01-21 11:48:10.568096411 +0000 UTC m=+2232.861125713" watchObservedRunningTime="2026-01-21 11:48:10.575717158 +0000 UTC m=+2232.868746449" Jan 21 11:48:23 crc kubenswrapper[4824]: I0121 11:48:23.049603 4824 scope.go:117] "RemoveContainer" containerID="9e4a3d10f6183509a66349b3f877d2ea1f0b26ebbfb7ecfc617c73f022a56788" Jan 21 11:48:23 crc kubenswrapper[4824]: E0121 11:48:23.050192 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 11:48:38 crc kubenswrapper[4824]: I0121 11:48:38.049999 4824 scope.go:117] "RemoveContainer" containerID="9e4a3d10f6183509a66349b3f877d2ea1f0b26ebbfb7ecfc617c73f022a56788" Jan 21 11:48:38 crc kubenswrapper[4824]: E0121 11:48:38.051009 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 11:48:51 crc kubenswrapper[4824]: I0121 11:48:51.049873 4824 scope.go:117] "RemoveContainer" containerID="9e4a3d10f6183509a66349b3f877d2ea1f0b26ebbfb7ecfc617c73f022a56788" Jan 21 11:48:51 crc kubenswrapper[4824]: E0121 11:48:51.051468 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 11:49:05 crc kubenswrapper[4824]: I0121 11:49:05.050016 4824 scope.go:117] "RemoveContainer" containerID="9e4a3d10f6183509a66349b3f877d2ea1f0b26ebbfb7ecfc617c73f022a56788" Jan 21 11:49:05 crc kubenswrapper[4824]: E0121 11:49:05.050731 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 11:49:18 crc kubenswrapper[4824]: I0121 11:49:18.053759 4824 scope.go:117] "RemoveContainer" containerID="9e4a3d10f6183509a66349b3f877d2ea1f0b26ebbfb7ecfc617c73f022a56788" Jan 21 11:49:18 crc kubenswrapper[4824]: E0121 11:49:18.054310 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 11:49:32 crc kubenswrapper[4824]: I0121 11:49:32.053014 4824 scope.go:117] "RemoveContainer" containerID="9e4a3d10f6183509a66349b3f877d2ea1f0b26ebbfb7ecfc617c73f022a56788" Jan 21 11:49:32 crc kubenswrapper[4824]: E0121 11:49:32.053768 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 11:49:43 crc kubenswrapper[4824]: I0121 11:49:43.049785 4824 scope.go:117] "RemoveContainer" containerID="9e4a3d10f6183509a66349b3f877d2ea1f0b26ebbfb7ecfc617c73f022a56788" Jan 21 11:49:43 crc kubenswrapper[4824]: E0121 11:49:43.050446 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 11:49:57 crc kubenswrapper[4824]: I0121 11:49:57.049223 4824 scope.go:117] "RemoveContainer" containerID="9e4a3d10f6183509a66349b3f877d2ea1f0b26ebbfb7ecfc617c73f022a56788" Jan 21 11:49:57 crc kubenswrapper[4824]: E0121 11:49:57.049801 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 11:50:12 crc kubenswrapper[4824]: I0121 11:50:12.051546 4824 scope.go:117] "RemoveContainer" containerID="9e4a3d10f6183509a66349b3f877d2ea1f0b26ebbfb7ecfc617c73f022a56788" Jan 21 11:50:12 crc kubenswrapper[4824]: E0121 11:50:12.052049 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 11:50:26 crc kubenswrapper[4824]: I0121 11:50:26.049548 4824 scope.go:117] "RemoveContainer" containerID="9e4a3d10f6183509a66349b3f877d2ea1f0b26ebbfb7ecfc617c73f022a56788" Jan 21 11:50:26 crc kubenswrapper[4824]: E0121 11:50:26.050258 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 11:50:38 crc kubenswrapper[4824]: I0121 11:50:38.053760 4824 scope.go:117] "RemoveContainer" containerID="9e4a3d10f6183509a66349b3f877d2ea1f0b26ebbfb7ecfc617c73f022a56788" Jan 21 11:50:38 crc kubenswrapper[4824]: E0121 11:50:38.054549 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 11:50:49 crc kubenswrapper[4824]: I0121 11:50:49.049268 4824 scope.go:117] "RemoveContainer" containerID="9e4a3d10f6183509a66349b3f877d2ea1f0b26ebbfb7ecfc617c73f022a56788" Jan 21 11:50:49 crc kubenswrapper[4824]: E0121 11:50:49.049819 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 11:51:02 crc kubenswrapper[4824]: I0121 11:51:02.050508 4824 scope.go:117] "RemoveContainer" containerID="9e4a3d10f6183509a66349b3f877d2ea1f0b26ebbfb7ecfc617c73f022a56788" Jan 21 11:51:02 crc kubenswrapper[4824]: E0121 11:51:02.051503 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 11:51:15 crc kubenswrapper[4824]: I0121 11:51:15.049793 4824 scope.go:117] "RemoveContainer" containerID="9e4a3d10f6183509a66349b3f877d2ea1f0b26ebbfb7ecfc617c73f022a56788" Jan 21 11:51:15 crc kubenswrapper[4824]: E0121 11:51:15.050401 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 11:51:28 crc kubenswrapper[4824]: I0121 11:51:28.053787 4824 scope.go:117] "RemoveContainer" containerID="9e4a3d10f6183509a66349b3f877d2ea1f0b26ebbfb7ecfc617c73f022a56788" Jan 21 11:51:28 crc kubenswrapper[4824]: E0121 11:51:28.054512 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 11:51:40 crc kubenswrapper[4824]: I0121 11:51:40.049410 4824 scope.go:117] "RemoveContainer" containerID="9e4a3d10f6183509a66349b3f877d2ea1f0b26ebbfb7ecfc617c73f022a56788" Jan 21 11:51:40 crc kubenswrapper[4824]: E0121 11:51:40.050079 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 11:51:52 crc kubenswrapper[4824]: I0121 11:51:52.049780 4824 scope.go:117] "RemoveContainer" containerID="9e4a3d10f6183509a66349b3f877d2ea1f0b26ebbfb7ecfc617c73f022a56788" Jan 21 11:51:52 crc kubenswrapper[4824]: E0121 11:51:52.050515 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 11:52:06 crc kubenswrapper[4824]: I0121 11:52:06.049735 4824 scope.go:117] "RemoveContainer" containerID="9e4a3d10f6183509a66349b3f877d2ea1f0b26ebbfb7ecfc617c73f022a56788" Jan 21 11:52:06 crc kubenswrapper[4824]: E0121 11:52:06.050643 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 11:52:19 crc kubenswrapper[4824]: I0121 11:52:19.049638 4824 scope.go:117] "RemoveContainer" containerID="9e4a3d10f6183509a66349b3f877d2ea1f0b26ebbfb7ecfc617c73f022a56788" Jan 21 11:52:19 crc kubenswrapper[4824]: E0121 11:52:19.050211 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 11:52:33 crc kubenswrapper[4824]: I0121 11:52:33.050144 4824 scope.go:117] "RemoveContainer" containerID="9e4a3d10f6183509a66349b3f877d2ea1f0b26ebbfb7ecfc617c73f022a56788" Jan 21 11:52:33 crc kubenswrapper[4824]: E0121 11:52:33.050686 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 11:52:35 crc kubenswrapper[4824]: I0121 11:52:35.618342 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-94jm9"] Jan 21 11:52:35 crc kubenswrapper[4824]: I0121 11:52:35.620093 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-94jm9" Jan 21 11:52:35 crc kubenswrapper[4824]: I0121 11:52:35.639412 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-94jm9"] Jan 21 11:52:35 crc kubenswrapper[4824]: I0121 11:52:35.704323 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/17ee7b33-8d12-49da-b45b-5b8d75e53a08-utilities\") pod \"certified-operators-94jm9\" (UID: \"17ee7b33-8d12-49da-b45b-5b8d75e53a08\") " pod="openshift-marketplace/certified-operators-94jm9" Jan 21 11:52:35 crc kubenswrapper[4824]: I0121 11:52:35.704492 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-swv92\" (UniqueName: \"kubernetes.io/projected/17ee7b33-8d12-49da-b45b-5b8d75e53a08-kube-api-access-swv92\") pod \"certified-operators-94jm9\" (UID: \"17ee7b33-8d12-49da-b45b-5b8d75e53a08\") " pod="openshift-marketplace/certified-operators-94jm9" Jan 21 11:52:35 crc kubenswrapper[4824]: I0121 11:52:35.704525 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/17ee7b33-8d12-49da-b45b-5b8d75e53a08-catalog-content\") pod \"certified-operators-94jm9\" (UID: \"17ee7b33-8d12-49da-b45b-5b8d75e53a08\") " pod="openshift-marketplace/certified-operators-94jm9" Jan 21 11:52:35 crc kubenswrapper[4824]: I0121 11:52:35.805664 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-swv92\" (UniqueName: \"kubernetes.io/projected/17ee7b33-8d12-49da-b45b-5b8d75e53a08-kube-api-access-swv92\") pod \"certified-operators-94jm9\" (UID: \"17ee7b33-8d12-49da-b45b-5b8d75e53a08\") " pod="openshift-marketplace/certified-operators-94jm9" Jan 21 11:52:35 crc kubenswrapper[4824]: I0121 11:52:35.805703 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/17ee7b33-8d12-49da-b45b-5b8d75e53a08-catalog-content\") pod \"certified-operators-94jm9\" (UID: \"17ee7b33-8d12-49da-b45b-5b8d75e53a08\") " pod="openshift-marketplace/certified-operators-94jm9" Jan 21 11:52:35 crc kubenswrapper[4824]: I0121 11:52:35.805877 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/17ee7b33-8d12-49da-b45b-5b8d75e53a08-utilities\") pod \"certified-operators-94jm9\" (UID: \"17ee7b33-8d12-49da-b45b-5b8d75e53a08\") " pod="openshift-marketplace/certified-operators-94jm9" Jan 21 11:52:35 crc kubenswrapper[4824]: I0121 11:52:35.806284 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/17ee7b33-8d12-49da-b45b-5b8d75e53a08-utilities\") pod \"certified-operators-94jm9\" (UID: \"17ee7b33-8d12-49da-b45b-5b8d75e53a08\") " pod="openshift-marketplace/certified-operators-94jm9" Jan 21 11:52:35 crc kubenswrapper[4824]: I0121 11:52:35.806340 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/17ee7b33-8d12-49da-b45b-5b8d75e53a08-catalog-content\") pod \"certified-operators-94jm9\" (UID: \"17ee7b33-8d12-49da-b45b-5b8d75e53a08\") " pod="openshift-marketplace/certified-operators-94jm9" Jan 21 11:52:35 crc kubenswrapper[4824]: I0121 11:52:35.824913 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-swv92\" (UniqueName: \"kubernetes.io/projected/17ee7b33-8d12-49da-b45b-5b8d75e53a08-kube-api-access-swv92\") pod \"certified-operators-94jm9\" (UID: \"17ee7b33-8d12-49da-b45b-5b8d75e53a08\") " pod="openshift-marketplace/certified-operators-94jm9" Jan 21 11:52:35 crc kubenswrapper[4824]: I0121 11:52:35.937309 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-94jm9" Jan 21 11:52:36 crc kubenswrapper[4824]: I0121 11:52:36.492817 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-94jm9"] Jan 21 11:52:37 crc kubenswrapper[4824]: I0121 11:52:37.284111 4824 generic.go:334] "Generic (PLEG): container finished" podID="17ee7b33-8d12-49da-b45b-5b8d75e53a08" containerID="5c3c56a19c488932d9fadfedab7a0d81e527d9aebbbb79aa6acb1aa4e01539ce" exitCode=0 Jan 21 11:52:37 crc kubenswrapper[4824]: I0121 11:52:37.284167 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-94jm9" event={"ID":"17ee7b33-8d12-49da-b45b-5b8d75e53a08","Type":"ContainerDied","Data":"5c3c56a19c488932d9fadfedab7a0d81e527d9aebbbb79aa6acb1aa4e01539ce"} Jan 21 11:52:37 crc kubenswrapper[4824]: I0121 11:52:37.284437 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-94jm9" event={"ID":"17ee7b33-8d12-49da-b45b-5b8d75e53a08","Type":"ContainerStarted","Data":"2f09f6b203e5c39ff281636d181c83fa7016bbf854baf158e8e90be501cf2e21"} Jan 21 11:52:37 crc kubenswrapper[4824]: I0121 11:52:37.285879 4824 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 21 11:52:38 crc kubenswrapper[4824]: I0121 11:52:38.293990 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-94jm9" event={"ID":"17ee7b33-8d12-49da-b45b-5b8d75e53a08","Type":"ContainerStarted","Data":"6a124101d97c8ef54c3a6d69ef834719e5219560c16dae90a8bb36ea450fa85e"} Jan 21 11:52:38 crc kubenswrapper[4824]: I0121 11:52:38.817321 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-f9255"] Jan 21 11:52:38 crc kubenswrapper[4824]: I0121 11:52:38.819113 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-f9255" Jan 21 11:52:38 crc kubenswrapper[4824]: I0121 11:52:38.839801 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-f9255"] Jan 21 11:52:38 crc kubenswrapper[4824]: I0121 11:52:38.861168 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tntsc\" (UniqueName: \"kubernetes.io/projected/f557c9a4-cd4b-467b-8d42-43c9f6e2d6f0-kube-api-access-tntsc\") pod \"redhat-operators-f9255\" (UID: \"f557c9a4-cd4b-467b-8d42-43c9f6e2d6f0\") " pod="openshift-marketplace/redhat-operators-f9255" Jan 21 11:52:38 crc kubenswrapper[4824]: I0121 11:52:38.861221 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f557c9a4-cd4b-467b-8d42-43c9f6e2d6f0-catalog-content\") pod \"redhat-operators-f9255\" (UID: \"f557c9a4-cd4b-467b-8d42-43c9f6e2d6f0\") " pod="openshift-marketplace/redhat-operators-f9255" Jan 21 11:52:38 crc kubenswrapper[4824]: I0121 11:52:38.861266 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f557c9a4-cd4b-467b-8d42-43c9f6e2d6f0-utilities\") pod \"redhat-operators-f9255\" (UID: \"f557c9a4-cd4b-467b-8d42-43c9f6e2d6f0\") " pod="openshift-marketplace/redhat-operators-f9255" Jan 21 11:52:38 crc kubenswrapper[4824]: I0121 11:52:38.963399 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tntsc\" (UniqueName: \"kubernetes.io/projected/f557c9a4-cd4b-467b-8d42-43c9f6e2d6f0-kube-api-access-tntsc\") pod \"redhat-operators-f9255\" (UID: \"f557c9a4-cd4b-467b-8d42-43c9f6e2d6f0\") " pod="openshift-marketplace/redhat-operators-f9255" Jan 21 11:52:38 crc kubenswrapper[4824]: I0121 11:52:38.963463 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f557c9a4-cd4b-467b-8d42-43c9f6e2d6f0-catalog-content\") pod \"redhat-operators-f9255\" (UID: \"f557c9a4-cd4b-467b-8d42-43c9f6e2d6f0\") " pod="openshift-marketplace/redhat-operators-f9255" Jan 21 11:52:38 crc kubenswrapper[4824]: I0121 11:52:38.963511 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f557c9a4-cd4b-467b-8d42-43c9f6e2d6f0-utilities\") pod \"redhat-operators-f9255\" (UID: \"f557c9a4-cd4b-467b-8d42-43c9f6e2d6f0\") " pod="openshift-marketplace/redhat-operators-f9255" Jan 21 11:52:38 crc kubenswrapper[4824]: I0121 11:52:38.964069 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f557c9a4-cd4b-467b-8d42-43c9f6e2d6f0-catalog-content\") pod \"redhat-operators-f9255\" (UID: \"f557c9a4-cd4b-467b-8d42-43c9f6e2d6f0\") " pod="openshift-marketplace/redhat-operators-f9255" Jan 21 11:52:38 crc kubenswrapper[4824]: I0121 11:52:38.964199 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f557c9a4-cd4b-467b-8d42-43c9f6e2d6f0-utilities\") pod \"redhat-operators-f9255\" (UID: \"f557c9a4-cd4b-467b-8d42-43c9f6e2d6f0\") " pod="openshift-marketplace/redhat-operators-f9255" Jan 21 11:52:38 crc kubenswrapper[4824]: I0121 11:52:38.985654 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tntsc\" (UniqueName: \"kubernetes.io/projected/f557c9a4-cd4b-467b-8d42-43c9f6e2d6f0-kube-api-access-tntsc\") pod \"redhat-operators-f9255\" (UID: \"f557c9a4-cd4b-467b-8d42-43c9f6e2d6f0\") " pod="openshift-marketplace/redhat-operators-f9255" Jan 21 11:52:39 crc kubenswrapper[4824]: I0121 11:52:39.141089 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-f9255" Jan 21 11:52:39 crc kubenswrapper[4824]: I0121 11:52:39.305643 4824 generic.go:334] "Generic (PLEG): container finished" podID="17ee7b33-8d12-49da-b45b-5b8d75e53a08" containerID="6a124101d97c8ef54c3a6d69ef834719e5219560c16dae90a8bb36ea450fa85e" exitCode=0 Jan 21 11:52:39 crc kubenswrapper[4824]: I0121 11:52:39.305748 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-94jm9" event={"ID":"17ee7b33-8d12-49da-b45b-5b8d75e53a08","Type":"ContainerDied","Data":"6a124101d97c8ef54c3a6d69ef834719e5219560c16dae90a8bb36ea450fa85e"} Jan 21 11:52:39 crc kubenswrapper[4824]: I0121 11:52:39.538266 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-f9255"] Jan 21 11:52:39 crc kubenswrapper[4824]: W0121 11:52:39.543769 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf557c9a4_cd4b_467b_8d42_43c9f6e2d6f0.slice/crio-749644413b97fec9cd23377ee2def7d6bcb9e73627dcf6518eaa3d3461f70acc WatchSource:0}: Error finding container 749644413b97fec9cd23377ee2def7d6bcb9e73627dcf6518eaa3d3461f70acc: Status 404 returned error can't find the container with id 749644413b97fec9cd23377ee2def7d6bcb9e73627dcf6518eaa3d3461f70acc Jan 21 11:52:40 crc kubenswrapper[4824]: I0121 11:52:40.313831 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-94jm9" event={"ID":"17ee7b33-8d12-49da-b45b-5b8d75e53a08","Type":"ContainerStarted","Data":"8b81e86fec0a01e32005faeaaae2b7652b7a81ce26fc7c8165d23642ab228d97"} Jan 21 11:52:40 crc kubenswrapper[4824]: I0121 11:52:40.315485 4824 generic.go:334] "Generic (PLEG): container finished" podID="f557c9a4-cd4b-467b-8d42-43c9f6e2d6f0" containerID="578ed9d39acbed22739419a347c18af6900aa1843784a13bd7014c78781a3ee9" exitCode=0 Jan 21 11:52:40 crc kubenswrapper[4824]: I0121 11:52:40.315522 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f9255" event={"ID":"f557c9a4-cd4b-467b-8d42-43c9f6e2d6f0","Type":"ContainerDied","Data":"578ed9d39acbed22739419a347c18af6900aa1843784a13bd7014c78781a3ee9"} Jan 21 11:52:40 crc kubenswrapper[4824]: I0121 11:52:40.315567 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f9255" event={"ID":"f557c9a4-cd4b-467b-8d42-43c9f6e2d6f0","Type":"ContainerStarted","Data":"749644413b97fec9cd23377ee2def7d6bcb9e73627dcf6518eaa3d3461f70acc"} Jan 21 11:52:40 crc kubenswrapper[4824]: I0121 11:52:40.335450 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-94jm9" podStartSLOduration=2.717146636 podStartE2EDuration="5.335437768s" podCreationTimestamp="2026-01-21 11:52:35 +0000 UTC" firstStartedPulling="2026-01-21 11:52:37.285679871 +0000 UTC m=+2499.578709163" lastFinishedPulling="2026-01-21 11:52:39.903971003 +0000 UTC m=+2502.197000295" observedRunningTime="2026-01-21 11:52:40.329934012 +0000 UTC m=+2502.622963305" watchObservedRunningTime="2026-01-21 11:52:40.335437768 +0000 UTC m=+2502.628467060" Jan 21 11:52:41 crc kubenswrapper[4824]: I0121 11:52:41.322823 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f9255" event={"ID":"f557c9a4-cd4b-467b-8d42-43c9f6e2d6f0","Type":"ContainerStarted","Data":"068161b300ed84c74e43ec69a803a06b692f64d2fbd3a1946d30310c5be5a5e3"} Jan 21 11:52:44 crc kubenswrapper[4824]: I0121 11:52:44.050378 4824 scope.go:117] "RemoveContainer" containerID="9e4a3d10f6183509a66349b3f877d2ea1f0b26ebbfb7ecfc617c73f022a56788" Jan 21 11:52:44 crc kubenswrapper[4824]: E0121 11:52:44.050584 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 11:52:44 crc kubenswrapper[4824]: I0121 11:52:44.341833 4824 generic.go:334] "Generic (PLEG): container finished" podID="f557c9a4-cd4b-467b-8d42-43c9f6e2d6f0" containerID="068161b300ed84c74e43ec69a803a06b692f64d2fbd3a1946d30310c5be5a5e3" exitCode=0 Jan 21 11:52:44 crc kubenswrapper[4824]: I0121 11:52:44.341923 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f9255" event={"ID":"f557c9a4-cd4b-467b-8d42-43c9f6e2d6f0","Type":"ContainerDied","Data":"068161b300ed84c74e43ec69a803a06b692f64d2fbd3a1946d30310c5be5a5e3"} Jan 21 11:52:45 crc kubenswrapper[4824]: I0121 11:52:45.357027 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f9255" event={"ID":"f557c9a4-cd4b-467b-8d42-43c9f6e2d6f0","Type":"ContainerStarted","Data":"e4dd9042ff54f92650b0bbe09c7e3ad51142fdbcb72213bc41f7f2b332907c9b"} Jan 21 11:52:45 crc kubenswrapper[4824]: I0121 11:52:45.377785 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-f9255" podStartSLOduration=2.867260868 podStartE2EDuration="7.377768786s" podCreationTimestamp="2026-01-21 11:52:38 +0000 UTC" firstStartedPulling="2026-01-21 11:52:40.317144439 +0000 UTC m=+2502.610173731" lastFinishedPulling="2026-01-21 11:52:44.827652357 +0000 UTC m=+2507.120681649" observedRunningTime="2026-01-21 11:52:45.371680447 +0000 UTC m=+2507.664709739" watchObservedRunningTime="2026-01-21 11:52:45.377768786 +0000 UTC m=+2507.670798067" Jan 21 11:52:45 crc kubenswrapper[4824]: I0121 11:52:45.937674 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-94jm9" Jan 21 11:52:45 crc kubenswrapper[4824]: I0121 11:52:45.937741 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-94jm9" Jan 21 11:52:45 crc kubenswrapper[4824]: I0121 11:52:45.972225 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-94jm9" Jan 21 11:52:46 crc kubenswrapper[4824]: I0121 11:52:46.395281 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-94jm9" Jan 21 11:52:47 crc kubenswrapper[4824]: I0121 11:52:47.007399 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-94jm9"] Jan 21 11:52:48 crc kubenswrapper[4824]: I0121 11:52:48.374165 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-94jm9" podUID="17ee7b33-8d12-49da-b45b-5b8d75e53a08" containerName="registry-server" containerID="cri-o://8b81e86fec0a01e32005faeaaae2b7652b7a81ce26fc7c8165d23642ab228d97" gracePeriod=2 Jan 21 11:52:49 crc kubenswrapper[4824]: I0121 11:52:49.141915 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-f9255" Jan 21 11:52:49 crc kubenswrapper[4824]: I0121 11:52:49.142068 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-f9255" Jan 21 11:52:49 crc kubenswrapper[4824]: I0121 11:52:49.309681 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-94jm9" Jan 21 11:52:49 crc kubenswrapper[4824]: I0121 11:52:49.350824 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/17ee7b33-8d12-49da-b45b-5b8d75e53a08-catalog-content\") pod \"17ee7b33-8d12-49da-b45b-5b8d75e53a08\" (UID: \"17ee7b33-8d12-49da-b45b-5b8d75e53a08\") " Jan 21 11:52:49 crc kubenswrapper[4824]: I0121 11:52:49.350875 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/17ee7b33-8d12-49da-b45b-5b8d75e53a08-utilities\") pod \"17ee7b33-8d12-49da-b45b-5b8d75e53a08\" (UID: \"17ee7b33-8d12-49da-b45b-5b8d75e53a08\") " Jan 21 11:52:49 crc kubenswrapper[4824]: I0121 11:52:49.351028 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-swv92\" (UniqueName: \"kubernetes.io/projected/17ee7b33-8d12-49da-b45b-5b8d75e53a08-kube-api-access-swv92\") pod \"17ee7b33-8d12-49da-b45b-5b8d75e53a08\" (UID: \"17ee7b33-8d12-49da-b45b-5b8d75e53a08\") " Jan 21 11:52:49 crc kubenswrapper[4824]: I0121 11:52:49.351403 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/17ee7b33-8d12-49da-b45b-5b8d75e53a08-utilities" (OuterVolumeSpecName: "utilities") pod "17ee7b33-8d12-49da-b45b-5b8d75e53a08" (UID: "17ee7b33-8d12-49da-b45b-5b8d75e53a08"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:52:49 crc kubenswrapper[4824]: I0121 11:52:49.351783 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/17ee7b33-8d12-49da-b45b-5b8d75e53a08-utilities\") on node \"crc\" DevicePath \"\"" Jan 21 11:52:49 crc kubenswrapper[4824]: I0121 11:52:49.355634 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/17ee7b33-8d12-49da-b45b-5b8d75e53a08-kube-api-access-swv92" (OuterVolumeSpecName: "kube-api-access-swv92") pod "17ee7b33-8d12-49da-b45b-5b8d75e53a08" (UID: "17ee7b33-8d12-49da-b45b-5b8d75e53a08"). InnerVolumeSpecName "kube-api-access-swv92". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:52:49 crc kubenswrapper[4824]: I0121 11:52:49.382199 4824 generic.go:334] "Generic (PLEG): container finished" podID="17ee7b33-8d12-49da-b45b-5b8d75e53a08" containerID="8b81e86fec0a01e32005faeaaae2b7652b7a81ce26fc7c8165d23642ab228d97" exitCode=0 Jan 21 11:52:49 crc kubenswrapper[4824]: I0121 11:52:49.382261 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-94jm9" Jan 21 11:52:49 crc kubenswrapper[4824]: I0121 11:52:49.382264 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-94jm9" event={"ID":"17ee7b33-8d12-49da-b45b-5b8d75e53a08","Type":"ContainerDied","Data":"8b81e86fec0a01e32005faeaaae2b7652b7a81ce26fc7c8165d23642ab228d97"} Jan 21 11:52:49 crc kubenswrapper[4824]: I0121 11:52:49.382931 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-94jm9" event={"ID":"17ee7b33-8d12-49da-b45b-5b8d75e53a08","Type":"ContainerDied","Data":"2f09f6b203e5c39ff281636d181c83fa7016bbf854baf158e8e90be501cf2e21"} Jan 21 11:52:49 crc kubenswrapper[4824]: I0121 11:52:49.382952 4824 scope.go:117] "RemoveContainer" containerID="8b81e86fec0a01e32005faeaaae2b7652b7a81ce26fc7c8165d23642ab228d97" Jan 21 11:52:49 crc kubenswrapper[4824]: I0121 11:52:49.386494 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/17ee7b33-8d12-49da-b45b-5b8d75e53a08-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "17ee7b33-8d12-49da-b45b-5b8d75e53a08" (UID: "17ee7b33-8d12-49da-b45b-5b8d75e53a08"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:52:49 crc kubenswrapper[4824]: I0121 11:52:49.399493 4824 scope.go:117] "RemoveContainer" containerID="6a124101d97c8ef54c3a6d69ef834719e5219560c16dae90a8bb36ea450fa85e" Jan 21 11:52:49 crc kubenswrapper[4824]: I0121 11:52:49.416985 4824 scope.go:117] "RemoveContainer" containerID="5c3c56a19c488932d9fadfedab7a0d81e527d9aebbbb79aa6acb1aa4e01539ce" Jan 21 11:52:49 crc kubenswrapper[4824]: I0121 11:52:49.446474 4824 scope.go:117] "RemoveContainer" containerID="8b81e86fec0a01e32005faeaaae2b7652b7a81ce26fc7c8165d23642ab228d97" Jan 21 11:52:49 crc kubenswrapper[4824]: E0121 11:52:49.446763 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8b81e86fec0a01e32005faeaaae2b7652b7a81ce26fc7c8165d23642ab228d97\": container with ID starting with 8b81e86fec0a01e32005faeaaae2b7652b7a81ce26fc7c8165d23642ab228d97 not found: ID does not exist" containerID="8b81e86fec0a01e32005faeaaae2b7652b7a81ce26fc7c8165d23642ab228d97" Jan 21 11:52:49 crc kubenswrapper[4824]: I0121 11:52:49.446792 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8b81e86fec0a01e32005faeaaae2b7652b7a81ce26fc7c8165d23642ab228d97"} err="failed to get container status \"8b81e86fec0a01e32005faeaaae2b7652b7a81ce26fc7c8165d23642ab228d97\": rpc error: code = NotFound desc = could not find container \"8b81e86fec0a01e32005faeaaae2b7652b7a81ce26fc7c8165d23642ab228d97\": container with ID starting with 8b81e86fec0a01e32005faeaaae2b7652b7a81ce26fc7c8165d23642ab228d97 not found: ID does not exist" Jan 21 11:52:49 crc kubenswrapper[4824]: I0121 11:52:49.446811 4824 scope.go:117] "RemoveContainer" containerID="6a124101d97c8ef54c3a6d69ef834719e5219560c16dae90a8bb36ea450fa85e" Jan 21 11:52:49 crc kubenswrapper[4824]: E0121 11:52:49.447063 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6a124101d97c8ef54c3a6d69ef834719e5219560c16dae90a8bb36ea450fa85e\": container with ID starting with 6a124101d97c8ef54c3a6d69ef834719e5219560c16dae90a8bb36ea450fa85e not found: ID does not exist" containerID="6a124101d97c8ef54c3a6d69ef834719e5219560c16dae90a8bb36ea450fa85e" Jan 21 11:52:49 crc kubenswrapper[4824]: I0121 11:52:49.447099 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6a124101d97c8ef54c3a6d69ef834719e5219560c16dae90a8bb36ea450fa85e"} err="failed to get container status \"6a124101d97c8ef54c3a6d69ef834719e5219560c16dae90a8bb36ea450fa85e\": rpc error: code = NotFound desc = could not find container \"6a124101d97c8ef54c3a6d69ef834719e5219560c16dae90a8bb36ea450fa85e\": container with ID starting with 6a124101d97c8ef54c3a6d69ef834719e5219560c16dae90a8bb36ea450fa85e not found: ID does not exist" Jan 21 11:52:49 crc kubenswrapper[4824]: I0121 11:52:49.447121 4824 scope.go:117] "RemoveContainer" containerID="5c3c56a19c488932d9fadfedab7a0d81e527d9aebbbb79aa6acb1aa4e01539ce" Jan 21 11:52:49 crc kubenswrapper[4824]: E0121 11:52:49.447403 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5c3c56a19c488932d9fadfedab7a0d81e527d9aebbbb79aa6acb1aa4e01539ce\": container with ID starting with 5c3c56a19c488932d9fadfedab7a0d81e527d9aebbbb79aa6acb1aa4e01539ce not found: ID does not exist" containerID="5c3c56a19c488932d9fadfedab7a0d81e527d9aebbbb79aa6acb1aa4e01539ce" Jan 21 11:52:49 crc kubenswrapper[4824]: I0121 11:52:49.447426 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5c3c56a19c488932d9fadfedab7a0d81e527d9aebbbb79aa6acb1aa4e01539ce"} err="failed to get container status \"5c3c56a19c488932d9fadfedab7a0d81e527d9aebbbb79aa6acb1aa4e01539ce\": rpc error: code = NotFound desc = could not find container \"5c3c56a19c488932d9fadfedab7a0d81e527d9aebbbb79aa6acb1aa4e01539ce\": container with ID starting with 5c3c56a19c488932d9fadfedab7a0d81e527d9aebbbb79aa6acb1aa4e01539ce not found: ID does not exist" Jan 21 11:52:49 crc kubenswrapper[4824]: I0121 11:52:49.453612 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-swv92\" (UniqueName: \"kubernetes.io/projected/17ee7b33-8d12-49da-b45b-5b8d75e53a08-kube-api-access-swv92\") on node \"crc\" DevicePath \"\"" Jan 21 11:52:49 crc kubenswrapper[4824]: I0121 11:52:49.453640 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/17ee7b33-8d12-49da-b45b-5b8d75e53a08-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 21 11:52:49 crc kubenswrapper[4824]: I0121 11:52:49.708022 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-94jm9"] Jan 21 11:52:49 crc kubenswrapper[4824]: I0121 11:52:49.713971 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-94jm9"] Jan 21 11:52:50 crc kubenswrapper[4824]: I0121 11:52:50.056943 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="17ee7b33-8d12-49da-b45b-5b8d75e53a08" path="/var/lib/kubelet/pods/17ee7b33-8d12-49da-b45b-5b8d75e53a08/volumes" Jan 21 11:52:50 crc kubenswrapper[4824]: I0121 11:52:50.177259 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-f9255" podUID="f557c9a4-cd4b-467b-8d42-43c9f6e2d6f0" containerName="registry-server" probeResult="failure" output=< Jan 21 11:52:50 crc kubenswrapper[4824]: timeout: failed to connect service ":50051" within 1s Jan 21 11:52:50 crc kubenswrapper[4824]: > Jan 21 11:52:58 crc kubenswrapper[4824]: I0121 11:52:58.054311 4824 scope.go:117] "RemoveContainer" containerID="9e4a3d10f6183509a66349b3f877d2ea1f0b26ebbfb7ecfc617c73f022a56788" Jan 21 11:52:58 crc kubenswrapper[4824]: I0121 11:52:58.445942 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" event={"ID":"33f3d922-4ffe-409b-a49a-d88c85898260","Type":"ContainerStarted","Data":"54c16394c863edb71506776097207cb6de1af7378f2e1570362c90829b6a2501"} Jan 21 11:52:59 crc kubenswrapper[4824]: I0121 11:52:59.174211 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-f9255" Jan 21 11:52:59 crc kubenswrapper[4824]: I0121 11:52:59.206558 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-f9255" Jan 21 11:52:59 crc kubenswrapper[4824]: I0121 11:52:59.403129 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-f9255"] Jan 21 11:53:00 crc kubenswrapper[4824]: I0121 11:53:00.457272 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-f9255" podUID="f557c9a4-cd4b-467b-8d42-43c9f6e2d6f0" containerName="registry-server" containerID="cri-o://e4dd9042ff54f92650b0bbe09c7e3ad51142fdbcb72213bc41f7f2b332907c9b" gracePeriod=2 Jan 21 11:53:00 crc kubenswrapper[4824]: I0121 11:53:00.835098 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-f9255" Jan 21 11:53:00 crc kubenswrapper[4824]: I0121 11:53:00.948819 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f557c9a4-cd4b-467b-8d42-43c9f6e2d6f0-catalog-content\") pod \"f557c9a4-cd4b-467b-8d42-43c9f6e2d6f0\" (UID: \"f557c9a4-cd4b-467b-8d42-43c9f6e2d6f0\") " Jan 21 11:53:00 crc kubenswrapper[4824]: I0121 11:53:00.948876 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tntsc\" (UniqueName: \"kubernetes.io/projected/f557c9a4-cd4b-467b-8d42-43c9f6e2d6f0-kube-api-access-tntsc\") pod \"f557c9a4-cd4b-467b-8d42-43c9f6e2d6f0\" (UID: \"f557c9a4-cd4b-467b-8d42-43c9f6e2d6f0\") " Jan 21 11:53:00 crc kubenswrapper[4824]: I0121 11:53:00.948995 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f557c9a4-cd4b-467b-8d42-43c9f6e2d6f0-utilities\") pod \"f557c9a4-cd4b-467b-8d42-43c9f6e2d6f0\" (UID: \"f557c9a4-cd4b-467b-8d42-43c9f6e2d6f0\") " Jan 21 11:53:00 crc kubenswrapper[4824]: I0121 11:53:00.949579 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f557c9a4-cd4b-467b-8d42-43c9f6e2d6f0-utilities" (OuterVolumeSpecName: "utilities") pod "f557c9a4-cd4b-467b-8d42-43c9f6e2d6f0" (UID: "f557c9a4-cd4b-467b-8d42-43c9f6e2d6f0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:53:00 crc kubenswrapper[4824]: I0121 11:53:00.955323 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f557c9a4-cd4b-467b-8d42-43c9f6e2d6f0-kube-api-access-tntsc" (OuterVolumeSpecName: "kube-api-access-tntsc") pod "f557c9a4-cd4b-467b-8d42-43c9f6e2d6f0" (UID: "f557c9a4-cd4b-467b-8d42-43c9f6e2d6f0"). InnerVolumeSpecName "kube-api-access-tntsc". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:53:01 crc kubenswrapper[4824]: I0121 11:53:01.044543 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f557c9a4-cd4b-467b-8d42-43c9f6e2d6f0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f557c9a4-cd4b-467b-8d42-43c9f6e2d6f0" (UID: "f557c9a4-cd4b-467b-8d42-43c9f6e2d6f0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:53:01 crc kubenswrapper[4824]: I0121 11:53:01.050937 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f557c9a4-cd4b-467b-8d42-43c9f6e2d6f0-utilities\") on node \"crc\" DevicePath \"\"" Jan 21 11:53:01 crc kubenswrapper[4824]: I0121 11:53:01.050981 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f557c9a4-cd4b-467b-8d42-43c9f6e2d6f0-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 21 11:53:01 crc kubenswrapper[4824]: I0121 11:53:01.050992 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tntsc\" (UniqueName: \"kubernetes.io/projected/f557c9a4-cd4b-467b-8d42-43c9f6e2d6f0-kube-api-access-tntsc\") on node \"crc\" DevicePath \"\"" Jan 21 11:53:01 crc kubenswrapper[4824]: I0121 11:53:01.466342 4824 generic.go:334] "Generic (PLEG): container finished" podID="f557c9a4-cd4b-467b-8d42-43c9f6e2d6f0" containerID="e4dd9042ff54f92650b0bbe09c7e3ad51142fdbcb72213bc41f7f2b332907c9b" exitCode=0 Jan 21 11:53:01 crc kubenswrapper[4824]: I0121 11:53:01.466439 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-f9255" Jan 21 11:53:01 crc kubenswrapper[4824]: I0121 11:53:01.466463 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f9255" event={"ID":"f557c9a4-cd4b-467b-8d42-43c9f6e2d6f0","Type":"ContainerDied","Data":"e4dd9042ff54f92650b0bbe09c7e3ad51142fdbcb72213bc41f7f2b332907c9b"} Jan 21 11:53:01 crc kubenswrapper[4824]: I0121 11:53:01.467420 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f9255" event={"ID":"f557c9a4-cd4b-467b-8d42-43c9f6e2d6f0","Type":"ContainerDied","Data":"749644413b97fec9cd23377ee2def7d6bcb9e73627dcf6518eaa3d3461f70acc"} Jan 21 11:53:01 crc kubenswrapper[4824]: I0121 11:53:01.467463 4824 scope.go:117] "RemoveContainer" containerID="e4dd9042ff54f92650b0bbe09c7e3ad51142fdbcb72213bc41f7f2b332907c9b" Jan 21 11:53:01 crc kubenswrapper[4824]: I0121 11:53:01.488048 4824 scope.go:117] "RemoveContainer" containerID="068161b300ed84c74e43ec69a803a06b692f64d2fbd3a1946d30310c5be5a5e3" Jan 21 11:53:01 crc kubenswrapper[4824]: I0121 11:53:01.495208 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-f9255"] Jan 21 11:53:01 crc kubenswrapper[4824]: I0121 11:53:01.500193 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-f9255"] Jan 21 11:53:01 crc kubenswrapper[4824]: I0121 11:53:01.520038 4824 scope.go:117] "RemoveContainer" containerID="578ed9d39acbed22739419a347c18af6900aa1843784a13bd7014c78781a3ee9" Jan 21 11:53:01 crc kubenswrapper[4824]: I0121 11:53:01.535609 4824 scope.go:117] "RemoveContainer" containerID="e4dd9042ff54f92650b0bbe09c7e3ad51142fdbcb72213bc41f7f2b332907c9b" Jan 21 11:53:01 crc kubenswrapper[4824]: E0121 11:53:01.535973 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e4dd9042ff54f92650b0bbe09c7e3ad51142fdbcb72213bc41f7f2b332907c9b\": container with ID starting with e4dd9042ff54f92650b0bbe09c7e3ad51142fdbcb72213bc41f7f2b332907c9b not found: ID does not exist" containerID="e4dd9042ff54f92650b0bbe09c7e3ad51142fdbcb72213bc41f7f2b332907c9b" Jan 21 11:53:01 crc kubenswrapper[4824]: I0121 11:53:01.536009 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e4dd9042ff54f92650b0bbe09c7e3ad51142fdbcb72213bc41f7f2b332907c9b"} err="failed to get container status \"e4dd9042ff54f92650b0bbe09c7e3ad51142fdbcb72213bc41f7f2b332907c9b\": rpc error: code = NotFound desc = could not find container \"e4dd9042ff54f92650b0bbe09c7e3ad51142fdbcb72213bc41f7f2b332907c9b\": container with ID starting with e4dd9042ff54f92650b0bbe09c7e3ad51142fdbcb72213bc41f7f2b332907c9b not found: ID does not exist" Jan 21 11:53:01 crc kubenswrapper[4824]: I0121 11:53:01.536033 4824 scope.go:117] "RemoveContainer" containerID="068161b300ed84c74e43ec69a803a06b692f64d2fbd3a1946d30310c5be5a5e3" Jan 21 11:53:01 crc kubenswrapper[4824]: E0121 11:53:01.536311 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"068161b300ed84c74e43ec69a803a06b692f64d2fbd3a1946d30310c5be5a5e3\": container with ID starting with 068161b300ed84c74e43ec69a803a06b692f64d2fbd3a1946d30310c5be5a5e3 not found: ID does not exist" containerID="068161b300ed84c74e43ec69a803a06b692f64d2fbd3a1946d30310c5be5a5e3" Jan 21 11:53:01 crc kubenswrapper[4824]: I0121 11:53:01.536338 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"068161b300ed84c74e43ec69a803a06b692f64d2fbd3a1946d30310c5be5a5e3"} err="failed to get container status \"068161b300ed84c74e43ec69a803a06b692f64d2fbd3a1946d30310c5be5a5e3\": rpc error: code = NotFound desc = could not find container \"068161b300ed84c74e43ec69a803a06b692f64d2fbd3a1946d30310c5be5a5e3\": container with ID starting with 068161b300ed84c74e43ec69a803a06b692f64d2fbd3a1946d30310c5be5a5e3 not found: ID does not exist" Jan 21 11:53:01 crc kubenswrapper[4824]: I0121 11:53:01.536358 4824 scope.go:117] "RemoveContainer" containerID="578ed9d39acbed22739419a347c18af6900aa1843784a13bd7014c78781a3ee9" Jan 21 11:53:01 crc kubenswrapper[4824]: E0121 11:53:01.536597 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"578ed9d39acbed22739419a347c18af6900aa1843784a13bd7014c78781a3ee9\": container with ID starting with 578ed9d39acbed22739419a347c18af6900aa1843784a13bd7014c78781a3ee9 not found: ID does not exist" containerID="578ed9d39acbed22739419a347c18af6900aa1843784a13bd7014c78781a3ee9" Jan 21 11:53:01 crc kubenswrapper[4824]: I0121 11:53:01.536630 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"578ed9d39acbed22739419a347c18af6900aa1843784a13bd7014c78781a3ee9"} err="failed to get container status \"578ed9d39acbed22739419a347c18af6900aa1843784a13bd7014c78781a3ee9\": rpc error: code = NotFound desc = could not find container \"578ed9d39acbed22739419a347c18af6900aa1843784a13bd7014c78781a3ee9\": container with ID starting with 578ed9d39acbed22739419a347c18af6900aa1843784a13bd7014c78781a3ee9 not found: ID does not exist" Jan 21 11:53:02 crc kubenswrapper[4824]: I0121 11:53:02.057385 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f557c9a4-cd4b-467b-8d42-43c9f6e2d6f0" path="/var/lib/kubelet/pods/f557c9a4-cd4b-467b-8d42-43c9f6e2d6f0/volumes" Jan 21 11:55:16 crc kubenswrapper[4824]: I0121 11:55:16.065233 4824 patch_prober.go:28] interesting pod/machine-config-daemon-zqjg9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 21 11:55:16 crc kubenswrapper[4824]: I0121 11:55:16.065654 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 21 11:55:46 crc kubenswrapper[4824]: I0121 11:55:46.065145 4824 patch_prober.go:28] interesting pod/machine-config-daemon-zqjg9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 21 11:55:46 crc kubenswrapper[4824]: I0121 11:55:46.065514 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 21 11:56:16 crc kubenswrapper[4824]: I0121 11:56:16.065216 4824 patch_prober.go:28] interesting pod/machine-config-daemon-zqjg9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 21 11:56:16 crc kubenswrapper[4824]: I0121 11:56:16.065612 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 21 11:56:16 crc kubenswrapper[4824]: I0121 11:56:16.065644 4824 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" Jan 21 11:56:16 crc kubenswrapper[4824]: I0121 11:56:16.066018 4824 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"54c16394c863edb71506776097207cb6de1af7378f2e1570362c90829b6a2501"} pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 21 11:56:16 crc kubenswrapper[4824]: I0121 11:56:16.066057 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" containerName="machine-config-daemon" containerID="cri-o://54c16394c863edb71506776097207cb6de1af7378f2e1570362c90829b6a2501" gracePeriod=600 Jan 21 11:56:16 crc kubenswrapper[4824]: I0121 11:56:16.694012 4824 generic.go:334] "Generic (PLEG): container finished" podID="33f3d922-4ffe-409b-a49a-d88c85898260" containerID="54c16394c863edb71506776097207cb6de1af7378f2e1570362c90829b6a2501" exitCode=0 Jan 21 11:56:16 crc kubenswrapper[4824]: I0121 11:56:16.694094 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" event={"ID":"33f3d922-4ffe-409b-a49a-d88c85898260","Type":"ContainerDied","Data":"54c16394c863edb71506776097207cb6de1af7378f2e1570362c90829b6a2501"} Jan 21 11:56:16 crc kubenswrapper[4824]: I0121 11:56:16.694366 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" event={"ID":"33f3d922-4ffe-409b-a49a-d88c85898260","Type":"ContainerStarted","Data":"753267d7f07800b4c7939c46da56af707c5831b8615d9de85460b5d273fc051a"} Jan 21 11:56:16 crc kubenswrapper[4824]: I0121 11:56:16.694386 4824 scope.go:117] "RemoveContainer" containerID="9e4a3d10f6183509a66349b3f877d2ea1f0b26ebbfb7ecfc617c73f022a56788" Jan 21 11:57:43 crc kubenswrapper[4824]: I0121 11:57:43.889321 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-f9xtx"] Jan 21 11:57:43 crc kubenswrapper[4824]: E0121 11:57:43.890002 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17ee7b33-8d12-49da-b45b-5b8d75e53a08" containerName="extract-utilities" Jan 21 11:57:43 crc kubenswrapper[4824]: I0121 11:57:43.890015 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="17ee7b33-8d12-49da-b45b-5b8d75e53a08" containerName="extract-utilities" Jan 21 11:57:43 crc kubenswrapper[4824]: E0121 11:57:43.890025 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f557c9a4-cd4b-467b-8d42-43c9f6e2d6f0" containerName="extract-content" Jan 21 11:57:43 crc kubenswrapper[4824]: I0121 11:57:43.890031 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="f557c9a4-cd4b-467b-8d42-43c9f6e2d6f0" containerName="extract-content" Jan 21 11:57:43 crc kubenswrapper[4824]: E0121 11:57:43.890041 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f557c9a4-cd4b-467b-8d42-43c9f6e2d6f0" containerName="registry-server" Jan 21 11:57:43 crc kubenswrapper[4824]: I0121 11:57:43.890046 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="f557c9a4-cd4b-467b-8d42-43c9f6e2d6f0" containerName="registry-server" Jan 21 11:57:43 crc kubenswrapper[4824]: E0121 11:57:43.890077 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f557c9a4-cd4b-467b-8d42-43c9f6e2d6f0" containerName="extract-utilities" Jan 21 11:57:43 crc kubenswrapper[4824]: I0121 11:57:43.890083 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="f557c9a4-cd4b-467b-8d42-43c9f6e2d6f0" containerName="extract-utilities" Jan 21 11:57:43 crc kubenswrapper[4824]: E0121 11:57:43.890095 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17ee7b33-8d12-49da-b45b-5b8d75e53a08" containerName="extract-content" Jan 21 11:57:43 crc kubenswrapper[4824]: I0121 11:57:43.890101 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="17ee7b33-8d12-49da-b45b-5b8d75e53a08" containerName="extract-content" Jan 21 11:57:43 crc kubenswrapper[4824]: E0121 11:57:43.890116 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17ee7b33-8d12-49da-b45b-5b8d75e53a08" containerName="registry-server" Jan 21 11:57:43 crc kubenswrapper[4824]: I0121 11:57:43.890122 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="17ee7b33-8d12-49da-b45b-5b8d75e53a08" containerName="registry-server" Jan 21 11:57:43 crc kubenswrapper[4824]: I0121 11:57:43.890306 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="17ee7b33-8d12-49da-b45b-5b8d75e53a08" containerName="registry-server" Jan 21 11:57:43 crc kubenswrapper[4824]: I0121 11:57:43.890327 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="f557c9a4-cd4b-467b-8d42-43c9f6e2d6f0" containerName="registry-server" Jan 21 11:57:43 crc kubenswrapper[4824]: I0121 11:57:43.891414 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-f9xtx" Jan 21 11:57:43 crc kubenswrapper[4824]: I0121 11:57:43.897281 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-f9xtx"] Jan 21 11:57:44 crc kubenswrapper[4824]: I0121 11:57:44.085865 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/95ce4f3d-0b58-42b2-9930-23952f0be687-utilities\") pod \"community-operators-f9xtx\" (UID: \"95ce4f3d-0b58-42b2-9930-23952f0be687\") " pod="openshift-marketplace/community-operators-f9xtx" Jan 21 11:57:44 crc kubenswrapper[4824]: I0121 11:57:44.085913 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9cfdg\" (UniqueName: \"kubernetes.io/projected/95ce4f3d-0b58-42b2-9930-23952f0be687-kube-api-access-9cfdg\") pod \"community-operators-f9xtx\" (UID: \"95ce4f3d-0b58-42b2-9930-23952f0be687\") " pod="openshift-marketplace/community-operators-f9xtx" Jan 21 11:57:44 crc kubenswrapper[4824]: I0121 11:57:44.086153 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/95ce4f3d-0b58-42b2-9930-23952f0be687-catalog-content\") pod \"community-operators-f9xtx\" (UID: \"95ce4f3d-0b58-42b2-9930-23952f0be687\") " pod="openshift-marketplace/community-operators-f9xtx" Jan 21 11:57:44 crc kubenswrapper[4824]: I0121 11:57:44.188331 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/95ce4f3d-0b58-42b2-9930-23952f0be687-utilities\") pod \"community-operators-f9xtx\" (UID: \"95ce4f3d-0b58-42b2-9930-23952f0be687\") " pod="openshift-marketplace/community-operators-f9xtx" Jan 21 11:57:44 crc kubenswrapper[4824]: I0121 11:57:44.188377 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9cfdg\" (UniqueName: \"kubernetes.io/projected/95ce4f3d-0b58-42b2-9930-23952f0be687-kube-api-access-9cfdg\") pod \"community-operators-f9xtx\" (UID: \"95ce4f3d-0b58-42b2-9930-23952f0be687\") " pod="openshift-marketplace/community-operators-f9xtx" Jan 21 11:57:44 crc kubenswrapper[4824]: I0121 11:57:44.188591 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/95ce4f3d-0b58-42b2-9930-23952f0be687-catalog-content\") pod \"community-operators-f9xtx\" (UID: \"95ce4f3d-0b58-42b2-9930-23952f0be687\") " pod="openshift-marketplace/community-operators-f9xtx" Jan 21 11:57:44 crc kubenswrapper[4824]: I0121 11:57:44.188791 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/95ce4f3d-0b58-42b2-9930-23952f0be687-utilities\") pod \"community-operators-f9xtx\" (UID: \"95ce4f3d-0b58-42b2-9930-23952f0be687\") " pod="openshift-marketplace/community-operators-f9xtx" Jan 21 11:57:44 crc kubenswrapper[4824]: I0121 11:57:44.188990 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/95ce4f3d-0b58-42b2-9930-23952f0be687-catalog-content\") pod \"community-operators-f9xtx\" (UID: \"95ce4f3d-0b58-42b2-9930-23952f0be687\") " pod="openshift-marketplace/community-operators-f9xtx" Jan 21 11:57:44 crc kubenswrapper[4824]: I0121 11:57:44.205552 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9cfdg\" (UniqueName: \"kubernetes.io/projected/95ce4f3d-0b58-42b2-9930-23952f0be687-kube-api-access-9cfdg\") pod \"community-operators-f9xtx\" (UID: \"95ce4f3d-0b58-42b2-9930-23952f0be687\") " pod="openshift-marketplace/community-operators-f9xtx" Jan 21 11:57:44 crc kubenswrapper[4824]: I0121 11:57:44.205853 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-f9xtx" Jan 21 11:57:44 crc kubenswrapper[4824]: I0121 11:57:44.622745 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-f9xtx"] Jan 21 11:57:45 crc kubenswrapper[4824]: I0121 11:57:45.258188 4824 generic.go:334] "Generic (PLEG): container finished" podID="95ce4f3d-0b58-42b2-9930-23952f0be687" containerID="49247b8448c2901e21cc6977d0f4ee04c18eb2071cb5727c538ddd9d2c738982" exitCode=0 Jan 21 11:57:45 crc kubenswrapper[4824]: I0121 11:57:45.258250 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f9xtx" event={"ID":"95ce4f3d-0b58-42b2-9930-23952f0be687","Type":"ContainerDied","Data":"49247b8448c2901e21cc6977d0f4ee04c18eb2071cb5727c538ddd9d2c738982"} Jan 21 11:57:45 crc kubenswrapper[4824]: I0121 11:57:45.258424 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f9xtx" event={"ID":"95ce4f3d-0b58-42b2-9930-23952f0be687","Type":"ContainerStarted","Data":"277001b0b5c366eff7d23d99e1d7e02281584f2e7f61c7ed99594590aa45edfa"} Jan 21 11:57:45 crc kubenswrapper[4824]: I0121 11:57:45.260750 4824 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 21 11:57:46 crc kubenswrapper[4824]: I0121 11:57:46.091593 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-dwwjr"] Jan 21 11:57:46 crc kubenswrapper[4824]: I0121 11:57:46.093749 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dwwjr" Jan 21 11:57:46 crc kubenswrapper[4824]: I0121 11:57:46.098811 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-dwwjr"] Jan 21 11:57:46 crc kubenswrapper[4824]: I0121 11:57:46.226452 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9ef4d10c-6070-49a6-89c7-cff8b6fc3efe-utilities\") pod \"redhat-marketplace-dwwjr\" (UID: \"9ef4d10c-6070-49a6-89c7-cff8b6fc3efe\") " pod="openshift-marketplace/redhat-marketplace-dwwjr" Jan 21 11:57:46 crc kubenswrapper[4824]: I0121 11:57:46.228670 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9ef4d10c-6070-49a6-89c7-cff8b6fc3efe-catalog-content\") pod \"redhat-marketplace-dwwjr\" (UID: \"9ef4d10c-6070-49a6-89c7-cff8b6fc3efe\") " pod="openshift-marketplace/redhat-marketplace-dwwjr" Jan 21 11:57:46 crc kubenswrapper[4824]: I0121 11:57:46.228833 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gsv99\" (UniqueName: \"kubernetes.io/projected/9ef4d10c-6070-49a6-89c7-cff8b6fc3efe-kube-api-access-gsv99\") pod \"redhat-marketplace-dwwjr\" (UID: \"9ef4d10c-6070-49a6-89c7-cff8b6fc3efe\") " pod="openshift-marketplace/redhat-marketplace-dwwjr" Jan 21 11:57:46 crc kubenswrapper[4824]: I0121 11:57:46.266974 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f9xtx" event={"ID":"95ce4f3d-0b58-42b2-9930-23952f0be687","Type":"ContainerStarted","Data":"638695f0fb682960f2f14ce2d183401c35f1c1d3726d270ab18d01bc6459dc50"} Jan 21 11:57:46 crc kubenswrapper[4824]: I0121 11:57:46.331641 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9ef4d10c-6070-49a6-89c7-cff8b6fc3efe-catalog-content\") pod \"redhat-marketplace-dwwjr\" (UID: \"9ef4d10c-6070-49a6-89c7-cff8b6fc3efe\") " pod="openshift-marketplace/redhat-marketplace-dwwjr" Jan 21 11:57:46 crc kubenswrapper[4824]: I0121 11:57:46.331741 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gsv99\" (UniqueName: \"kubernetes.io/projected/9ef4d10c-6070-49a6-89c7-cff8b6fc3efe-kube-api-access-gsv99\") pod \"redhat-marketplace-dwwjr\" (UID: \"9ef4d10c-6070-49a6-89c7-cff8b6fc3efe\") " pod="openshift-marketplace/redhat-marketplace-dwwjr" Jan 21 11:57:46 crc kubenswrapper[4824]: I0121 11:57:46.331822 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9ef4d10c-6070-49a6-89c7-cff8b6fc3efe-utilities\") pod \"redhat-marketplace-dwwjr\" (UID: \"9ef4d10c-6070-49a6-89c7-cff8b6fc3efe\") " pod="openshift-marketplace/redhat-marketplace-dwwjr" Jan 21 11:57:46 crc kubenswrapper[4824]: I0121 11:57:46.332012 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9ef4d10c-6070-49a6-89c7-cff8b6fc3efe-catalog-content\") pod \"redhat-marketplace-dwwjr\" (UID: \"9ef4d10c-6070-49a6-89c7-cff8b6fc3efe\") " pod="openshift-marketplace/redhat-marketplace-dwwjr" Jan 21 11:57:46 crc kubenswrapper[4824]: I0121 11:57:46.332305 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9ef4d10c-6070-49a6-89c7-cff8b6fc3efe-utilities\") pod \"redhat-marketplace-dwwjr\" (UID: \"9ef4d10c-6070-49a6-89c7-cff8b6fc3efe\") " pod="openshift-marketplace/redhat-marketplace-dwwjr" Jan 21 11:57:46 crc kubenswrapper[4824]: I0121 11:57:46.352699 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gsv99\" (UniqueName: \"kubernetes.io/projected/9ef4d10c-6070-49a6-89c7-cff8b6fc3efe-kube-api-access-gsv99\") pod \"redhat-marketplace-dwwjr\" (UID: \"9ef4d10c-6070-49a6-89c7-cff8b6fc3efe\") " pod="openshift-marketplace/redhat-marketplace-dwwjr" Jan 21 11:57:46 crc kubenswrapper[4824]: I0121 11:57:46.413357 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dwwjr" Jan 21 11:57:46 crc kubenswrapper[4824]: I0121 11:57:46.821661 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-dwwjr"] Jan 21 11:57:47 crc kubenswrapper[4824]: I0121 11:57:47.275111 4824 generic.go:334] "Generic (PLEG): container finished" podID="9ef4d10c-6070-49a6-89c7-cff8b6fc3efe" containerID="4348e65f757892dad73a8da976eba5527fd6deb6ba5e99e3204b12a7526894f3" exitCode=0 Jan 21 11:57:47 crc kubenswrapper[4824]: I0121 11:57:47.275185 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dwwjr" event={"ID":"9ef4d10c-6070-49a6-89c7-cff8b6fc3efe","Type":"ContainerDied","Data":"4348e65f757892dad73a8da976eba5527fd6deb6ba5e99e3204b12a7526894f3"} Jan 21 11:57:47 crc kubenswrapper[4824]: I0121 11:57:47.275216 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dwwjr" event={"ID":"9ef4d10c-6070-49a6-89c7-cff8b6fc3efe","Type":"ContainerStarted","Data":"6aa35848e24a34105b74ff6810b0b5a65321efe98eb936afe9a804fe7f741e36"} Jan 21 11:57:47 crc kubenswrapper[4824]: I0121 11:57:47.276882 4824 generic.go:334] "Generic (PLEG): container finished" podID="95ce4f3d-0b58-42b2-9930-23952f0be687" containerID="638695f0fb682960f2f14ce2d183401c35f1c1d3726d270ab18d01bc6459dc50" exitCode=0 Jan 21 11:57:47 crc kubenswrapper[4824]: I0121 11:57:47.276924 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f9xtx" event={"ID":"95ce4f3d-0b58-42b2-9930-23952f0be687","Type":"ContainerDied","Data":"638695f0fb682960f2f14ce2d183401c35f1c1d3726d270ab18d01bc6459dc50"} Jan 21 11:57:48 crc kubenswrapper[4824]: I0121 11:57:48.289556 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f9xtx" event={"ID":"95ce4f3d-0b58-42b2-9930-23952f0be687","Type":"ContainerStarted","Data":"2f9dc3b067e921b54712fd313ffde00f4358f700d5d2951838163c8e760f1fef"} Jan 21 11:57:48 crc kubenswrapper[4824]: I0121 11:57:48.292253 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dwwjr" event={"ID":"9ef4d10c-6070-49a6-89c7-cff8b6fc3efe","Type":"ContainerStarted","Data":"1395c6331c55dda72692c82cfb7f405a6f14ad9c4c33c90a877304f74be6c19d"} Jan 21 11:57:48 crc kubenswrapper[4824]: I0121 11:57:48.316133 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-f9xtx" podStartSLOduration=2.845292348 podStartE2EDuration="5.316117569s" podCreationTimestamp="2026-01-21 11:57:43 +0000 UTC" firstStartedPulling="2026-01-21 11:57:45.260524009 +0000 UTC m=+2807.553553301" lastFinishedPulling="2026-01-21 11:57:47.73134923 +0000 UTC m=+2810.024378522" observedRunningTime="2026-01-21 11:57:48.308101306 +0000 UTC m=+2810.601130598" watchObservedRunningTime="2026-01-21 11:57:48.316117569 +0000 UTC m=+2810.609146850" Jan 21 11:57:49 crc kubenswrapper[4824]: I0121 11:57:49.300626 4824 generic.go:334] "Generic (PLEG): container finished" podID="9ef4d10c-6070-49a6-89c7-cff8b6fc3efe" containerID="1395c6331c55dda72692c82cfb7f405a6f14ad9c4c33c90a877304f74be6c19d" exitCode=0 Jan 21 11:57:49 crc kubenswrapper[4824]: I0121 11:57:49.300713 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dwwjr" event={"ID":"9ef4d10c-6070-49a6-89c7-cff8b6fc3efe","Type":"ContainerDied","Data":"1395c6331c55dda72692c82cfb7f405a6f14ad9c4c33c90a877304f74be6c19d"} Jan 21 11:57:50 crc kubenswrapper[4824]: I0121 11:57:50.311151 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dwwjr" event={"ID":"9ef4d10c-6070-49a6-89c7-cff8b6fc3efe","Type":"ContainerStarted","Data":"56a664df70a0da76de98566146b77a6f9c3a743b2a196b8ba3f9c56f6ec5dd11"} Jan 21 11:57:50 crc kubenswrapper[4824]: I0121 11:57:50.333859 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-dwwjr" podStartSLOduration=1.835427191 podStartE2EDuration="4.333843129s" podCreationTimestamp="2026-01-21 11:57:46 +0000 UTC" firstStartedPulling="2026-01-21 11:57:47.276396156 +0000 UTC m=+2809.569425447" lastFinishedPulling="2026-01-21 11:57:49.774812093 +0000 UTC m=+2812.067841385" observedRunningTime="2026-01-21 11:57:50.326997343 +0000 UTC m=+2812.620026635" watchObservedRunningTime="2026-01-21 11:57:50.333843129 +0000 UTC m=+2812.626872411" Jan 21 11:57:54 crc kubenswrapper[4824]: I0121 11:57:54.207094 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-f9xtx" Jan 21 11:57:54 crc kubenswrapper[4824]: I0121 11:57:54.207493 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-f9xtx" Jan 21 11:57:54 crc kubenswrapper[4824]: I0121 11:57:54.238211 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-f9xtx" Jan 21 11:57:54 crc kubenswrapper[4824]: I0121 11:57:54.367183 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-f9xtx" Jan 21 11:57:54 crc kubenswrapper[4824]: I0121 11:57:54.881753 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-f9xtx"] Jan 21 11:57:56 crc kubenswrapper[4824]: I0121 11:57:56.348408 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-f9xtx" podUID="95ce4f3d-0b58-42b2-9930-23952f0be687" containerName="registry-server" containerID="cri-o://2f9dc3b067e921b54712fd313ffde00f4358f700d5d2951838163c8e760f1fef" gracePeriod=2 Jan 21 11:57:56 crc kubenswrapper[4824]: I0121 11:57:56.413460 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-dwwjr" Jan 21 11:57:56 crc kubenswrapper[4824]: I0121 11:57:56.413518 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-dwwjr" Jan 21 11:57:56 crc kubenswrapper[4824]: I0121 11:57:56.450445 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-dwwjr" Jan 21 11:57:56 crc kubenswrapper[4824]: I0121 11:57:56.735361 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-f9xtx" Jan 21 11:57:56 crc kubenswrapper[4824]: I0121 11:57:56.830998 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9cfdg\" (UniqueName: \"kubernetes.io/projected/95ce4f3d-0b58-42b2-9930-23952f0be687-kube-api-access-9cfdg\") pod \"95ce4f3d-0b58-42b2-9930-23952f0be687\" (UID: \"95ce4f3d-0b58-42b2-9930-23952f0be687\") " Jan 21 11:57:56 crc kubenswrapper[4824]: I0121 11:57:56.831323 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/95ce4f3d-0b58-42b2-9930-23952f0be687-utilities\") pod \"95ce4f3d-0b58-42b2-9930-23952f0be687\" (UID: \"95ce4f3d-0b58-42b2-9930-23952f0be687\") " Jan 21 11:57:56 crc kubenswrapper[4824]: I0121 11:57:56.831419 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/95ce4f3d-0b58-42b2-9930-23952f0be687-catalog-content\") pod \"95ce4f3d-0b58-42b2-9930-23952f0be687\" (UID: \"95ce4f3d-0b58-42b2-9930-23952f0be687\") " Jan 21 11:57:56 crc kubenswrapper[4824]: I0121 11:57:56.831995 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/95ce4f3d-0b58-42b2-9930-23952f0be687-utilities" (OuterVolumeSpecName: "utilities") pod "95ce4f3d-0b58-42b2-9930-23952f0be687" (UID: "95ce4f3d-0b58-42b2-9930-23952f0be687"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:57:56 crc kubenswrapper[4824]: I0121 11:57:56.832331 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/95ce4f3d-0b58-42b2-9930-23952f0be687-utilities\") on node \"crc\" DevicePath \"\"" Jan 21 11:57:56 crc kubenswrapper[4824]: I0121 11:57:56.844676 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/95ce4f3d-0b58-42b2-9930-23952f0be687-kube-api-access-9cfdg" (OuterVolumeSpecName: "kube-api-access-9cfdg") pod "95ce4f3d-0b58-42b2-9930-23952f0be687" (UID: "95ce4f3d-0b58-42b2-9930-23952f0be687"). InnerVolumeSpecName "kube-api-access-9cfdg". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:57:56 crc kubenswrapper[4824]: I0121 11:57:56.869394 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/95ce4f3d-0b58-42b2-9930-23952f0be687-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "95ce4f3d-0b58-42b2-9930-23952f0be687" (UID: "95ce4f3d-0b58-42b2-9930-23952f0be687"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:57:56 crc kubenswrapper[4824]: I0121 11:57:56.934529 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9cfdg\" (UniqueName: \"kubernetes.io/projected/95ce4f3d-0b58-42b2-9930-23952f0be687-kube-api-access-9cfdg\") on node \"crc\" DevicePath \"\"" Jan 21 11:57:56 crc kubenswrapper[4824]: I0121 11:57:56.934690 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/95ce4f3d-0b58-42b2-9930-23952f0be687-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 21 11:57:57 crc kubenswrapper[4824]: I0121 11:57:57.372366 4824 generic.go:334] "Generic (PLEG): container finished" podID="95ce4f3d-0b58-42b2-9930-23952f0be687" containerID="2f9dc3b067e921b54712fd313ffde00f4358f700d5d2951838163c8e760f1fef" exitCode=0 Jan 21 11:57:57 crc kubenswrapper[4824]: I0121 11:57:57.372488 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-f9xtx" Jan 21 11:57:57 crc kubenswrapper[4824]: I0121 11:57:57.372529 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f9xtx" event={"ID":"95ce4f3d-0b58-42b2-9930-23952f0be687","Type":"ContainerDied","Data":"2f9dc3b067e921b54712fd313ffde00f4358f700d5d2951838163c8e760f1fef"} Jan 21 11:57:57 crc kubenswrapper[4824]: I0121 11:57:57.372571 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f9xtx" event={"ID":"95ce4f3d-0b58-42b2-9930-23952f0be687","Type":"ContainerDied","Data":"277001b0b5c366eff7d23d99e1d7e02281584f2e7f61c7ed99594590aa45edfa"} Jan 21 11:57:57 crc kubenswrapper[4824]: I0121 11:57:57.372588 4824 scope.go:117] "RemoveContainer" containerID="2f9dc3b067e921b54712fd313ffde00f4358f700d5d2951838163c8e760f1fef" Jan 21 11:57:57 crc kubenswrapper[4824]: I0121 11:57:57.391420 4824 scope.go:117] "RemoveContainer" containerID="638695f0fb682960f2f14ce2d183401c35f1c1d3726d270ab18d01bc6459dc50" Jan 21 11:57:57 crc kubenswrapper[4824]: I0121 11:57:57.400582 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-f9xtx"] Jan 21 11:57:57 crc kubenswrapper[4824]: I0121 11:57:57.407568 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-dwwjr" Jan 21 11:57:57 crc kubenswrapper[4824]: I0121 11:57:57.409499 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-f9xtx"] Jan 21 11:57:57 crc kubenswrapper[4824]: I0121 11:57:57.424835 4824 scope.go:117] "RemoveContainer" containerID="49247b8448c2901e21cc6977d0f4ee04c18eb2071cb5727c538ddd9d2c738982" Jan 21 11:57:57 crc kubenswrapper[4824]: I0121 11:57:57.440549 4824 scope.go:117] "RemoveContainer" containerID="2f9dc3b067e921b54712fd313ffde00f4358f700d5d2951838163c8e760f1fef" Jan 21 11:57:57 crc kubenswrapper[4824]: E0121 11:57:57.440931 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2f9dc3b067e921b54712fd313ffde00f4358f700d5d2951838163c8e760f1fef\": container with ID starting with 2f9dc3b067e921b54712fd313ffde00f4358f700d5d2951838163c8e760f1fef not found: ID does not exist" containerID="2f9dc3b067e921b54712fd313ffde00f4358f700d5d2951838163c8e760f1fef" Jan 21 11:57:57 crc kubenswrapper[4824]: I0121 11:57:57.440996 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2f9dc3b067e921b54712fd313ffde00f4358f700d5d2951838163c8e760f1fef"} err="failed to get container status \"2f9dc3b067e921b54712fd313ffde00f4358f700d5d2951838163c8e760f1fef\": rpc error: code = NotFound desc = could not find container \"2f9dc3b067e921b54712fd313ffde00f4358f700d5d2951838163c8e760f1fef\": container with ID starting with 2f9dc3b067e921b54712fd313ffde00f4358f700d5d2951838163c8e760f1fef not found: ID does not exist" Jan 21 11:57:57 crc kubenswrapper[4824]: I0121 11:57:57.441020 4824 scope.go:117] "RemoveContainer" containerID="638695f0fb682960f2f14ce2d183401c35f1c1d3726d270ab18d01bc6459dc50" Jan 21 11:57:57 crc kubenswrapper[4824]: E0121 11:57:57.441307 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"638695f0fb682960f2f14ce2d183401c35f1c1d3726d270ab18d01bc6459dc50\": container with ID starting with 638695f0fb682960f2f14ce2d183401c35f1c1d3726d270ab18d01bc6459dc50 not found: ID does not exist" containerID="638695f0fb682960f2f14ce2d183401c35f1c1d3726d270ab18d01bc6459dc50" Jan 21 11:57:57 crc kubenswrapper[4824]: I0121 11:57:57.441340 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"638695f0fb682960f2f14ce2d183401c35f1c1d3726d270ab18d01bc6459dc50"} err="failed to get container status \"638695f0fb682960f2f14ce2d183401c35f1c1d3726d270ab18d01bc6459dc50\": rpc error: code = NotFound desc = could not find container \"638695f0fb682960f2f14ce2d183401c35f1c1d3726d270ab18d01bc6459dc50\": container with ID starting with 638695f0fb682960f2f14ce2d183401c35f1c1d3726d270ab18d01bc6459dc50 not found: ID does not exist" Jan 21 11:57:57 crc kubenswrapper[4824]: I0121 11:57:57.441382 4824 scope.go:117] "RemoveContainer" containerID="49247b8448c2901e21cc6977d0f4ee04c18eb2071cb5727c538ddd9d2c738982" Jan 21 11:57:57 crc kubenswrapper[4824]: E0121 11:57:57.441763 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"49247b8448c2901e21cc6977d0f4ee04c18eb2071cb5727c538ddd9d2c738982\": container with ID starting with 49247b8448c2901e21cc6977d0f4ee04c18eb2071cb5727c538ddd9d2c738982 not found: ID does not exist" containerID="49247b8448c2901e21cc6977d0f4ee04c18eb2071cb5727c538ddd9d2c738982" Jan 21 11:57:57 crc kubenswrapper[4824]: I0121 11:57:57.441785 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"49247b8448c2901e21cc6977d0f4ee04c18eb2071cb5727c538ddd9d2c738982"} err="failed to get container status \"49247b8448c2901e21cc6977d0f4ee04c18eb2071cb5727c538ddd9d2c738982\": rpc error: code = NotFound desc = could not find container \"49247b8448c2901e21cc6977d0f4ee04c18eb2071cb5727c538ddd9d2c738982\": container with ID starting with 49247b8448c2901e21cc6977d0f4ee04c18eb2071cb5727c538ddd9d2c738982 not found: ID does not exist" Jan 21 11:57:58 crc kubenswrapper[4824]: I0121 11:57:58.058234 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="95ce4f3d-0b58-42b2-9930-23952f0be687" path="/var/lib/kubelet/pods/95ce4f3d-0b58-42b2-9930-23952f0be687/volumes" Jan 21 11:58:01 crc kubenswrapper[4824]: I0121 11:58:01.281934 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-dwwjr"] Jan 21 11:58:01 crc kubenswrapper[4824]: I0121 11:58:01.282499 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-dwwjr" podUID="9ef4d10c-6070-49a6-89c7-cff8b6fc3efe" containerName="registry-server" containerID="cri-o://56a664df70a0da76de98566146b77a6f9c3a743b2a196b8ba3f9c56f6ec5dd11" gracePeriod=2 Jan 21 11:58:01 crc kubenswrapper[4824]: I0121 11:58:01.415459 4824 generic.go:334] "Generic (PLEG): container finished" podID="9ef4d10c-6070-49a6-89c7-cff8b6fc3efe" containerID="56a664df70a0da76de98566146b77a6f9c3a743b2a196b8ba3f9c56f6ec5dd11" exitCode=0 Jan 21 11:58:01 crc kubenswrapper[4824]: I0121 11:58:01.415700 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dwwjr" event={"ID":"9ef4d10c-6070-49a6-89c7-cff8b6fc3efe","Type":"ContainerDied","Data":"56a664df70a0da76de98566146b77a6f9c3a743b2a196b8ba3f9c56f6ec5dd11"} Jan 21 11:58:01 crc kubenswrapper[4824]: I0121 11:58:01.689479 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dwwjr" Jan 21 11:58:01 crc kubenswrapper[4824]: I0121 11:58:01.722978 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9ef4d10c-6070-49a6-89c7-cff8b6fc3efe-catalog-content\") pod \"9ef4d10c-6070-49a6-89c7-cff8b6fc3efe\" (UID: \"9ef4d10c-6070-49a6-89c7-cff8b6fc3efe\") " Jan 21 11:58:01 crc kubenswrapper[4824]: I0121 11:58:01.723032 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9ef4d10c-6070-49a6-89c7-cff8b6fc3efe-utilities\") pod \"9ef4d10c-6070-49a6-89c7-cff8b6fc3efe\" (UID: \"9ef4d10c-6070-49a6-89c7-cff8b6fc3efe\") " Jan 21 11:58:01 crc kubenswrapper[4824]: I0121 11:58:01.723231 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gsv99\" (UniqueName: \"kubernetes.io/projected/9ef4d10c-6070-49a6-89c7-cff8b6fc3efe-kube-api-access-gsv99\") pod \"9ef4d10c-6070-49a6-89c7-cff8b6fc3efe\" (UID: \"9ef4d10c-6070-49a6-89c7-cff8b6fc3efe\") " Jan 21 11:58:01 crc kubenswrapper[4824]: I0121 11:58:01.723973 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9ef4d10c-6070-49a6-89c7-cff8b6fc3efe-utilities" (OuterVolumeSpecName: "utilities") pod "9ef4d10c-6070-49a6-89c7-cff8b6fc3efe" (UID: "9ef4d10c-6070-49a6-89c7-cff8b6fc3efe"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:58:01 crc kubenswrapper[4824]: I0121 11:58:01.728171 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9ef4d10c-6070-49a6-89c7-cff8b6fc3efe-kube-api-access-gsv99" (OuterVolumeSpecName: "kube-api-access-gsv99") pod "9ef4d10c-6070-49a6-89c7-cff8b6fc3efe" (UID: "9ef4d10c-6070-49a6-89c7-cff8b6fc3efe"). InnerVolumeSpecName "kube-api-access-gsv99". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:58:01 crc kubenswrapper[4824]: I0121 11:58:01.739222 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9ef4d10c-6070-49a6-89c7-cff8b6fc3efe-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9ef4d10c-6070-49a6-89c7-cff8b6fc3efe" (UID: "9ef4d10c-6070-49a6-89c7-cff8b6fc3efe"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:58:01 crc kubenswrapper[4824]: I0121 11:58:01.826163 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gsv99\" (UniqueName: \"kubernetes.io/projected/9ef4d10c-6070-49a6-89c7-cff8b6fc3efe-kube-api-access-gsv99\") on node \"crc\" DevicePath \"\"" Jan 21 11:58:01 crc kubenswrapper[4824]: I0121 11:58:01.826435 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9ef4d10c-6070-49a6-89c7-cff8b6fc3efe-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 21 11:58:01 crc kubenswrapper[4824]: I0121 11:58:01.826446 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9ef4d10c-6070-49a6-89c7-cff8b6fc3efe-utilities\") on node \"crc\" DevicePath \"\"" Jan 21 11:58:02 crc kubenswrapper[4824]: I0121 11:58:02.424814 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dwwjr" event={"ID":"9ef4d10c-6070-49a6-89c7-cff8b6fc3efe","Type":"ContainerDied","Data":"6aa35848e24a34105b74ff6810b0b5a65321efe98eb936afe9a804fe7f741e36"} Jan 21 11:58:02 crc kubenswrapper[4824]: I0121 11:58:02.424863 4824 scope.go:117] "RemoveContainer" containerID="56a664df70a0da76de98566146b77a6f9c3a743b2a196b8ba3f9c56f6ec5dd11" Jan 21 11:58:02 crc kubenswrapper[4824]: I0121 11:58:02.424865 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dwwjr" Jan 21 11:58:02 crc kubenswrapper[4824]: I0121 11:58:02.442087 4824 scope.go:117] "RemoveContainer" containerID="1395c6331c55dda72692c82cfb7f405a6f14ad9c4c33c90a877304f74be6c19d" Jan 21 11:58:02 crc kubenswrapper[4824]: I0121 11:58:02.443396 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-dwwjr"] Jan 21 11:58:02 crc kubenswrapper[4824]: I0121 11:58:02.451285 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-dwwjr"] Jan 21 11:58:02 crc kubenswrapper[4824]: I0121 11:58:02.457663 4824 scope.go:117] "RemoveContainer" containerID="4348e65f757892dad73a8da976eba5527fd6deb6ba5e99e3204b12a7526894f3" Jan 21 11:58:04 crc kubenswrapper[4824]: I0121 11:58:04.056829 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9ef4d10c-6070-49a6-89c7-cff8b6fc3efe" path="/var/lib/kubelet/pods/9ef4d10c-6070-49a6-89c7-cff8b6fc3efe/volumes" Jan 21 11:58:16 crc kubenswrapper[4824]: I0121 11:58:16.064940 4824 patch_prober.go:28] interesting pod/machine-config-daemon-zqjg9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 21 11:58:16 crc kubenswrapper[4824]: I0121 11:58:16.065635 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 21 11:58:46 crc kubenswrapper[4824]: I0121 11:58:46.065040 4824 patch_prober.go:28] interesting pod/machine-config-daemon-zqjg9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 21 11:58:46 crc kubenswrapper[4824]: I0121 11:58:46.065513 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 21 11:59:16 crc kubenswrapper[4824]: I0121 11:59:16.064845 4824 patch_prober.go:28] interesting pod/machine-config-daemon-zqjg9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 21 11:59:16 crc kubenswrapper[4824]: I0121 11:59:16.065249 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 21 11:59:16 crc kubenswrapper[4824]: I0121 11:59:16.065289 4824 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" Jan 21 11:59:16 crc kubenswrapper[4824]: I0121 11:59:16.065691 4824 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"753267d7f07800b4c7939c46da56af707c5831b8615d9de85460b5d273fc051a"} pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 21 11:59:16 crc kubenswrapper[4824]: I0121 11:59:16.065743 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" containerName="machine-config-daemon" containerID="cri-o://753267d7f07800b4c7939c46da56af707c5831b8615d9de85460b5d273fc051a" gracePeriod=600 Jan 21 11:59:16 crc kubenswrapper[4824]: E0121 11:59:16.180343 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 11:59:16 crc kubenswrapper[4824]: I0121 11:59:16.910739 4824 generic.go:334] "Generic (PLEG): container finished" podID="33f3d922-4ffe-409b-a49a-d88c85898260" containerID="753267d7f07800b4c7939c46da56af707c5831b8615d9de85460b5d273fc051a" exitCode=0 Jan 21 11:59:16 crc kubenswrapper[4824]: I0121 11:59:16.910782 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" event={"ID":"33f3d922-4ffe-409b-a49a-d88c85898260","Type":"ContainerDied","Data":"753267d7f07800b4c7939c46da56af707c5831b8615d9de85460b5d273fc051a"} Jan 21 11:59:16 crc kubenswrapper[4824]: I0121 11:59:16.910823 4824 scope.go:117] "RemoveContainer" containerID="54c16394c863edb71506776097207cb6de1af7378f2e1570362c90829b6a2501" Jan 21 11:59:16 crc kubenswrapper[4824]: I0121 11:59:16.911450 4824 scope.go:117] "RemoveContainer" containerID="753267d7f07800b4c7939c46da56af707c5831b8615d9de85460b5d273fc051a" Jan 21 11:59:16 crc kubenswrapper[4824]: E0121 11:59:16.911891 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 11:59:28 crc kubenswrapper[4824]: I0121 11:59:28.054513 4824 scope.go:117] "RemoveContainer" containerID="753267d7f07800b4c7939c46da56af707c5831b8615d9de85460b5d273fc051a" Jan 21 11:59:28 crc kubenswrapper[4824]: E0121 11:59:28.055251 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 11:59:40 crc kubenswrapper[4824]: I0121 11:59:40.049230 4824 scope.go:117] "RemoveContainer" containerID="753267d7f07800b4c7939c46da56af707c5831b8615d9de85460b5d273fc051a" Jan 21 11:59:40 crc kubenswrapper[4824]: E0121 11:59:40.050523 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 11:59:53 crc kubenswrapper[4824]: I0121 11:59:53.133829 4824 generic.go:334] "Generic (PLEG): container finished" podID="58d30951-ee56-4217-a614-a04951369c8b" containerID="f56ac3b5212e3537571012922916244691168864ad9d36d1f4cde9480e97ed7a" exitCode=0 Jan 21 11:59:53 crc kubenswrapper[4824]: I0121 11:59:53.133925 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"58d30951-ee56-4217-a614-a04951369c8b","Type":"ContainerDied","Data":"f56ac3b5212e3537571012922916244691168864ad9d36d1f4cde9480e97ed7a"} Jan 21 11:59:54 crc kubenswrapper[4824]: I0121 11:59:54.049810 4824 scope.go:117] "RemoveContainer" containerID="753267d7f07800b4c7939c46da56af707c5831b8615d9de85460b5d273fc051a" Jan 21 11:59:54 crc kubenswrapper[4824]: E0121 11:59:54.050032 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 11:59:54 crc kubenswrapper[4824]: I0121 11:59:54.563024 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Jan 21 11:59:54 crc kubenswrapper[4824]: I0121 11:59:54.604887 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/58d30951-ee56-4217-a614-a04951369c8b-test-operator-ephemeral-temporary\") pod \"58d30951-ee56-4217-a614-a04951369c8b\" (UID: \"58d30951-ee56-4217-a614-a04951369c8b\") " Jan 21 11:59:54 crc kubenswrapper[4824]: I0121 11:59:54.604935 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/58d30951-ee56-4217-a614-a04951369c8b-ca-certs\") pod \"58d30951-ee56-4217-a614-a04951369c8b\" (UID: \"58d30951-ee56-4217-a614-a04951369c8b\") " Jan 21 11:59:54 crc kubenswrapper[4824]: I0121 11:59:54.604998 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/58d30951-ee56-4217-a614-a04951369c8b-ssh-key\") pod \"58d30951-ee56-4217-a614-a04951369c8b\" (UID: \"58d30951-ee56-4217-a614-a04951369c8b\") " Jan 21 11:59:54 crc kubenswrapper[4824]: I0121 11:59:54.605087 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/58d30951-ee56-4217-a614-a04951369c8b-test-operator-ephemeral-workdir\") pod \"58d30951-ee56-4217-a614-a04951369c8b\" (UID: \"58d30951-ee56-4217-a614-a04951369c8b\") " Jan 21 11:59:54 crc kubenswrapper[4824]: I0121 11:59:54.605113 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dr6pd\" (UniqueName: \"kubernetes.io/projected/58d30951-ee56-4217-a614-a04951369c8b-kube-api-access-dr6pd\") pod \"58d30951-ee56-4217-a614-a04951369c8b\" (UID: \"58d30951-ee56-4217-a614-a04951369c8b\") " Jan 21 11:59:54 crc kubenswrapper[4824]: I0121 11:59:54.605711 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/58d30951-ee56-4217-a614-a04951369c8b-openstack-config-secret\") pod \"58d30951-ee56-4217-a614-a04951369c8b\" (UID: \"58d30951-ee56-4217-a614-a04951369c8b\") " Jan 21 11:59:54 crc kubenswrapper[4824]: I0121 11:59:54.605730 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/58d30951-ee56-4217-a614-a04951369c8b-test-operator-ephemeral-temporary" (OuterVolumeSpecName: "test-operator-ephemeral-temporary") pod "58d30951-ee56-4217-a614-a04951369c8b" (UID: "58d30951-ee56-4217-a614-a04951369c8b"). InnerVolumeSpecName "test-operator-ephemeral-temporary". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:59:54 crc kubenswrapper[4824]: I0121 11:59:54.605740 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-logs\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"58d30951-ee56-4217-a614-a04951369c8b\" (UID: \"58d30951-ee56-4217-a614-a04951369c8b\") " Jan 21 11:59:54 crc kubenswrapper[4824]: I0121 11:59:54.605806 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/58d30951-ee56-4217-a614-a04951369c8b-config-data\") pod \"58d30951-ee56-4217-a614-a04951369c8b\" (UID: \"58d30951-ee56-4217-a614-a04951369c8b\") " Jan 21 11:59:54 crc kubenswrapper[4824]: I0121 11:59:54.605837 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/58d30951-ee56-4217-a614-a04951369c8b-openstack-config\") pod \"58d30951-ee56-4217-a614-a04951369c8b\" (UID: \"58d30951-ee56-4217-a614-a04951369c8b\") " Jan 21 11:59:54 crc kubenswrapper[4824]: I0121 11:59:54.606422 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/58d30951-ee56-4217-a614-a04951369c8b-config-data" (OuterVolumeSpecName: "config-data") pod "58d30951-ee56-4217-a614-a04951369c8b" (UID: "58d30951-ee56-4217-a614-a04951369c8b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:59:54 crc kubenswrapper[4824]: I0121 11:59:54.606716 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/58d30951-ee56-4217-a614-a04951369c8b-config-data\") on node \"crc\" DevicePath \"\"" Jan 21 11:59:54 crc kubenswrapper[4824]: I0121 11:59:54.606739 4824 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/58d30951-ee56-4217-a614-a04951369c8b-test-operator-ephemeral-temporary\") on node \"crc\" DevicePath \"\"" Jan 21 11:59:54 crc kubenswrapper[4824]: I0121 11:59:54.607773 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/58d30951-ee56-4217-a614-a04951369c8b-test-operator-ephemeral-workdir" (OuterVolumeSpecName: "test-operator-ephemeral-workdir") pod "58d30951-ee56-4217-a614-a04951369c8b" (UID: "58d30951-ee56-4217-a614-a04951369c8b"). InnerVolumeSpecName "test-operator-ephemeral-workdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 11:59:54 crc kubenswrapper[4824]: I0121 11:59:54.609491 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage08-crc" (OuterVolumeSpecName: "test-operator-logs") pod "58d30951-ee56-4217-a614-a04951369c8b" (UID: "58d30951-ee56-4217-a614-a04951369c8b"). InnerVolumeSpecName "local-storage08-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Jan 21 11:59:54 crc kubenswrapper[4824]: I0121 11:59:54.609560 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/58d30951-ee56-4217-a614-a04951369c8b-kube-api-access-dr6pd" (OuterVolumeSpecName: "kube-api-access-dr6pd") pod "58d30951-ee56-4217-a614-a04951369c8b" (UID: "58d30951-ee56-4217-a614-a04951369c8b"). InnerVolumeSpecName "kube-api-access-dr6pd". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 11:59:54 crc kubenswrapper[4824]: I0121 11:59:54.626266 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/58d30951-ee56-4217-a614-a04951369c8b-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "58d30951-ee56-4217-a614-a04951369c8b" (UID: "58d30951-ee56-4217-a614-a04951369c8b"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:59:54 crc kubenswrapper[4824]: I0121 11:59:54.626648 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/58d30951-ee56-4217-a614-a04951369c8b-ca-certs" (OuterVolumeSpecName: "ca-certs") pod "58d30951-ee56-4217-a614-a04951369c8b" (UID: "58d30951-ee56-4217-a614-a04951369c8b"). InnerVolumeSpecName "ca-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:59:54 crc kubenswrapper[4824]: I0121 11:59:54.627424 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/58d30951-ee56-4217-a614-a04951369c8b-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "58d30951-ee56-4217-a614-a04951369c8b" (UID: "58d30951-ee56-4217-a614-a04951369c8b"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 11:59:54 crc kubenswrapper[4824]: I0121 11:59:54.640170 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/58d30951-ee56-4217-a614-a04951369c8b-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "58d30951-ee56-4217-a614-a04951369c8b" (UID: "58d30951-ee56-4217-a614-a04951369c8b"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 11:59:54 crc kubenswrapper[4824]: I0121 11:59:54.708884 4824 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/58d30951-ee56-4217-a614-a04951369c8b-test-operator-ephemeral-workdir\") on node \"crc\" DevicePath \"\"" Jan 21 11:59:54 crc kubenswrapper[4824]: I0121 11:59:54.709010 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dr6pd\" (UniqueName: \"kubernetes.io/projected/58d30951-ee56-4217-a614-a04951369c8b-kube-api-access-dr6pd\") on node \"crc\" DevicePath \"\"" Jan 21 11:59:54 crc kubenswrapper[4824]: I0121 11:59:54.709082 4824 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/58d30951-ee56-4217-a614-a04951369c8b-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Jan 21 11:59:54 crc kubenswrapper[4824]: I0121 11:59:54.709159 4824 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" " Jan 21 11:59:54 crc kubenswrapper[4824]: I0121 11:59:54.709224 4824 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/58d30951-ee56-4217-a614-a04951369c8b-openstack-config\") on node \"crc\" DevicePath \"\"" Jan 21 11:59:54 crc kubenswrapper[4824]: I0121 11:59:54.709283 4824 reconciler_common.go:293] "Volume detached for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/58d30951-ee56-4217-a614-a04951369c8b-ca-certs\") on node \"crc\" DevicePath \"\"" Jan 21 11:59:54 crc kubenswrapper[4824]: I0121 11:59:54.709342 4824 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/58d30951-ee56-4217-a614-a04951369c8b-ssh-key\") on node \"crc\" DevicePath \"\"" Jan 21 11:59:54 crc kubenswrapper[4824]: I0121 11:59:54.723431 4824 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage08-crc" (UniqueName: "kubernetes.io/local-volume/local-storage08-crc") on node "crc" Jan 21 11:59:54 crc kubenswrapper[4824]: I0121 11:59:54.812173 4824 reconciler_common.go:293] "Volume detached for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" DevicePath \"\"" Jan 21 11:59:55 crc kubenswrapper[4824]: I0121 11:59:55.147144 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"58d30951-ee56-4217-a614-a04951369c8b","Type":"ContainerDied","Data":"5bc5adacec79286d2293a6c562788392bcdca8f811b129442be3e2105c5568bf"} Jan 21 11:59:55 crc kubenswrapper[4824]: I0121 11:59:55.147356 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5bc5adacec79286d2293a6c562788392bcdca8f811b129442be3e2105c5568bf" Jan 21 11:59:55 crc kubenswrapper[4824]: I0121 11:59:55.147204 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Jan 21 12:00:00 crc kubenswrapper[4824]: I0121 12:00:00.145120 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29483280-5hscj"] Jan 21 12:00:00 crc kubenswrapper[4824]: E0121 12:00:00.146267 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95ce4f3d-0b58-42b2-9930-23952f0be687" containerName="registry-server" Jan 21 12:00:00 crc kubenswrapper[4824]: I0121 12:00:00.146282 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="95ce4f3d-0b58-42b2-9930-23952f0be687" containerName="registry-server" Jan 21 12:00:00 crc kubenswrapper[4824]: E0121 12:00:00.146301 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ef4d10c-6070-49a6-89c7-cff8b6fc3efe" containerName="registry-server" Jan 21 12:00:00 crc kubenswrapper[4824]: I0121 12:00:00.146307 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ef4d10c-6070-49a6-89c7-cff8b6fc3efe" containerName="registry-server" Jan 21 12:00:00 crc kubenswrapper[4824]: E0121 12:00:00.146322 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95ce4f3d-0b58-42b2-9930-23952f0be687" containerName="extract-content" Jan 21 12:00:00 crc kubenswrapper[4824]: I0121 12:00:00.146328 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="95ce4f3d-0b58-42b2-9930-23952f0be687" containerName="extract-content" Jan 21 12:00:00 crc kubenswrapper[4824]: E0121 12:00:00.146344 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95ce4f3d-0b58-42b2-9930-23952f0be687" containerName="extract-utilities" Jan 21 12:00:00 crc kubenswrapper[4824]: I0121 12:00:00.146350 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="95ce4f3d-0b58-42b2-9930-23952f0be687" containerName="extract-utilities" Jan 21 12:00:00 crc kubenswrapper[4824]: E0121 12:00:00.146358 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58d30951-ee56-4217-a614-a04951369c8b" containerName="tempest-tests-tempest-tests-runner" Jan 21 12:00:00 crc kubenswrapper[4824]: I0121 12:00:00.146364 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="58d30951-ee56-4217-a614-a04951369c8b" containerName="tempest-tests-tempest-tests-runner" Jan 21 12:00:00 crc kubenswrapper[4824]: E0121 12:00:00.146386 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ef4d10c-6070-49a6-89c7-cff8b6fc3efe" containerName="extract-utilities" Jan 21 12:00:00 crc kubenswrapper[4824]: I0121 12:00:00.146392 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ef4d10c-6070-49a6-89c7-cff8b6fc3efe" containerName="extract-utilities" Jan 21 12:00:00 crc kubenswrapper[4824]: E0121 12:00:00.146399 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ef4d10c-6070-49a6-89c7-cff8b6fc3efe" containerName="extract-content" Jan 21 12:00:00 crc kubenswrapper[4824]: I0121 12:00:00.146404 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ef4d10c-6070-49a6-89c7-cff8b6fc3efe" containerName="extract-content" Jan 21 12:00:00 crc kubenswrapper[4824]: I0121 12:00:00.146600 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="9ef4d10c-6070-49a6-89c7-cff8b6fc3efe" containerName="registry-server" Jan 21 12:00:00 crc kubenswrapper[4824]: I0121 12:00:00.146629 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="58d30951-ee56-4217-a614-a04951369c8b" containerName="tempest-tests-tempest-tests-runner" Jan 21 12:00:00 crc kubenswrapper[4824]: I0121 12:00:00.146641 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="95ce4f3d-0b58-42b2-9930-23952f0be687" containerName="registry-server" Jan 21 12:00:00 crc kubenswrapper[4824]: I0121 12:00:00.147415 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29483280-5hscj" Jan 21 12:00:00 crc kubenswrapper[4824]: I0121 12:00:00.149468 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Jan 21 12:00:00 crc kubenswrapper[4824]: I0121 12:00:00.149695 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Jan 21 12:00:00 crc kubenswrapper[4824]: I0121 12:00:00.155569 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29483280-5hscj"] Jan 21 12:00:00 crc kubenswrapper[4824]: I0121 12:00:00.214313 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/51822659-c0a9-4335-ad4d-8d1bff0a72fd-config-volume\") pod \"collect-profiles-29483280-5hscj\" (UID: \"51822659-c0a9-4335-ad4d-8d1bff0a72fd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29483280-5hscj" Jan 21 12:00:00 crc kubenswrapper[4824]: I0121 12:00:00.214408 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7hcrn\" (UniqueName: \"kubernetes.io/projected/51822659-c0a9-4335-ad4d-8d1bff0a72fd-kube-api-access-7hcrn\") pod \"collect-profiles-29483280-5hscj\" (UID: \"51822659-c0a9-4335-ad4d-8d1bff0a72fd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29483280-5hscj" Jan 21 12:00:00 crc kubenswrapper[4824]: I0121 12:00:00.214457 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/51822659-c0a9-4335-ad4d-8d1bff0a72fd-secret-volume\") pod \"collect-profiles-29483280-5hscj\" (UID: \"51822659-c0a9-4335-ad4d-8d1bff0a72fd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29483280-5hscj" Jan 21 12:00:00 crc kubenswrapper[4824]: I0121 12:00:00.315177 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/51822659-c0a9-4335-ad4d-8d1bff0a72fd-config-volume\") pod \"collect-profiles-29483280-5hscj\" (UID: \"51822659-c0a9-4335-ad4d-8d1bff0a72fd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29483280-5hscj" Jan 21 12:00:00 crc kubenswrapper[4824]: I0121 12:00:00.315246 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7hcrn\" (UniqueName: \"kubernetes.io/projected/51822659-c0a9-4335-ad4d-8d1bff0a72fd-kube-api-access-7hcrn\") pod \"collect-profiles-29483280-5hscj\" (UID: \"51822659-c0a9-4335-ad4d-8d1bff0a72fd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29483280-5hscj" Jan 21 12:00:00 crc kubenswrapper[4824]: I0121 12:00:00.315277 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/51822659-c0a9-4335-ad4d-8d1bff0a72fd-secret-volume\") pod \"collect-profiles-29483280-5hscj\" (UID: \"51822659-c0a9-4335-ad4d-8d1bff0a72fd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29483280-5hscj" Jan 21 12:00:00 crc kubenswrapper[4824]: I0121 12:00:00.316381 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/51822659-c0a9-4335-ad4d-8d1bff0a72fd-config-volume\") pod \"collect-profiles-29483280-5hscj\" (UID: \"51822659-c0a9-4335-ad4d-8d1bff0a72fd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29483280-5hscj" Jan 21 12:00:00 crc kubenswrapper[4824]: I0121 12:00:00.324853 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/51822659-c0a9-4335-ad4d-8d1bff0a72fd-secret-volume\") pod \"collect-profiles-29483280-5hscj\" (UID: \"51822659-c0a9-4335-ad4d-8d1bff0a72fd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29483280-5hscj" Jan 21 12:00:00 crc kubenswrapper[4824]: I0121 12:00:00.330753 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7hcrn\" (UniqueName: \"kubernetes.io/projected/51822659-c0a9-4335-ad4d-8d1bff0a72fd-kube-api-access-7hcrn\") pod \"collect-profiles-29483280-5hscj\" (UID: \"51822659-c0a9-4335-ad4d-8d1bff0a72fd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29483280-5hscj" Jan 21 12:00:00 crc kubenswrapper[4824]: I0121 12:00:00.466176 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29483280-5hscj" Jan 21 12:00:00 crc kubenswrapper[4824]: I0121 12:00:00.867186 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29483280-5hscj"] Jan 21 12:00:01 crc kubenswrapper[4824]: I0121 12:00:01.189853 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29483280-5hscj" event={"ID":"51822659-c0a9-4335-ad4d-8d1bff0a72fd","Type":"ContainerStarted","Data":"03ed238d87b1dad4b9d3e888042bef192d20cdc482bbe612d269cd50129896fd"} Jan 21 12:00:01 crc kubenswrapper[4824]: I0121 12:00:01.190186 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29483280-5hscj" event={"ID":"51822659-c0a9-4335-ad4d-8d1bff0a72fd","Type":"ContainerStarted","Data":"dd3f55cae47f30206c87faa984421a17438917a85679db9f01459ec6d110d380"} Jan 21 12:00:01 crc kubenswrapper[4824]: I0121 12:00:01.205584 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29483280-5hscj" podStartSLOduration=1.205569927 podStartE2EDuration="1.205569927s" podCreationTimestamp="2026-01-21 12:00:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 12:00:01.202116295 +0000 UTC m=+2943.495145588" watchObservedRunningTime="2026-01-21 12:00:01.205569927 +0000 UTC m=+2943.498599219" Jan 21 12:00:02 crc kubenswrapper[4824]: I0121 12:00:02.200438 4824 generic.go:334] "Generic (PLEG): container finished" podID="51822659-c0a9-4335-ad4d-8d1bff0a72fd" containerID="03ed238d87b1dad4b9d3e888042bef192d20cdc482bbe612d269cd50129896fd" exitCode=0 Jan 21 12:00:02 crc kubenswrapper[4824]: I0121 12:00:02.200489 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29483280-5hscj" event={"ID":"51822659-c0a9-4335-ad4d-8d1bff0a72fd","Type":"ContainerDied","Data":"03ed238d87b1dad4b9d3e888042bef192d20cdc482bbe612d269cd50129896fd"} Jan 21 12:00:03 crc kubenswrapper[4824]: I0121 12:00:03.505564 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29483280-5hscj" Jan 21 12:00:03 crc kubenswrapper[4824]: I0121 12:00:03.573229 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/51822659-c0a9-4335-ad4d-8d1bff0a72fd-config-volume\") pod \"51822659-c0a9-4335-ad4d-8d1bff0a72fd\" (UID: \"51822659-c0a9-4335-ad4d-8d1bff0a72fd\") " Jan 21 12:00:03 crc kubenswrapper[4824]: I0121 12:00:03.573297 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7hcrn\" (UniqueName: \"kubernetes.io/projected/51822659-c0a9-4335-ad4d-8d1bff0a72fd-kube-api-access-7hcrn\") pod \"51822659-c0a9-4335-ad4d-8d1bff0a72fd\" (UID: \"51822659-c0a9-4335-ad4d-8d1bff0a72fd\") " Jan 21 12:00:03 crc kubenswrapper[4824]: I0121 12:00:03.573365 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/51822659-c0a9-4335-ad4d-8d1bff0a72fd-secret-volume\") pod \"51822659-c0a9-4335-ad4d-8d1bff0a72fd\" (UID: \"51822659-c0a9-4335-ad4d-8d1bff0a72fd\") " Jan 21 12:00:03 crc kubenswrapper[4824]: I0121 12:00:03.574157 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/51822659-c0a9-4335-ad4d-8d1bff0a72fd-config-volume" (OuterVolumeSpecName: "config-volume") pod "51822659-c0a9-4335-ad4d-8d1bff0a72fd" (UID: "51822659-c0a9-4335-ad4d-8d1bff0a72fd"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 12:00:03 crc kubenswrapper[4824]: I0121 12:00:03.579463 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/51822659-c0a9-4335-ad4d-8d1bff0a72fd-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "51822659-c0a9-4335-ad4d-8d1bff0a72fd" (UID: "51822659-c0a9-4335-ad4d-8d1bff0a72fd"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 12:00:03 crc kubenswrapper[4824]: I0121 12:00:03.579677 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/51822659-c0a9-4335-ad4d-8d1bff0a72fd-kube-api-access-7hcrn" (OuterVolumeSpecName: "kube-api-access-7hcrn") pod "51822659-c0a9-4335-ad4d-8d1bff0a72fd" (UID: "51822659-c0a9-4335-ad4d-8d1bff0a72fd"). InnerVolumeSpecName "kube-api-access-7hcrn". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 12:00:03 crc kubenswrapper[4824]: I0121 12:00:03.676250 4824 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/51822659-c0a9-4335-ad4d-8d1bff0a72fd-config-volume\") on node \"crc\" DevicePath \"\"" Jan 21 12:00:03 crc kubenswrapper[4824]: I0121 12:00:03.676285 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7hcrn\" (UniqueName: \"kubernetes.io/projected/51822659-c0a9-4335-ad4d-8d1bff0a72fd-kube-api-access-7hcrn\") on node \"crc\" DevicePath \"\"" Jan 21 12:00:03 crc kubenswrapper[4824]: I0121 12:00:03.676298 4824 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/51822659-c0a9-4335-ad4d-8d1bff0a72fd-secret-volume\") on node \"crc\" DevicePath \"\"" Jan 21 12:00:04 crc kubenswrapper[4824]: I0121 12:00:04.218525 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29483280-5hscj" event={"ID":"51822659-c0a9-4335-ad4d-8d1bff0a72fd","Type":"ContainerDied","Data":"dd3f55cae47f30206c87faa984421a17438917a85679db9f01459ec6d110d380"} Jan 21 12:00:04 crc kubenswrapper[4824]: I0121 12:00:04.218575 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dd3f55cae47f30206c87faa984421a17438917a85679db9f01459ec6d110d380" Jan 21 12:00:04 crc kubenswrapper[4824]: I0121 12:00:04.218622 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29483280-5hscj" Jan 21 12:00:04 crc kubenswrapper[4824]: I0121 12:00:04.278034 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29483235-4zbf5"] Jan 21 12:00:04 crc kubenswrapper[4824]: I0121 12:00:04.285979 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29483235-4zbf5"] Jan 21 12:00:05 crc kubenswrapper[4824]: I0121 12:00:05.050011 4824 scope.go:117] "RemoveContainer" containerID="753267d7f07800b4c7939c46da56af707c5831b8615d9de85460b5d273fc051a" Jan 21 12:00:05 crc kubenswrapper[4824]: E0121 12:00:05.051272 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 12:00:05 crc kubenswrapper[4824]: I0121 12:00:05.414333 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Jan 21 12:00:05 crc kubenswrapper[4824]: E0121 12:00:05.414828 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="51822659-c0a9-4335-ad4d-8d1bff0a72fd" containerName="collect-profiles" Jan 21 12:00:05 crc kubenswrapper[4824]: I0121 12:00:05.414847 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="51822659-c0a9-4335-ad4d-8d1bff0a72fd" containerName="collect-profiles" Jan 21 12:00:05 crc kubenswrapper[4824]: I0121 12:00:05.415118 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="51822659-c0a9-4335-ad4d-8d1bff0a72fd" containerName="collect-profiles" Jan 21 12:00:05 crc kubenswrapper[4824]: I0121 12:00:05.415795 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Jan 21 12:00:05 crc kubenswrapper[4824]: I0121 12:00:05.418312 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-qs54z" Jan 21 12:00:05 crc kubenswrapper[4824]: I0121 12:00:05.419497 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Jan 21 12:00:05 crc kubenswrapper[4824]: I0121 12:00:05.611405 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cjl5c\" (UniqueName: \"kubernetes.io/projected/333160e2-3c05-401f-af89-abae8c0cb9d9-kube-api-access-cjl5c\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"333160e2-3c05-401f-af89-abae8c0cb9d9\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Jan 21 12:00:05 crc kubenswrapper[4824]: I0121 12:00:05.611731 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"333160e2-3c05-401f-af89-abae8c0cb9d9\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Jan 21 12:00:05 crc kubenswrapper[4824]: I0121 12:00:05.713106 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cjl5c\" (UniqueName: \"kubernetes.io/projected/333160e2-3c05-401f-af89-abae8c0cb9d9-kube-api-access-cjl5c\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"333160e2-3c05-401f-af89-abae8c0cb9d9\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Jan 21 12:00:05 crc kubenswrapper[4824]: I0121 12:00:05.713175 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"333160e2-3c05-401f-af89-abae8c0cb9d9\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Jan 21 12:00:05 crc kubenswrapper[4824]: I0121 12:00:05.713650 4824 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"333160e2-3c05-401f-af89-abae8c0cb9d9\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Jan 21 12:00:05 crc kubenswrapper[4824]: I0121 12:00:05.731629 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cjl5c\" (UniqueName: \"kubernetes.io/projected/333160e2-3c05-401f-af89-abae8c0cb9d9-kube-api-access-cjl5c\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"333160e2-3c05-401f-af89-abae8c0cb9d9\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Jan 21 12:00:05 crc kubenswrapper[4824]: I0121 12:00:05.734915 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"333160e2-3c05-401f-af89-abae8c0cb9d9\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Jan 21 12:00:06 crc kubenswrapper[4824]: I0121 12:00:06.030795 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Jan 21 12:00:06 crc kubenswrapper[4824]: I0121 12:00:06.061787 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="123c89e0-3113-4b28-9129-3848a7288052" path="/var/lib/kubelet/pods/123c89e0-3113-4b28-9129-3848a7288052/volumes" Jan 21 12:00:06 crc kubenswrapper[4824]: I0121 12:00:06.440644 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Jan 21 12:00:07 crc kubenswrapper[4824]: I0121 12:00:07.249719 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"333160e2-3c05-401f-af89-abae8c0cb9d9","Type":"ContainerStarted","Data":"7b61a3f10a73097a6b45f89bcec677347f8c66f5dfa2b3f35db827e53eab7185"} Jan 21 12:00:08 crc kubenswrapper[4824]: I0121 12:00:08.258867 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"333160e2-3c05-401f-af89-abae8c0cb9d9","Type":"ContainerStarted","Data":"8032ae7e95d06d6b6e62f69d1cbf031c04ad5d6ec99a69cbbb9694bd7abedba3"} Jan 21 12:00:08 crc kubenswrapper[4824]: I0121 12:00:08.275032 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" podStartSLOduration=2.325221253 podStartE2EDuration="3.275009063s" podCreationTimestamp="2026-01-21 12:00:05 +0000 UTC" firstStartedPulling="2026-01-21 12:00:06.443953385 +0000 UTC m=+2948.736982678" lastFinishedPulling="2026-01-21 12:00:07.393741196 +0000 UTC m=+2949.686770488" observedRunningTime="2026-01-21 12:00:08.269543549 +0000 UTC m=+2950.562572841" watchObservedRunningTime="2026-01-21 12:00:08.275009063 +0000 UTC m=+2950.568038355" Jan 21 12:00:17 crc kubenswrapper[4824]: I0121 12:00:17.050843 4824 scope.go:117] "RemoveContainer" containerID="753267d7f07800b4c7939c46da56af707c5831b8615d9de85460b5d273fc051a" Jan 21 12:00:17 crc kubenswrapper[4824]: E0121 12:00:17.051611 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 12:00:26 crc kubenswrapper[4824]: I0121 12:00:26.038429 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-87fjw/must-gather-z5tfn"] Jan 21 12:00:26 crc kubenswrapper[4824]: I0121 12:00:26.040583 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-87fjw/must-gather-z5tfn" Jan 21 12:00:26 crc kubenswrapper[4824]: I0121 12:00:26.043913 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-87fjw"/"kube-root-ca.crt" Jan 21 12:00:26 crc kubenswrapper[4824]: I0121 12:00:26.049297 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-87fjw"/"openshift-service-ca.crt" Jan 21 12:00:26 crc kubenswrapper[4824]: I0121 12:00:26.066751 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-87fjw/must-gather-z5tfn"] Jan 21 12:00:26 crc kubenswrapper[4824]: I0121 12:00:26.100079 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/2245f836-eda4-4bf5-aa8a-5b18de2927b2-must-gather-output\") pod \"must-gather-z5tfn\" (UID: \"2245f836-eda4-4bf5-aa8a-5b18de2927b2\") " pod="openshift-must-gather-87fjw/must-gather-z5tfn" Jan 21 12:00:26 crc kubenswrapper[4824]: I0121 12:00:26.100310 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qjtv2\" (UniqueName: \"kubernetes.io/projected/2245f836-eda4-4bf5-aa8a-5b18de2927b2-kube-api-access-qjtv2\") pod \"must-gather-z5tfn\" (UID: \"2245f836-eda4-4bf5-aa8a-5b18de2927b2\") " pod="openshift-must-gather-87fjw/must-gather-z5tfn" Jan 21 12:00:26 crc kubenswrapper[4824]: I0121 12:00:26.206544 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qjtv2\" (UniqueName: \"kubernetes.io/projected/2245f836-eda4-4bf5-aa8a-5b18de2927b2-kube-api-access-qjtv2\") pod \"must-gather-z5tfn\" (UID: \"2245f836-eda4-4bf5-aa8a-5b18de2927b2\") " pod="openshift-must-gather-87fjw/must-gather-z5tfn" Jan 21 12:00:26 crc kubenswrapper[4824]: I0121 12:00:26.209709 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/2245f836-eda4-4bf5-aa8a-5b18de2927b2-must-gather-output\") pod \"must-gather-z5tfn\" (UID: \"2245f836-eda4-4bf5-aa8a-5b18de2927b2\") " pod="openshift-must-gather-87fjw/must-gather-z5tfn" Jan 21 12:00:26 crc kubenswrapper[4824]: I0121 12:00:26.210220 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/2245f836-eda4-4bf5-aa8a-5b18de2927b2-must-gather-output\") pod \"must-gather-z5tfn\" (UID: \"2245f836-eda4-4bf5-aa8a-5b18de2927b2\") " pod="openshift-must-gather-87fjw/must-gather-z5tfn" Jan 21 12:00:26 crc kubenswrapper[4824]: I0121 12:00:26.260046 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qjtv2\" (UniqueName: \"kubernetes.io/projected/2245f836-eda4-4bf5-aa8a-5b18de2927b2-kube-api-access-qjtv2\") pod \"must-gather-z5tfn\" (UID: \"2245f836-eda4-4bf5-aa8a-5b18de2927b2\") " pod="openshift-must-gather-87fjw/must-gather-z5tfn" Jan 21 12:00:26 crc kubenswrapper[4824]: I0121 12:00:26.357679 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-87fjw/must-gather-z5tfn" Jan 21 12:00:26 crc kubenswrapper[4824]: W0121 12:00:26.802668 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2245f836_eda4_4bf5_aa8a_5b18de2927b2.slice/crio-ba666897ba27d94dafdd68ff02864be8c54ee5955437ed8916a10c4f73931701 WatchSource:0}: Error finding container ba666897ba27d94dafdd68ff02864be8c54ee5955437ed8916a10c4f73931701: Status 404 returned error can't find the container with id ba666897ba27d94dafdd68ff02864be8c54ee5955437ed8916a10c4f73931701 Jan 21 12:00:26 crc kubenswrapper[4824]: I0121 12:00:26.802952 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-87fjw/must-gather-z5tfn"] Jan 21 12:00:27 crc kubenswrapper[4824]: I0121 12:00:27.428937 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-87fjw/must-gather-z5tfn" event={"ID":"2245f836-eda4-4bf5-aa8a-5b18de2927b2","Type":"ContainerStarted","Data":"ba666897ba27d94dafdd68ff02864be8c54ee5955437ed8916a10c4f73931701"} Jan 21 12:00:31 crc kubenswrapper[4824]: I0121 12:00:31.049716 4824 scope.go:117] "RemoveContainer" containerID="753267d7f07800b4c7939c46da56af707c5831b8615d9de85460b5d273fc051a" Jan 21 12:00:31 crc kubenswrapper[4824]: E0121 12:00:31.050542 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 12:00:33 crc kubenswrapper[4824]: I0121 12:00:33.495695 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-87fjw/must-gather-z5tfn" event={"ID":"2245f836-eda4-4bf5-aa8a-5b18de2927b2","Type":"ContainerStarted","Data":"a7da4d1f02d49f551acb1d97262ed02dfab6791465333cc45925d14ffc0a4320"} Jan 21 12:00:33 crc kubenswrapper[4824]: I0121 12:00:33.496481 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-87fjw/must-gather-z5tfn" event={"ID":"2245f836-eda4-4bf5-aa8a-5b18de2927b2","Type":"ContainerStarted","Data":"a6e5e2a199134ec83203b77006aebfa54870b884b3f5b127c99a1be886340ffe"} Jan 21 12:00:33 crc kubenswrapper[4824]: I0121 12:00:33.518817 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-87fjw/must-gather-z5tfn" podStartSLOduration=1.94938882 podStartE2EDuration="7.518416432s" podCreationTimestamp="2026-01-21 12:00:26 +0000 UTC" firstStartedPulling="2026-01-21 12:00:26.805298386 +0000 UTC m=+2969.098327678" lastFinishedPulling="2026-01-21 12:00:32.374325999 +0000 UTC m=+2974.667355290" observedRunningTime="2026-01-21 12:00:33.512850889 +0000 UTC m=+2975.805880180" watchObservedRunningTime="2026-01-21 12:00:33.518416432 +0000 UTC m=+2975.811445724" Jan 21 12:00:35 crc kubenswrapper[4824]: I0121 12:00:35.635024 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-87fjw/crc-debug-cs7ct"] Jan 21 12:00:35 crc kubenswrapper[4824]: I0121 12:00:35.637551 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-87fjw/crc-debug-cs7ct" Jan 21 12:00:35 crc kubenswrapper[4824]: I0121 12:00:35.639493 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-87fjw"/"default-dockercfg-xkx7x" Jan 21 12:00:35 crc kubenswrapper[4824]: I0121 12:00:35.733806 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c7mlv\" (UniqueName: \"kubernetes.io/projected/67011da1-4c52-4b94-8f0f-c6f8349af235-kube-api-access-c7mlv\") pod \"crc-debug-cs7ct\" (UID: \"67011da1-4c52-4b94-8f0f-c6f8349af235\") " pod="openshift-must-gather-87fjw/crc-debug-cs7ct" Jan 21 12:00:35 crc kubenswrapper[4824]: I0121 12:00:35.733898 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/67011da1-4c52-4b94-8f0f-c6f8349af235-host\") pod \"crc-debug-cs7ct\" (UID: \"67011da1-4c52-4b94-8f0f-c6f8349af235\") " pod="openshift-must-gather-87fjw/crc-debug-cs7ct" Jan 21 12:00:35 crc kubenswrapper[4824]: I0121 12:00:35.837391 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c7mlv\" (UniqueName: \"kubernetes.io/projected/67011da1-4c52-4b94-8f0f-c6f8349af235-kube-api-access-c7mlv\") pod \"crc-debug-cs7ct\" (UID: \"67011da1-4c52-4b94-8f0f-c6f8349af235\") " pod="openshift-must-gather-87fjw/crc-debug-cs7ct" Jan 21 12:00:35 crc kubenswrapper[4824]: I0121 12:00:35.837505 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/67011da1-4c52-4b94-8f0f-c6f8349af235-host\") pod \"crc-debug-cs7ct\" (UID: \"67011da1-4c52-4b94-8f0f-c6f8349af235\") " pod="openshift-must-gather-87fjw/crc-debug-cs7ct" Jan 21 12:00:35 crc kubenswrapper[4824]: I0121 12:00:35.838034 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/67011da1-4c52-4b94-8f0f-c6f8349af235-host\") pod \"crc-debug-cs7ct\" (UID: \"67011da1-4c52-4b94-8f0f-c6f8349af235\") " pod="openshift-must-gather-87fjw/crc-debug-cs7ct" Jan 21 12:00:35 crc kubenswrapper[4824]: I0121 12:00:35.863832 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c7mlv\" (UniqueName: \"kubernetes.io/projected/67011da1-4c52-4b94-8f0f-c6f8349af235-kube-api-access-c7mlv\") pod \"crc-debug-cs7ct\" (UID: \"67011da1-4c52-4b94-8f0f-c6f8349af235\") " pod="openshift-must-gather-87fjw/crc-debug-cs7ct" Jan 21 12:00:35 crc kubenswrapper[4824]: I0121 12:00:35.957666 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-87fjw/crc-debug-cs7ct" Jan 21 12:00:36 crc kubenswrapper[4824]: I0121 12:00:36.522950 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-87fjw/crc-debug-cs7ct" event={"ID":"67011da1-4c52-4b94-8f0f-c6f8349af235","Type":"ContainerStarted","Data":"fc5413e9aaabe44b332afe86bee51d572012c0e71b94f4311343b145fb32e634"} Jan 21 12:00:38 crc kubenswrapper[4824]: I0121 12:00:38.045173 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-6c7f965fb4-8slr2_3f7f97b5-8abd-4351-8396-cf1b76c376a7/barbican-api-log/0.log" Jan 21 12:00:38 crc kubenswrapper[4824]: I0121 12:00:38.055399 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-6c7f965fb4-8slr2_3f7f97b5-8abd-4351-8396-cf1b76c376a7/barbican-api/0.log" Jan 21 12:00:38 crc kubenswrapper[4824]: I0121 12:00:38.160545 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-79b7cfb8cd-ktpp5_8b84edbe-4485-4027-bfc4-cebbb1b10543/barbican-keystone-listener-log/0.log" Jan 21 12:00:38 crc kubenswrapper[4824]: I0121 12:00:38.166319 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-79b7cfb8cd-ktpp5_8b84edbe-4485-4027-bfc4-cebbb1b10543/barbican-keystone-listener/0.log" Jan 21 12:00:38 crc kubenswrapper[4824]: I0121 12:00:38.180838 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-7cf777545f-f8drb_283e6982-7444-48c7-b7ac-ed4210329a08/barbican-worker-log/0.log" Jan 21 12:00:38 crc kubenswrapper[4824]: I0121 12:00:38.187213 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-7cf777545f-f8drb_283e6982-7444-48c7-b7ac-ed4210329a08/barbican-worker/0.log" Jan 21 12:00:38 crc kubenswrapper[4824]: I0121 12:00:38.216468 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-9z7z6_300a037e-ac25-4a9e-9339-6cc707969f7c/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Jan 21 12:00:38 crc kubenswrapper[4824]: I0121 12:00:38.256812 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_adc98d07-2b32-4f30-bdb7-40923468389e/ceilometer-central-agent/0.log" Jan 21 12:00:38 crc kubenswrapper[4824]: I0121 12:00:38.278842 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_adc98d07-2b32-4f30-bdb7-40923468389e/ceilometer-notification-agent/0.log" Jan 21 12:00:38 crc kubenswrapper[4824]: I0121 12:00:38.283392 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_adc98d07-2b32-4f30-bdb7-40923468389e/sg-core/0.log" Jan 21 12:00:38 crc kubenswrapper[4824]: I0121 12:00:38.292430 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_adc98d07-2b32-4f30-bdb7-40923468389e/proxy-httpd/0.log" Jan 21 12:00:38 crc kubenswrapper[4824]: I0121 12:00:38.307332 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_ff9de537-ecd2-469d-b90e-5857bbf64c2d/cinder-api-log/0.log" Jan 21 12:00:38 crc kubenswrapper[4824]: I0121 12:00:38.336936 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_ff9de537-ecd2-469d-b90e-5857bbf64c2d/cinder-api/0.log" Jan 21 12:00:38 crc kubenswrapper[4824]: I0121 12:00:38.364235 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_2e0f0267-58d2-4690-9187-8ddf03ef082e/cinder-scheduler/0.log" Jan 21 12:00:38 crc kubenswrapper[4824]: I0121 12:00:38.394653 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_2e0f0267-58d2-4690-9187-8ddf03ef082e/probe/0.log" Jan 21 12:00:38 crc kubenswrapper[4824]: I0121 12:00:38.416871 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-wqzrz_b8c7ec7b-ee3f-46bc-95f8-4406713ba43a/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Jan 21 12:00:38 crc kubenswrapper[4824]: I0121 12:00:38.436084 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-4bd5n_8d46d1eb-4230-49c7-8376-698e7689ad2b/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Jan 21 12:00:38 crc kubenswrapper[4824]: I0121 12:00:38.470767 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-6f6df4f56c-26d8b_976a6446-e79c-42db-be45-43147767f1da/dnsmasq-dns/0.log" Jan 21 12:00:38 crc kubenswrapper[4824]: I0121 12:00:38.475201 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-6f6df4f56c-26d8b_976a6446-e79c-42db-be45-43147767f1da/init/0.log" Jan 21 12:00:38 crc kubenswrapper[4824]: I0121 12:00:38.500849 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-dpgt9_8dbaf72e-72a7-4073-a4c1-95cbdabaa64b/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Jan 21 12:00:38 crc kubenswrapper[4824]: I0121 12:00:38.515643 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_219725d0-9bef-4e42-8b78-cbc74a8ac9db/glance-log/0.log" Jan 21 12:00:38 crc kubenswrapper[4824]: I0121 12:00:38.532915 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_219725d0-9bef-4e42-8b78-cbc74a8ac9db/glance-httpd/0.log" Jan 21 12:00:38 crc kubenswrapper[4824]: I0121 12:00:38.543598 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_7419ae4a-e8ea-4384-a9b8-3a3e02e27f25/glance-log/0.log" Jan 21 12:00:38 crc kubenswrapper[4824]: I0121 12:00:38.563609 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_7419ae4a-e8ea-4384-a9b8-3a3e02e27f25/glance-httpd/0.log" Jan 21 12:00:39 crc kubenswrapper[4824]: I0121 12:00:39.005307 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-api-549445f5fc-fwxlz_8655cad5-2680-4cf6-b50a-0745adeeb469/heat-api/0.log" Jan 21 12:00:39 crc kubenswrapper[4824]: I0121 12:00:39.244169 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-cfnapi-68b8cf894c-ws5zd_a1aefb21-aac8-4f8f-a402-16006c1f336a/heat-cfnapi/0.log" Jan 21 12:00:39 crc kubenswrapper[4824]: I0121 12:00:39.364625 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-engine-5bc6fc99dd-nv45v_ca8530d7-bcfb-45fa-b0ff-a2702c081bda/heat-engine/0.log" Jan 21 12:00:39 crc kubenswrapper[4824]: I0121 12:00:39.385917 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2_b821ea8d-9327-4623-b5be-3da7d3872cab/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Jan 21 12:00:39 crc kubenswrapper[4824]: I0121 12:00:39.414363 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-b74gh_60fedbf7-07e6-41be-ada3-2a06f28835e3/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Jan 21 12:00:39 crc kubenswrapper[4824]: I0121 12:00:39.768293 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-6c5fc769c9-spmzr_7d943b5e-d05f-445a-888d-7ebe340468b9/keystone-api/0.log" Jan 21 12:00:39 crc kubenswrapper[4824]: I0121 12:00:39.780425 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_d42233a9-cfd9-486d-b314-f17d6a9d3b03/kube-state-metrics/0.log" Jan 21 12:00:39 crc kubenswrapper[4824]: I0121 12:00:39.815675 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-pgjpn_fdde8991-c204-45a0-b344-dcdc41a9a275/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Jan 21 12:00:43 crc kubenswrapper[4824]: I0121 12:00:43.050349 4824 scope.go:117] "RemoveContainer" containerID="753267d7f07800b4c7939c46da56af707c5831b8615d9de85460b5d273fc051a" Jan 21 12:00:43 crc kubenswrapper[4824]: E0121 12:00:43.051138 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 12:00:47 crc kubenswrapper[4824]: I0121 12:00:47.633198 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-87fjw/crc-debug-cs7ct" event={"ID":"67011da1-4c52-4b94-8f0f-c6f8349af235","Type":"ContainerStarted","Data":"c3e2158290b1bee0a6e844d1cfc3b1ea1c8ca3e18fabfc983d0a39b05fe5af6d"} Jan 21 12:00:47 crc kubenswrapper[4824]: I0121 12:00:47.649586 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-87fjw/crc-debug-cs7ct" podStartSLOduration=1.220151406 podStartE2EDuration="12.649570589s" podCreationTimestamp="2026-01-21 12:00:35 +0000 UTC" firstStartedPulling="2026-01-21 12:00:36.003670906 +0000 UTC m=+2978.296700198" lastFinishedPulling="2026-01-21 12:00:47.433090089 +0000 UTC m=+2989.726119381" observedRunningTime="2026-01-21 12:00:47.648243057 +0000 UTC m=+2989.941272359" watchObservedRunningTime="2026-01-21 12:00:47.649570589 +0000 UTC m=+2989.942599881" Jan 21 12:00:54 crc kubenswrapper[4824]: I0121 12:00:54.049368 4824 scope.go:117] "RemoveContainer" containerID="753267d7f07800b4c7939c46da56af707c5831b8615d9de85460b5d273fc051a" Jan 21 12:00:54 crc kubenswrapper[4824]: E0121 12:00:54.050592 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 12:00:55 crc kubenswrapper[4824]: I0121 12:00:55.089342 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_dba0050b-8a73-4355-a1e0-7c9a03557ead/memcached/0.log" Jan 21 12:00:55 crc kubenswrapper[4824]: I0121 12:00:55.194635 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-659565f76f-jdg2q_b6f2ba03-746e-465c-bde2-c917129fb125/neutron-api/0.log" Jan 21 12:00:55 crc kubenswrapper[4824]: I0121 12:00:55.252817 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-659565f76f-jdg2q_b6f2ba03-746e-465c-bde2-c917129fb125/neutron-httpd/0.log" Jan 21 12:00:55 crc kubenswrapper[4824]: I0121 12:00:55.274019 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-rf47l_4f3c3a58-5d6b-4cc2-9707-d088b2fd463d/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Jan 21 12:00:55 crc kubenswrapper[4824]: I0121 12:00:55.482387 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_55525bd3-ae2b-494d-9a33-dd7d00c576b2/nova-api-log/0.log" Jan 21 12:00:55 crc kubenswrapper[4824]: I0121 12:00:55.763104 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_55525bd3-ae2b-494d-9a33-dd7d00c576b2/nova-api-api/0.log" Jan 21 12:00:55 crc kubenswrapper[4824]: I0121 12:00:55.860705 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_8effdacf-8fec-40b4-8f61-8856ac0f232c/nova-cell0-conductor-conductor/0.log" Jan 21 12:00:55 crc kubenswrapper[4824]: I0121 12:00:55.934030 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_c398d1e9-13e1-400c-a609-bceea49aea51/nova-cell1-conductor-conductor/0.log" Jan 21 12:00:55 crc kubenswrapper[4824]: I0121 12:00:55.983437 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_8e42e9d6-485f-4b97-b049-316cd47af30e/nova-cell1-novncproxy-novncproxy/0.log" Jan 21 12:00:56 crc kubenswrapper[4824]: I0121 12:00:56.032610 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-7crw8_7355a8fc-8c18-4de1-8baf-f6294ea9e11c/nova-edpm-deployment-openstack-edpm-ipam/0.log" Jan 21 12:00:56 crc kubenswrapper[4824]: I0121 12:00:56.146502 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_654a41d0-b44a-4194-a514-a3c60126d0d9/nova-metadata-log/0.log" Jan 21 12:00:56 crc kubenswrapper[4824]: I0121 12:00:56.823499 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_654a41d0-b44a-4194-a514-a3c60126d0d9/nova-metadata-metadata/0.log" Jan 21 12:00:56 crc kubenswrapper[4824]: I0121 12:00:56.922520 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_7860ffbb-b435-4736-b2d3-f8871e70dc7a/nova-scheduler-scheduler/0.log" Jan 21 12:00:56 crc kubenswrapper[4824]: I0121 12:00:56.943166 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_7cd8d62b-ab74-4a1b-95f5-7253ed6c0346/galera/0.log" Jan 21 12:00:56 crc kubenswrapper[4824]: I0121 12:00:56.954721 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_7cd8d62b-ab74-4a1b-95f5-7253ed6c0346/mysql-bootstrap/0.log" Jan 21 12:00:56 crc kubenswrapper[4824]: I0121 12:00:56.975725 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_218aca21-4c37-4862-b891-fc228baa5348/galera/0.log" Jan 21 12:00:56 crc kubenswrapper[4824]: I0121 12:00:56.987750 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_218aca21-4c37-4862-b891-fc228baa5348/mysql-bootstrap/0.log" Jan 21 12:00:56 crc kubenswrapper[4824]: I0121 12:00:56.996590 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_1af94d08-6d1e-4a97-8f82-27a1755d17a1/openstackclient/0.log" Jan 21 12:00:57 crc kubenswrapper[4824]: I0121 12:00:57.010852 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-htwd6_511f2b6d-a08d-49f8-b393-ab222219d4a7/ovn-controller/0.log" Jan 21 12:00:57 crc kubenswrapper[4824]: I0121 12:00:57.020926 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-vrpmw_8652f3f6-1ad7-49d9-9048-827be17a3dd0/openstack-network-exporter/0.log" Jan 21 12:00:57 crc kubenswrapper[4824]: I0121 12:00:57.034410 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-gr9gj_b27b3f77-8476-4f3e-a946-f4f26fc66cb5/ovsdb-server/0.log" Jan 21 12:00:57 crc kubenswrapper[4824]: I0121 12:00:57.045322 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-gr9gj_b27b3f77-8476-4f3e-a946-f4f26fc66cb5/ovs-vswitchd/0.log" Jan 21 12:00:57 crc kubenswrapper[4824]: I0121 12:00:57.052919 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-gr9gj_b27b3f77-8476-4f3e-a946-f4f26fc66cb5/ovsdb-server-init/0.log" Jan 21 12:00:57 crc kubenswrapper[4824]: I0121 12:00:57.090327 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-gf4zq_18e65eb9-3b96-48cd-88b2-9f7fb0e54b6f/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Jan 21 12:00:57 crc kubenswrapper[4824]: I0121 12:00:57.100092 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_8b9c25b1-d0b9-430e-88cf-0aeaf6785461/ovn-northd/0.log" Jan 21 12:00:57 crc kubenswrapper[4824]: I0121 12:00:57.104664 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_8b9c25b1-d0b9-430e-88cf-0aeaf6785461/openstack-network-exporter/0.log" Jan 21 12:00:57 crc kubenswrapper[4824]: I0121 12:00:57.118863 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_5c0b5876-da04-49e6-8946-4fc6060a64ec/ovsdbserver-nb/0.log" Jan 21 12:00:57 crc kubenswrapper[4824]: I0121 12:00:57.127455 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_5c0b5876-da04-49e6-8946-4fc6060a64ec/openstack-network-exporter/0.log" Jan 21 12:00:57 crc kubenswrapper[4824]: I0121 12:00:57.139837 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_e86f23d4-099d-4e8b-b3a4-d2fc7152c1ef/ovsdbserver-sb/0.log" Jan 21 12:00:57 crc kubenswrapper[4824]: I0121 12:00:57.148098 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_e86f23d4-099d-4e8b-b3a4-d2fc7152c1ef/openstack-network-exporter/0.log" Jan 21 12:00:57 crc kubenswrapper[4824]: I0121 12:00:57.216846 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-8654bbf998-p8crm_c23477ba-5a41-44da-a8b7-6606595b9726/placement-log/0.log" Jan 21 12:00:57 crc kubenswrapper[4824]: I0121 12:00:57.271722 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-8654bbf998-p8crm_c23477ba-5a41-44da-a8b7-6606595b9726/placement-api/0.log" Jan 21 12:00:57 crc kubenswrapper[4824]: I0121 12:00:57.295540 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_5d8e256a-aa1b-4ccb-8e8c-c86be1e31762/rabbitmq/0.log" Jan 21 12:00:57 crc kubenswrapper[4824]: I0121 12:00:57.299177 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_5d8e256a-aa1b-4ccb-8e8c-c86be1e31762/setup-container/0.log" Jan 21 12:00:57 crc kubenswrapper[4824]: I0121 12:00:57.324599 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_e4a4d725-1a49-4292-af74-6ab09ab15cdb/rabbitmq/0.log" Jan 21 12:00:57 crc kubenswrapper[4824]: I0121 12:00:57.328714 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_e4a4d725-1a49-4292-af74-6ab09ab15cdb/setup-container/0.log" Jan 21 12:00:57 crc kubenswrapper[4824]: I0121 12:00:57.341701 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-8w57w_890acdab-9f31-4afc-80e4-c1df308ac5a6/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Jan 21 12:00:57 crc kubenswrapper[4824]: I0121 12:00:57.350996 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-2flzn_34c68ecf-cf29-44cb-9880-a3d45f625454/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Jan 21 12:00:57 crc kubenswrapper[4824]: I0121 12:00:57.361391 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-2mbcf_e3664d03-d53d-4a49-8cab-c4266f2d1426/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Jan 21 12:00:57 crc kubenswrapper[4824]: I0121 12:00:57.370449 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-pn4sw_ad533443-01b7-4860-9920-93ed67f6b52f/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Jan 21 12:00:57 crc kubenswrapper[4824]: I0121 12:00:57.380423 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-wbrwb_4529344f-3dfd-4858-83f1-4abffe47f2ad/ssh-known-hosts-edpm-deployment/0.log" Jan 21 12:00:57 crc kubenswrapper[4824]: I0121 12:00:57.463486 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-79f4c948b7-4hz4p_01d4ce4d-3728-4f4c-9fbf-ce58793b6b7d/proxy-httpd/0.log" Jan 21 12:00:57 crc kubenswrapper[4824]: I0121 12:00:57.478080 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-79f4c948b7-4hz4p_01d4ce4d-3728-4f4c-9fbf-ce58793b6b7d/proxy-server/0.log" Jan 21 12:00:57 crc kubenswrapper[4824]: I0121 12:00:57.486124 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-nc6vf_29473359-3517-4b86-bd5e-80e25706ff27/swift-ring-rebalance/0.log" Jan 21 12:00:57 crc kubenswrapper[4824]: I0121 12:00:57.509216 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_f980e493-ca74-47f1-8f6f-aca9c9011bf5/account-server/0.log" Jan 21 12:00:57 crc kubenswrapper[4824]: I0121 12:00:57.525284 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_f980e493-ca74-47f1-8f6f-aca9c9011bf5/account-replicator/0.log" Jan 21 12:00:57 crc kubenswrapper[4824]: I0121 12:00:57.529593 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_f980e493-ca74-47f1-8f6f-aca9c9011bf5/account-auditor/0.log" Jan 21 12:00:57 crc kubenswrapper[4824]: I0121 12:00:57.536101 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_f980e493-ca74-47f1-8f6f-aca9c9011bf5/account-reaper/0.log" Jan 21 12:00:57 crc kubenswrapper[4824]: I0121 12:00:57.545939 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_f980e493-ca74-47f1-8f6f-aca9c9011bf5/container-server/0.log" Jan 21 12:00:57 crc kubenswrapper[4824]: I0121 12:00:57.562450 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_f980e493-ca74-47f1-8f6f-aca9c9011bf5/container-replicator/0.log" Jan 21 12:00:57 crc kubenswrapper[4824]: I0121 12:00:57.566104 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_f980e493-ca74-47f1-8f6f-aca9c9011bf5/container-auditor/0.log" Jan 21 12:00:57 crc kubenswrapper[4824]: I0121 12:00:57.574170 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_f980e493-ca74-47f1-8f6f-aca9c9011bf5/container-updater/0.log" Jan 21 12:00:57 crc kubenswrapper[4824]: I0121 12:00:57.582348 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_f980e493-ca74-47f1-8f6f-aca9c9011bf5/object-server/0.log" Jan 21 12:00:57 crc kubenswrapper[4824]: I0121 12:00:57.596440 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_f980e493-ca74-47f1-8f6f-aca9c9011bf5/object-replicator/0.log" Jan 21 12:00:57 crc kubenswrapper[4824]: I0121 12:00:57.605243 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_f980e493-ca74-47f1-8f6f-aca9c9011bf5/object-auditor/0.log" Jan 21 12:00:57 crc kubenswrapper[4824]: I0121 12:00:57.613219 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_f980e493-ca74-47f1-8f6f-aca9c9011bf5/object-updater/0.log" Jan 21 12:00:57 crc kubenswrapper[4824]: I0121 12:00:57.623134 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_f980e493-ca74-47f1-8f6f-aca9c9011bf5/object-expirer/0.log" Jan 21 12:00:57 crc kubenswrapper[4824]: I0121 12:00:57.627057 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_f980e493-ca74-47f1-8f6f-aca9c9011bf5/rsync/0.log" Jan 21 12:00:57 crc kubenswrapper[4824]: I0121 12:00:57.633312 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_f980e493-ca74-47f1-8f6f-aca9c9011bf5/swift-recon-cron/0.log" Jan 21 12:00:57 crc kubenswrapper[4824]: I0121 12:00:57.693596 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-g8gx6_49b2a136-e5c3-4741-80c6-4c545d4a82a0/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Jan 21 12:00:57 crc kubenswrapper[4824]: I0121 12:00:57.720571 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_58d30951-ee56-4217-a614-a04951369c8b/tempest-tests-tempest-tests-runner/0.log" Jan 21 12:00:57 crc kubenswrapper[4824]: I0121 12:00:57.729009 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_333160e2-3c05-401f-af89-abae8c0cb9d9/test-operator-logs-container/0.log" Jan 21 12:00:57 crc kubenswrapper[4824]: I0121 12:00:57.743933 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-5b6wz_d672f153-6ef2-432d-829a-db63c8daab69/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Jan 21 12:00:59 crc kubenswrapper[4824]: I0121 12:00:59.602479 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-6968d8fdc4-bbv7p_d30695b8-6f6a-48f0-88ed-9181484634b8/controller/0.log" Jan 21 12:00:59 crc kubenswrapper[4824]: I0121 12:00:59.608679 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-6968d8fdc4-bbv7p_d30695b8-6f6a-48f0-88ed-9181484634b8/kube-rbac-proxy/0.log" Jan 21 12:00:59 crc kubenswrapper[4824]: I0121 12:00:59.626302 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9q9wj_e38576f0-c423-497a-a30d-60c2e8edd93a/controller/0.log" Jan 21 12:01:00 crc kubenswrapper[4824]: I0121 12:01:00.149302 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29483281-bz8hk"] Jan 21 12:01:00 crc kubenswrapper[4824]: I0121 12:01:00.158318 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29483281-bz8hk"] Jan 21 12:01:00 crc kubenswrapper[4824]: I0121 12:01:00.158529 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29483281-bz8hk" Jan 21 12:01:00 crc kubenswrapper[4824]: I0121 12:01:00.179212 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c92d6f21-1fe2-41a2-a516-a0d0740ab36f-config-data\") pod \"keystone-cron-29483281-bz8hk\" (UID: \"c92d6f21-1fe2-41a2-a516-a0d0740ab36f\") " pod="openstack/keystone-cron-29483281-bz8hk" Jan 21 12:01:00 crc kubenswrapper[4824]: I0121 12:01:00.179492 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c92d6f21-1fe2-41a2-a516-a0d0740ab36f-combined-ca-bundle\") pod \"keystone-cron-29483281-bz8hk\" (UID: \"c92d6f21-1fe2-41a2-a516-a0d0740ab36f\") " pod="openstack/keystone-cron-29483281-bz8hk" Jan 21 12:01:00 crc kubenswrapper[4824]: I0121 12:01:00.179542 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c92d6f21-1fe2-41a2-a516-a0d0740ab36f-fernet-keys\") pod \"keystone-cron-29483281-bz8hk\" (UID: \"c92d6f21-1fe2-41a2-a516-a0d0740ab36f\") " pod="openstack/keystone-cron-29483281-bz8hk" Jan 21 12:01:00 crc kubenswrapper[4824]: I0121 12:01:00.179615 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l5rns\" (UniqueName: \"kubernetes.io/projected/c92d6f21-1fe2-41a2-a516-a0d0740ab36f-kube-api-access-l5rns\") pod \"keystone-cron-29483281-bz8hk\" (UID: \"c92d6f21-1fe2-41a2-a516-a0d0740ab36f\") " pod="openstack/keystone-cron-29483281-bz8hk" Jan 21 12:01:00 crc kubenswrapper[4824]: I0121 12:01:00.281546 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c92d6f21-1fe2-41a2-a516-a0d0740ab36f-combined-ca-bundle\") pod \"keystone-cron-29483281-bz8hk\" (UID: \"c92d6f21-1fe2-41a2-a516-a0d0740ab36f\") " pod="openstack/keystone-cron-29483281-bz8hk" Jan 21 12:01:00 crc kubenswrapper[4824]: I0121 12:01:00.281606 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c92d6f21-1fe2-41a2-a516-a0d0740ab36f-fernet-keys\") pod \"keystone-cron-29483281-bz8hk\" (UID: \"c92d6f21-1fe2-41a2-a516-a0d0740ab36f\") " pod="openstack/keystone-cron-29483281-bz8hk" Jan 21 12:01:00 crc kubenswrapper[4824]: I0121 12:01:00.281671 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l5rns\" (UniqueName: \"kubernetes.io/projected/c92d6f21-1fe2-41a2-a516-a0d0740ab36f-kube-api-access-l5rns\") pod \"keystone-cron-29483281-bz8hk\" (UID: \"c92d6f21-1fe2-41a2-a516-a0d0740ab36f\") " pod="openstack/keystone-cron-29483281-bz8hk" Jan 21 12:01:00 crc kubenswrapper[4824]: I0121 12:01:00.281760 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c92d6f21-1fe2-41a2-a516-a0d0740ab36f-config-data\") pod \"keystone-cron-29483281-bz8hk\" (UID: \"c92d6f21-1fe2-41a2-a516-a0d0740ab36f\") " pod="openstack/keystone-cron-29483281-bz8hk" Jan 21 12:01:00 crc kubenswrapper[4824]: I0121 12:01:00.290623 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c92d6f21-1fe2-41a2-a516-a0d0740ab36f-fernet-keys\") pod \"keystone-cron-29483281-bz8hk\" (UID: \"c92d6f21-1fe2-41a2-a516-a0d0740ab36f\") " pod="openstack/keystone-cron-29483281-bz8hk" Jan 21 12:01:00 crc kubenswrapper[4824]: I0121 12:01:00.306351 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c92d6f21-1fe2-41a2-a516-a0d0740ab36f-combined-ca-bundle\") pod \"keystone-cron-29483281-bz8hk\" (UID: \"c92d6f21-1fe2-41a2-a516-a0d0740ab36f\") " pod="openstack/keystone-cron-29483281-bz8hk" Jan 21 12:01:00 crc kubenswrapper[4824]: I0121 12:01:00.312261 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l5rns\" (UniqueName: \"kubernetes.io/projected/c92d6f21-1fe2-41a2-a516-a0d0740ab36f-kube-api-access-l5rns\") pod \"keystone-cron-29483281-bz8hk\" (UID: \"c92d6f21-1fe2-41a2-a516-a0d0740ab36f\") " pod="openstack/keystone-cron-29483281-bz8hk" Jan 21 12:01:00 crc kubenswrapper[4824]: I0121 12:01:00.319797 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c92d6f21-1fe2-41a2-a516-a0d0740ab36f-config-data\") pod \"keystone-cron-29483281-bz8hk\" (UID: \"c92d6f21-1fe2-41a2-a516-a0d0740ab36f\") " pod="openstack/keystone-cron-29483281-bz8hk" Jan 21 12:01:00 crc kubenswrapper[4824]: I0121 12:01:00.487186 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29483281-bz8hk" Jan 21 12:01:00 crc kubenswrapper[4824]: I0121 12:01:00.897259 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9q9wj_e38576f0-c423-497a-a30d-60c2e8edd93a/frr/0.log" Jan 21 12:01:00 crc kubenswrapper[4824]: I0121 12:01:00.907936 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9q9wj_e38576f0-c423-497a-a30d-60c2e8edd93a/reloader/0.log" Jan 21 12:01:00 crc kubenswrapper[4824]: I0121 12:01:00.917684 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9q9wj_e38576f0-c423-497a-a30d-60c2e8edd93a/frr-metrics/0.log" Jan 21 12:01:00 crc kubenswrapper[4824]: I0121 12:01:00.924526 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9q9wj_e38576f0-c423-497a-a30d-60c2e8edd93a/kube-rbac-proxy/0.log" Jan 21 12:01:00 crc kubenswrapper[4824]: I0121 12:01:00.932260 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9q9wj_e38576f0-c423-497a-a30d-60c2e8edd93a/kube-rbac-proxy-frr/0.log" Jan 21 12:01:00 crc kubenswrapper[4824]: I0121 12:01:00.937781 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9q9wj_e38576f0-c423-497a-a30d-60c2e8edd93a/cp-frr-files/0.log" Jan 21 12:01:00 crc kubenswrapper[4824]: I0121 12:01:00.951261 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9q9wj_e38576f0-c423-497a-a30d-60c2e8edd93a/cp-reloader/0.log" Jan 21 12:01:00 crc kubenswrapper[4824]: I0121 12:01:00.958298 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9q9wj_e38576f0-c423-497a-a30d-60c2e8edd93a/cp-metrics/0.log" Jan 21 12:01:00 crc kubenswrapper[4824]: I0121 12:01:00.969316 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7df86c4f6c-fdblb_d71f993c-b2c7-47af-884f-4cbe6dcf3c1a/frr-k8s-webhook-server/0.log" Jan 21 12:01:00 crc kubenswrapper[4824]: I0121 12:01:00.989744 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-5c58bcfcff-69lct_714d2131-afc8-4828-840f-e2f0ce727d5d/manager/0.log" Jan 21 12:01:00 crc kubenswrapper[4824]: I0121 12:01:00.997352 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-68bc9bfcb6-hzmp5_81fed5ab-ec3d-46f0-8998-199393bffba0/webhook-server/0.log" Jan 21 12:01:01 crc kubenswrapper[4824]: I0121 12:01:01.047140 4824 scope.go:117] "RemoveContainer" containerID="1c72418277c3fff2993fb8936f438b1d2049c9f13bdefd24ee7c2900f6904eca" Jan 21 12:01:01 crc kubenswrapper[4824]: I0121 12:01:01.066100 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29483281-bz8hk"] Jan 21 12:01:01 crc kubenswrapper[4824]: I0121 12:01:01.369278 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-8tsb9_baf5db51-d3ac-4b97-9c5f-afc598d59ab7/speaker/0.log" Jan 21 12:01:01 crc kubenswrapper[4824]: I0121 12:01:01.377051 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-8tsb9_baf5db51-d3ac-4b97-9c5f-afc598d59ab7/kube-rbac-proxy/0.log" Jan 21 12:01:01 crc kubenswrapper[4824]: I0121 12:01:01.744410 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29483281-bz8hk" event={"ID":"c92d6f21-1fe2-41a2-a516-a0d0740ab36f","Type":"ContainerStarted","Data":"0f8293175001be831fd07c8e6b87e9d57d97179a89c7a7a13cc4a25f076d8348"} Jan 21 12:01:01 crc kubenswrapper[4824]: I0121 12:01:01.744462 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29483281-bz8hk" event={"ID":"c92d6f21-1fe2-41a2-a516-a0d0740ab36f","Type":"ContainerStarted","Data":"afdbd9e5b38ea1c61d377b1aa74001c97326aa386ba38e7c7774fbbd18823a66"} Jan 21 12:01:01 crc kubenswrapper[4824]: I0121 12:01:01.759240 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29483281-bz8hk" podStartSLOduration=1.7592209479999998 podStartE2EDuration="1.759220948s" podCreationTimestamp="2026-01-21 12:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 12:01:01.75765157 +0000 UTC m=+3004.050680862" watchObservedRunningTime="2026-01-21 12:01:01.759220948 +0000 UTC m=+3004.052250240" Jan 21 12:01:01 crc kubenswrapper[4824]: I0121 12:01:01.888699 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7ddb5c749-bvptq_8ed6bb2c-2770-4785-bf38-9b6fd0bd8472/manager/0.log" Jan 21 12:01:01 crc kubenswrapper[4824]: I0121 12:01:01.934450 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-9b68f5989-24r54_a25733d4-a5f8-4c97-a897-8d96e637c253/manager/0.log" Jan 21 12:01:01 crc kubenswrapper[4824]: I0121 12:01:01.943357 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_db4b9aacd6f51b5aff2d66474158631eb1e58fb4d1358e39c8aaaca1bdn7cxr_72ab26cb-149e-4825-9587-3acfa66a368f/extract/0.log" Jan 21 12:01:01 crc kubenswrapper[4824]: I0121 12:01:01.951907 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_db4b9aacd6f51b5aff2d66474158631eb1e58fb4d1358e39c8aaaca1bdn7cxr_72ab26cb-149e-4825-9587-3acfa66a368f/util/0.log" Jan 21 12:01:01 crc kubenswrapper[4824]: I0121 12:01:01.964573 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_db4b9aacd6f51b5aff2d66474158631eb1e58fb4d1358e39c8aaaca1bdn7cxr_72ab26cb-149e-4825-9587-3acfa66a368f/pull/0.log" Jan 21 12:01:01 crc kubenswrapper[4824]: I0121 12:01:01.984405 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-9f958b845-wtfj4_14dd1f12-3943-42e4-be1b-c9e37e49d21b/manager/0.log" Jan 21 12:01:02 crc kubenswrapper[4824]: I0121 12:01:02.056416 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-c6994669c-hvq6s_b7053a68-ce89-48b6-9b5d-1c6dac813a75/manager/0.log" Jan 21 12:01:02 crc kubenswrapper[4824]: I0121 12:01:02.116735 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5fbc59b5b6-hgmf4_944c3123-9063-4731-9581-5b683741edc4/manager/0.log" Jan 21 12:01:02 crc kubenswrapper[4824]: I0121 12:01:02.124746 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-77d5c5b54f-mp5ml_2463afb1-0271-4ab3-ad1a-9329ecaeec4d/manager/0.log" Jan 21 12:01:02 crc kubenswrapper[4824]: I0121 12:01:02.371183 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-77c48c7859-dzn7w_57e1fe13-b2c8-4711-9efe-1ebbc7e9d12e/manager/0.log" Jan 21 12:01:02 crc kubenswrapper[4824]: I0121 12:01:02.381671 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-78757b4889-gkwjs_fc911701-7c15-4a20-b89e-709521bdd3a1/manager/0.log" Jan 21 12:01:02 crc kubenswrapper[4824]: I0121 12:01:02.453548 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-767fdc4f47-5mhmd_2e5efa9b-40bc-4860-967d-aa37e1838f81/manager/0.log" Jan 21 12:01:02 crc kubenswrapper[4824]: I0121 12:01:02.462972 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-864f6b75bf-c7vsf_18017a19-1676-4fe9-a97e-ae878f423847/manager/0.log" Jan 21 12:01:02 crc kubenswrapper[4824]: I0121 12:01:02.493579 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-c87fff755-d5hn9_19b1c150-f5ed-4432-bfc2-717b78eebc7e/manager/0.log" Jan 21 12:01:02 crc kubenswrapper[4824]: I0121 12:01:02.534605 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-cb4666565-7w8p4_8c74d3da-d870-45e9-96b9-70eb2395c977/manager/0.log" Jan 21 12:01:02 crc kubenswrapper[4824]: I0121 12:01:02.606530 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-65849867d6-8ngx6_b38ac60b-1643-44d9-a1fe-5f75129247dc/manager/0.log" Jan 21 12:01:02 crc kubenswrapper[4824]: I0121 12:01:02.616805 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-7fc9b76cf6-tncbx_5429da27-a618-44da-8aeb-12b917af8734/manager/0.log" Jan 21 12:01:02 crc kubenswrapper[4824]: I0121 12:01:02.628626 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-6b68b8b854zxcbl_3fa95966-3147-4e50-8ed8-f702a87992c4/manager/0.log" Jan 21 12:01:02 crc kubenswrapper[4824]: I0121 12:01:02.784206 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-init-dbdbd6855-gkm5b_35334861-8e16-4b16-8cf5-167a049836af/operator/0.log" Jan 21 12:01:03 crc kubenswrapper[4824]: I0121 12:01:03.762286 4824 generic.go:334] "Generic (PLEG): container finished" podID="c92d6f21-1fe2-41a2-a516-a0d0740ab36f" containerID="0f8293175001be831fd07c8e6b87e9d57d97179a89c7a7a13cc4a25f076d8348" exitCode=0 Jan 21 12:01:03 crc kubenswrapper[4824]: I0121 12:01:03.762336 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29483281-bz8hk" event={"ID":"c92d6f21-1fe2-41a2-a516-a0d0740ab36f","Type":"ContainerDied","Data":"0f8293175001be831fd07c8e6b87e9d57d97179a89c7a7a13cc4a25f076d8348"} Jan 21 12:01:03 crc kubenswrapper[4824]: I0121 12:01:03.877025 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-74d84f758c-vjgps_c1128e01-8b7c-4742-b1da-dbf06aa8b4d8/manager/0.log" Jan 21 12:01:03 crc kubenswrapper[4824]: I0121 12:01:03.884170 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-zggxm_78eb1906-d9e6-49f7-a512-ed3cc0a07da7/registry-server/0.log" Jan 21 12:01:03 crc kubenswrapper[4824]: I0121 12:01:03.932332 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-55db956ddc-78z2q_f5639f17-3a5b-4ef4-bd6e-6eeb3274636a/manager/0.log" Jan 21 12:01:03 crc kubenswrapper[4824]: I0121 12:01:03.953621 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-686df47fcb-c8xvs_d8b36106-8f21-4873-890d-c1879ea49068/manager/0.log" Jan 21 12:01:03 crc kubenswrapper[4824]: I0121 12:01:03.973526 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-94lqx_782093ab-751c-4e38-bba3-c5f76308c82e/operator/0.log" Jan 21 12:01:03 crc kubenswrapper[4824]: I0121 12:01:03.993578 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-85dd56d4cc-wlrdt_b4cb80e6-c720-4875-9848-7d9597a354e4/manager/0.log" Jan 21 12:01:04 crc kubenswrapper[4824]: I0121 12:01:04.049503 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-5f8f495fcf-6rz5h_db84b172-59d8-4995-815b-50b492283bee/manager/0.log" Jan 21 12:01:04 crc kubenswrapper[4824]: I0121 12:01:04.063146 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-7cd8bc9dbb-xkp8x_52491e7b-cd8c-403a-942c-312debd6f1c6/manager/0.log" Jan 21 12:01:04 crc kubenswrapper[4824]: I0121 12:01:04.069531 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-64cd966744-9w4ch_0715aa91-2ae5-4e07-8388-c37def4c7743/manager/0.log" Jan 21 12:01:05 crc kubenswrapper[4824]: I0121 12:01:05.048432 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29483281-bz8hk" Jan 21 12:01:05 crc kubenswrapper[4824]: I0121 12:01:05.113054 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l5rns\" (UniqueName: \"kubernetes.io/projected/c92d6f21-1fe2-41a2-a516-a0d0740ab36f-kube-api-access-l5rns\") pod \"c92d6f21-1fe2-41a2-a516-a0d0740ab36f\" (UID: \"c92d6f21-1fe2-41a2-a516-a0d0740ab36f\") " Jan 21 12:01:05 crc kubenswrapper[4824]: I0121 12:01:05.113374 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c92d6f21-1fe2-41a2-a516-a0d0740ab36f-combined-ca-bundle\") pod \"c92d6f21-1fe2-41a2-a516-a0d0740ab36f\" (UID: \"c92d6f21-1fe2-41a2-a516-a0d0740ab36f\") " Jan 21 12:01:05 crc kubenswrapper[4824]: I0121 12:01:05.113425 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c92d6f21-1fe2-41a2-a516-a0d0740ab36f-config-data\") pod \"c92d6f21-1fe2-41a2-a516-a0d0740ab36f\" (UID: \"c92d6f21-1fe2-41a2-a516-a0d0740ab36f\") " Jan 21 12:01:05 crc kubenswrapper[4824]: I0121 12:01:05.113497 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c92d6f21-1fe2-41a2-a516-a0d0740ab36f-fernet-keys\") pod \"c92d6f21-1fe2-41a2-a516-a0d0740ab36f\" (UID: \"c92d6f21-1fe2-41a2-a516-a0d0740ab36f\") " Jan 21 12:01:05 crc kubenswrapper[4824]: I0121 12:01:05.120090 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c92d6f21-1fe2-41a2-a516-a0d0740ab36f-kube-api-access-l5rns" (OuterVolumeSpecName: "kube-api-access-l5rns") pod "c92d6f21-1fe2-41a2-a516-a0d0740ab36f" (UID: "c92d6f21-1fe2-41a2-a516-a0d0740ab36f"). InnerVolumeSpecName "kube-api-access-l5rns". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 12:01:05 crc kubenswrapper[4824]: I0121 12:01:05.125065 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c92d6f21-1fe2-41a2-a516-a0d0740ab36f-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "c92d6f21-1fe2-41a2-a516-a0d0740ab36f" (UID: "c92d6f21-1fe2-41a2-a516-a0d0740ab36f"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 12:01:05 crc kubenswrapper[4824]: I0121 12:01:05.164314 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c92d6f21-1fe2-41a2-a516-a0d0740ab36f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c92d6f21-1fe2-41a2-a516-a0d0740ab36f" (UID: "c92d6f21-1fe2-41a2-a516-a0d0740ab36f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 12:01:05 crc kubenswrapper[4824]: I0121 12:01:05.178775 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c92d6f21-1fe2-41a2-a516-a0d0740ab36f-config-data" (OuterVolumeSpecName: "config-data") pod "c92d6f21-1fe2-41a2-a516-a0d0740ab36f" (UID: "c92d6f21-1fe2-41a2-a516-a0d0740ab36f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 12:01:05 crc kubenswrapper[4824]: I0121 12:01:05.225215 4824 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c92d6f21-1fe2-41a2-a516-a0d0740ab36f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 21 12:01:05 crc kubenswrapper[4824]: I0121 12:01:05.225241 4824 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c92d6f21-1fe2-41a2-a516-a0d0740ab36f-config-data\") on node \"crc\" DevicePath \"\"" Jan 21 12:01:05 crc kubenswrapper[4824]: I0121 12:01:05.225251 4824 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c92d6f21-1fe2-41a2-a516-a0d0740ab36f-fernet-keys\") on node \"crc\" DevicePath \"\"" Jan 21 12:01:05 crc kubenswrapper[4824]: I0121 12:01:05.225263 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l5rns\" (UniqueName: \"kubernetes.io/projected/c92d6f21-1fe2-41a2-a516-a0d0740ab36f-kube-api-access-l5rns\") on node \"crc\" DevicePath \"\"" Jan 21 12:01:05 crc kubenswrapper[4824]: I0121 12:01:05.779594 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29483281-bz8hk" event={"ID":"c92d6f21-1fe2-41a2-a516-a0d0740ab36f","Type":"ContainerDied","Data":"afdbd9e5b38ea1c61d377b1aa74001c97326aa386ba38e7c7774fbbd18823a66"} Jan 21 12:01:05 crc kubenswrapper[4824]: I0121 12:01:05.779654 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="afdbd9e5b38ea1c61d377b1aa74001c97326aa386ba38e7c7774fbbd18823a66" Jan 21 12:01:05 crc kubenswrapper[4824]: I0121 12:01:05.779678 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29483281-bz8hk" Jan 21 12:01:05 crc kubenswrapper[4824]: E0121 12:01:05.976669 4824 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc92d6f21_1fe2_41a2_a516_a0d0740ab36f.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc92d6f21_1fe2_41a2_a516_a0d0740ab36f.slice/crio-afdbd9e5b38ea1c61d377b1aa74001c97326aa386ba38e7c7774fbbd18823a66\": RecentStats: unable to find data in memory cache]" Jan 21 12:01:06 crc kubenswrapper[4824]: I0121 12:01:06.050489 4824 scope.go:117] "RemoveContainer" containerID="753267d7f07800b4c7939c46da56af707c5831b8615d9de85460b5d273fc051a" Jan 21 12:01:06 crc kubenswrapper[4824]: E0121 12:01:06.051366 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 12:01:09 crc kubenswrapper[4824]: I0121 12:01:09.116639 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-r65dm_c117483a-fe5c-423b-995e-9c2e7f59566a/control-plane-machine-set-operator/0.log" Jan 21 12:01:09 crc kubenswrapper[4824]: I0121 12:01:09.129430 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-7grz2_fb8bdb2b-bdc7-4433-be10-296fc67dc1c3/kube-rbac-proxy/0.log" Jan 21 12:01:09 crc kubenswrapper[4824]: I0121 12:01:09.137531 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-7grz2_fb8bdb2b-bdc7-4433-be10-296fc67dc1c3/machine-api-operator/0.log" Jan 21 12:01:20 crc kubenswrapper[4824]: I0121 12:01:20.061560 4824 scope.go:117] "RemoveContainer" containerID="753267d7f07800b4c7939c46da56af707c5831b8615d9de85460b5d273fc051a" Jan 21 12:01:20 crc kubenswrapper[4824]: E0121 12:01:20.062723 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 12:01:22 crc kubenswrapper[4824]: I0121 12:01:22.929371 4824 generic.go:334] "Generic (PLEG): container finished" podID="67011da1-4c52-4b94-8f0f-c6f8349af235" containerID="c3e2158290b1bee0a6e844d1cfc3b1ea1c8ca3e18fabfc983d0a39b05fe5af6d" exitCode=0 Jan 21 12:01:22 crc kubenswrapper[4824]: I0121 12:01:22.929455 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-87fjw/crc-debug-cs7ct" event={"ID":"67011da1-4c52-4b94-8f0f-c6f8349af235","Type":"ContainerDied","Data":"c3e2158290b1bee0a6e844d1cfc3b1ea1c8ca3e18fabfc983d0a39b05fe5af6d"} Jan 21 12:01:24 crc kubenswrapper[4824]: I0121 12:01:24.012313 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-87fjw/crc-debug-cs7ct" Jan 21 12:01:24 crc kubenswrapper[4824]: I0121 12:01:24.037023 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c7mlv\" (UniqueName: \"kubernetes.io/projected/67011da1-4c52-4b94-8f0f-c6f8349af235-kube-api-access-c7mlv\") pod \"67011da1-4c52-4b94-8f0f-c6f8349af235\" (UID: \"67011da1-4c52-4b94-8f0f-c6f8349af235\") " Jan 21 12:01:24 crc kubenswrapper[4824]: I0121 12:01:24.037192 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/67011da1-4c52-4b94-8f0f-c6f8349af235-host\") pod \"67011da1-4c52-4b94-8f0f-c6f8349af235\" (UID: \"67011da1-4c52-4b94-8f0f-c6f8349af235\") " Jan 21 12:01:24 crc kubenswrapper[4824]: I0121 12:01:24.037322 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/67011da1-4c52-4b94-8f0f-c6f8349af235-host" (OuterVolumeSpecName: "host") pod "67011da1-4c52-4b94-8f0f-c6f8349af235" (UID: "67011da1-4c52-4b94-8f0f-c6f8349af235"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 21 12:01:24 crc kubenswrapper[4824]: I0121 12:01:24.038151 4824 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/67011da1-4c52-4b94-8f0f-c6f8349af235-host\") on node \"crc\" DevicePath \"\"" Jan 21 12:01:24 crc kubenswrapper[4824]: I0121 12:01:24.041427 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-87fjw/crc-debug-cs7ct"] Jan 21 12:01:24 crc kubenswrapper[4824]: I0121 12:01:24.043589 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/67011da1-4c52-4b94-8f0f-c6f8349af235-kube-api-access-c7mlv" (OuterVolumeSpecName: "kube-api-access-c7mlv") pod "67011da1-4c52-4b94-8f0f-c6f8349af235" (UID: "67011da1-4c52-4b94-8f0f-c6f8349af235"). InnerVolumeSpecName "kube-api-access-c7mlv". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 12:01:24 crc kubenswrapper[4824]: I0121 12:01:24.058744 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-87fjw/crc-debug-cs7ct"] Jan 21 12:01:24 crc kubenswrapper[4824]: I0121 12:01:24.140345 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c7mlv\" (UniqueName: \"kubernetes.io/projected/67011da1-4c52-4b94-8f0f-c6f8349af235-kube-api-access-c7mlv\") on node \"crc\" DevicePath \"\"" Jan 21 12:01:24 crc kubenswrapper[4824]: I0121 12:01:24.950418 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fc5413e9aaabe44b332afe86bee51d572012c0e71b94f4311343b145fb32e634" Jan 21 12:01:24 crc kubenswrapper[4824]: I0121 12:01:24.950486 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-87fjw/crc-debug-cs7ct" Jan 21 12:01:25 crc kubenswrapper[4824]: I0121 12:01:25.181891 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-87fjw/crc-debug-jqpvk"] Jan 21 12:01:25 crc kubenswrapper[4824]: E0121 12:01:25.182599 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c92d6f21-1fe2-41a2-a516-a0d0740ab36f" containerName="keystone-cron" Jan 21 12:01:25 crc kubenswrapper[4824]: I0121 12:01:25.182616 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="c92d6f21-1fe2-41a2-a516-a0d0740ab36f" containerName="keystone-cron" Jan 21 12:01:25 crc kubenswrapper[4824]: E0121 12:01:25.182635 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="67011da1-4c52-4b94-8f0f-c6f8349af235" containerName="container-00" Jan 21 12:01:25 crc kubenswrapper[4824]: I0121 12:01:25.182641 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="67011da1-4c52-4b94-8f0f-c6f8349af235" containerName="container-00" Jan 21 12:01:25 crc kubenswrapper[4824]: I0121 12:01:25.182885 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="67011da1-4c52-4b94-8f0f-c6f8349af235" containerName="container-00" Jan 21 12:01:25 crc kubenswrapper[4824]: I0121 12:01:25.182903 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="c92d6f21-1fe2-41a2-a516-a0d0740ab36f" containerName="keystone-cron" Jan 21 12:01:25 crc kubenswrapper[4824]: I0121 12:01:25.183553 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-87fjw/crc-debug-jqpvk" Jan 21 12:01:25 crc kubenswrapper[4824]: I0121 12:01:25.185414 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-87fjw"/"default-dockercfg-xkx7x" Jan 21 12:01:25 crc kubenswrapper[4824]: I0121 12:01:25.258186 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dnpcq\" (UniqueName: \"kubernetes.io/projected/87698512-b8b0-416a-b593-d9ca8bc2eeac-kube-api-access-dnpcq\") pod \"crc-debug-jqpvk\" (UID: \"87698512-b8b0-416a-b593-d9ca8bc2eeac\") " pod="openshift-must-gather-87fjw/crc-debug-jqpvk" Jan 21 12:01:25 crc kubenswrapper[4824]: I0121 12:01:25.258520 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/87698512-b8b0-416a-b593-d9ca8bc2eeac-host\") pod \"crc-debug-jqpvk\" (UID: \"87698512-b8b0-416a-b593-d9ca8bc2eeac\") " pod="openshift-must-gather-87fjw/crc-debug-jqpvk" Jan 21 12:01:25 crc kubenswrapper[4824]: I0121 12:01:25.359478 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dnpcq\" (UniqueName: \"kubernetes.io/projected/87698512-b8b0-416a-b593-d9ca8bc2eeac-kube-api-access-dnpcq\") pod \"crc-debug-jqpvk\" (UID: \"87698512-b8b0-416a-b593-d9ca8bc2eeac\") " pod="openshift-must-gather-87fjw/crc-debug-jqpvk" Jan 21 12:01:25 crc kubenswrapper[4824]: I0121 12:01:25.359562 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/87698512-b8b0-416a-b593-d9ca8bc2eeac-host\") pod \"crc-debug-jqpvk\" (UID: \"87698512-b8b0-416a-b593-d9ca8bc2eeac\") " pod="openshift-must-gather-87fjw/crc-debug-jqpvk" Jan 21 12:01:25 crc kubenswrapper[4824]: I0121 12:01:25.359683 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/87698512-b8b0-416a-b593-d9ca8bc2eeac-host\") pod \"crc-debug-jqpvk\" (UID: \"87698512-b8b0-416a-b593-d9ca8bc2eeac\") " pod="openshift-must-gather-87fjw/crc-debug-jqpvk" Jan 21 12:01:25 crc kubenswrapper[4824]: I0121 12:01:25.377403 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dnpcq\" (UniqueName: \"kubernetes.io/projected/87698512-b8b0-416a-b593-d9ca8bc2eeac-kube-api-access-dnpcq\") pod \"crc-debug-jqpvk\" (UID: \"87698512-b8b0-416a-b593-d9ca8bc2eeac\") " pod="openshift-must-gather-87fjw/crc-debug-jqpvk" Jan 21 12:01:25 crc kubenswrapper[4824]: I0121 12:01:25.501710 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-87fjw/crc-debug-jqpvk" Jan 21 12:01:25 crc kubenswrapper[4824]: I0121 12:01:25.963989 4824 generic.go:334] "Generic (PLEG): container finished" podID="87698512-b8b0-416a-b593-d9ca8bc2eeac" containerID="c6793d88c2985e5e3d28f1236121b047a23b50c741fdeb44a8e3cc182cf06683" exitCode=0 Jan 21 12:01:25 crc kubenswrapper[4824]: I0121 12:01:25.964076 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-87fjw/crc-debug-jqpvk" event={"ID":"87698512-b8b0-416a-b593-d9ca8bc2eeac","Type":"ContainerDied","Data":"c6793d88c2985e5e3d28f1236121b047a23b50c741fdeb44a8e3cc182cf06683"} Jan 21 12:01:25 crc kubenswrapper[4824]: I0121 12:01:25.964369 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-87fjw/crc-debug-jqpvk" event={"ID":"87698512-b8b0-416a-b593-d9ca8bc2eeac","Type":"ContainerStarted","Data":"a448dccd21cb3b27e3293d09c067735c54887c539a7fd0b2a40524c5960f4e56"} Jan 21 12:01:26 crc kubenswrapper[4824]: I0121 12:01:26.058610 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="67011da1-4c52-4b94-8f0f-c6f8349af235" path="/var/lib/kubelet/pods/67011da1-4c52-4b94-8f0f-c6f8349af235/volumes" Jan 21 12:01:27 crc kubenswrapper[4824]: I0121 12:01:27.057834 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-87fjw/crc-debug-jqpvk" Jan 21 12:01:27 crc kubenswrapper[4824]: I0121 12:01:27.195090 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/87698512-b8b0-416a-b593-d9ca8bc2eeac-host\") pod \"87698512-b8b0-416a-b593-d9ca8bc2eeac\" (UID: \"87698512-b8b0-416a-b593-d9ca8bc2eeac\") " Jan 21 12:01:27 crc kubenswrapper[4824]: I0121 12:01:27.195191 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/87698512-b8b0-416a-b593-d9ca8bc2eeac-host" (OuterVolumeSpecName: "host") pod "87698512-b8b0-416a-b593-d9ca8bc2eeac" (UID: "87698512-b8b0-416a-b593-d9ca8bc2eeac"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 21 12:01:27 crc kubenswrapper[4824]: I0121 12:01:27.195364 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dnpcq\" (UniqueName: \"kubernetes.io/projected/87698512-b8b0-416a-b593-d9ca8bc2eeac-kube-api-access-dnpcq\") pod \"87698512-b8b0-416a-b593-d9ca8bc2eeac\" (UID: \"87698512-b8b0-416a-b593-d9ca8bc2eeac\") " Jan 21 12:01:27 crc kubenswrapper[4824]: I0121 12:01:27.196995 4824 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/87698512-b8b0-416a-b593-d9ca8bc2eeac-host\") on node \"crc\" DevicePath \"\"" Jan 21 12:01:27 crc kubenswrapper[4824]: I0121 12:01:27.201557 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87698512-b8b0-416a-b593-d9ca8bc2eeac-kube-api-access-dnpcq" (OuterVolumeSpecName: "kube-api-access-dnpcq") pod "87698512-b8b0-416a-b593-d9ca8bc2eeac" (UID: "87698512-b8b0-416a-b593-d9ca8bc2eeac"). InnerVolumeSpecName "kube-api-access-dnpcq". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 12:01:27 crc kubenswrapper[4824]: I0121 12:01:27.298597 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dnpcq\" (UniqueName: \"kubernetes.io/projected/87698512-b8b0-416a-b593-d9ca8bc2eeac-kube-api-access-dnpcq\") on node \"crc\" DevicePath \"\"" Jan 21 12:01:27 crc kubenswrapper[4824]: I0121 12:01:27.988533 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-87fjw/crc-debug-jqpvk" event={"ID":"87698512-b8b0-416a-b593-d9ca8bc2eeac","Type":"ContainerDied","Data":"a448dccd21cb3b27e3293d09c067735c54887c539a7fd0b2a40524c5960f4e56"} Jan 21 12:01:27 crc kubenswrapper[4824]: I0121 12:01:27.989077 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a448dccd21cb3b27e3293d09c067735c54887c539a7fd0b2a40524c5960f4e56" Jan 21 12:01:27 crc kubenswrapper[4824]: I0121 12:01:27.988597 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-87fjw/crc-debug-jqpvk" Jan 21 12:01:28 crc kubenswrapper[4824]: I0121 12:01:28.114412 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-87fjw/crc-debug-jqpvk"] Jan 21 12:01:28 crc kubenswrapper[4824]: I0121 12:01:28.123021 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-87fjw/crc-debug-jqpvk"] Jan 21 12:01:29 crc kubenswrapper[4824]: I0121 12:01:29.272359 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-87fjw/crc-debug-7b2zl"] Jan 21 12:01:29 crc kubenswrapper[4824]: E0121 12:01:29.272857 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="87698512-b8b0-416a-b593-d9ca8bc2eeac" containerName="container-00" Jan 21 12:01:29 crc kubenswrapper[4824]: I0121 12:01:29.272871 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="87698512-b8b0-416a-b593-d9ca8bc2eeac" containerName="container-00" Jan 21 12:01:29 crc kubenswrapper[4824]: I0121 12:01:29.273122 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="87698512-b8b0-416a-b593-d9ca8bc2eeac" containerName="container-00" Jan 21 12:01:29 crc kubenswrapper[4824]: I0121 12:01:29.273780 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-87fjw/crc-debug-7b2zl" Jan 21 12:01:29 crc kubenswrapper[4824]: I0121 12:01:29.276067 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-87fjw"/"default-dockercfg-xkx7x" Jan 21 12:01:29 crc kubenswrapper[4824]: I0121 12:01:29.434433 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ad457817-57db-426b-8191-f35e1d82d266-host\") pod \"crc-debug-7b2zl\" (UID: \"ad457817-57db-426b-8191-f35e1d82d266\") " pod="openshift-must-gather-87fjw/crc-debug-7b2zl" Jan 21 12:01:29 crc kubenswrapper[4824]: I0121 12:01:29.434516 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bdwzx\" (UniqueName: \"kubernetes.io/projected/ad457817-57db-426b-8191-f35e1d82d266-kube-api-access-bdwzx\") pod \"crc-debug-7b2zl\" (UID: \"ad457817-57db-426b-8191-f35e1d82d266\") " pod="openshift-must-gather-87fjw/crc-debug-7b2zl" Jan 21 12:01:29 crc kubenswrapper[4824]: I0121 12:01:29.536377 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bdwzx\" (UniqueName: \"kubernetes.io/projected/ad457817-57db-426b-8191-f35e1d82d266-kube-api-access-bdwzx\") pod \"crc-debug-7b2zl\" (UID: \"ad457817-57db-426b-8191-f35e1d82d266\") " pod="openshift-must-gather-87fjw/crc-debug-7b2zl" Jan 21 12:01:29 crc kubenswrapper[4824]: I0121 12:01:29.536552 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ad457817-57db-426b-8191-f35e1d82d266-host\") pod \"crc-debug-7b2zl\" (UID: \"ad457817-57db-426b-8191-f35e1d82d266\") " pod="openshift-must-gather-87fjw/crc-debug-7b2zl" Jan 21 12:01:29 crc kubenswrapper[4824]: I0121 12:01:29.536633 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ad457817-57db-426b-8191-f35e1d82d266-host\") pod \"crc-debug-7b2zl\" (UID: \"ad457817-57db-426b-8191-f35e1d82d266\") " pod="openshift-must-gather-87fjw/crc-debug-7b2zl" Jan 21 12:01:29 crc kubenswrapper[4824]: I0121 12:01:29.556402 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bdwzx\" (UniqueName: \"kubernetes.io/projected/ad457817-57db-426b-8191-f35e1d82d266-kube-api-access-bdwzx\") pod \"crc-debug-7b2zl\" (UID: \"ad457817-57db-426b-8191-f35e1d82d266\") " pod="openshift-must-gather-87fjw/crc-debug-7b2zl" Jan 21 12:01:29 crc kubenswrapper[4824]: I0121 12:01:29.595941 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-87fjw/crc-debug-7b2zl" Jan 21 12:01:29 crc kubenswrapper[4824]: W0121 12:01:29.622993 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podad457817_57db_426b_8191_f35e1d82d266.slice/crio-ae1325f602a0dbb91e07fcc9773e501ad40e003b6336ac64860e84a0cb84054f WatchSource:0}: Error finding container ae1325f602a0dbb91e07fcc9773e501ad40e003b6336ac64860e84a0cb84054f: Status 404 returned error can't find the container with id ae1325f602a0dbb91e07fcc9773e501ad40e003b6336ac64860e84a0cb84054f Jan 21 12:01:30 crc kubenswrapper[4824]: I0121 12:01:30.010467 4824 generic.go:334] "Generic (PLEG): container finished" podID="ad457817-57db-426b-8191-f35e1d82d266" containerID="6e4471e366c5d457dbdb31676ebe2e10113437d05155002808726047c10699c0" exitCode=0 Jan 21 12:01:30 crc kubenswrapper[4824]: I0121 12:01:30.010575 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-87fjw/crc-debug-7b2zl" event={"ID":"ad457817-57db-426b-8191-f35e1d82d266","Type":"ContainerDied","Data":"6e4471e366c5d457dbdb31676ebe2e10113437d05155002808726047c10699c0"} Jan 21 12:01:30 crc kubenswrapper[4824]: I0121 12:01:30.010899 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-87fjw/crc-debug-7b2zl" event={"ID":"ad457817-57db-426b-8191-f35e1d82d266","Type":"ContainerStarted","Data":"ae1325f602a0dbb91e07fcc9773e501ad40e003b6336ac64860e84a0cb84054f"} Jan 21 12:01:30 crc kubenswrapper[4824]: I0121 12:01:30.060973 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87698512-b8b0-416a-b593-d9ca8bc2eeac" path="/var/lib/kubelet/pods/87698512-b8b0-416a-b593-d9ca8bc2eeac/volumes" Jan 21 12:01:30 crc kubenswrapper[4824]: I0121 12:01:30.061771 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-87fjw/crc-debug-7b2zl"] Jan 21 12:01:30 crc kubenswrapper[4824]: I0121 12:01:30.061818 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-87fjw/crc-debug-7b2zl"] Jan 21 12:01:30 crc kubenswrapper[4824]: I0121 12:01:30.516222 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-858654f9db-qhlzr_b9405067-2dfa-46f9-9471-fe456b3ca5aa/cert-manager-controller/0.log" Jan 21 12:01:30 crc kubenswrapper[4824]: I0121 12:01:30.528035 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-cf98fcc89-xjwkp_d783bb77-baab-4e1c-99e9-ddba7ef78fc2/cert-manager-cainjector/0.log" Jan 21 12:01:30 crc kubenswrapper[4824]: I0121 12:01:30.535795 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-687f57d79b-r6qj7_b12640f9-bf75-4056-9c20-a75d63a0e177/cert-manager-webhook/0.log" Jan 21 12:01:31 crc kubenswrapper[4824]: I0121 12:01:31.112567 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-87fjw/crc-debug-7b2zl" Jan 21 12:01:31 crc kubenswrapper[4824]: I0121 12:01:31.168067 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bdwzx\" (UniqueName: \"kubernetes.io/projected/ad457817-57db-426b-8191-f35e1d82d266-kube-api-access-bdwzx\") pod \"ad457817-57db-426b-8191-f35e1d82d266\" (UID: \"ad457817-57db-426b-8191-f35e1d82d266\") " Jan 21 12:01:31 crc kubenswrapper[4824]: I0121 12:01:31.168141 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ad457817-57db-426b-8191-f35e1d82d266-host\") pod \"ad457817-57db-426b-8191-f35e1d82d266\" (UID: \"ad457817-57db-426b-8191-f35e1d82d266\") " Jan 21 12:01:31 crc kubenswrapper[4824]: I0121 12:01:31.168381 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ad457817-57db-426b-8191-f35e1d82d266-host" (OuterVolumeSpecName: "host") pod "ad457817-57db-426b-8191-f35e1d82d266" (UID: "ad457817-57db-426b-8191-f35e1d82d266"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 21 12:01:31 crc kubenswrapper[4824]: I0121 12:01:31.168682 4824 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ad457817-57db-426b-8191-f35e1d82d266-host\") on node \"crc\" DevicePath \"\"" Jan 21 12:01:31 crc kubenswrapper[4824]: I0121 12:01:31.174200 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ad457817-57db-426b-8191-f35e1d82d266-kube-api-access-bdwzx" (OuterVolumeSpecName: "kube-api-access-bdwzx") pod "ad457817-57db-426b-8191-f35e1d82d266" (UID: "ad457817-57db-426b-8191-f35e1d82d266"). InnerVolumeSpecName "kube-api-access-bdwzx". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 12:01:31 crc kubenswrapper[4824]: I0121 12:01:31.270583 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bdwzx\" (UniqueName: \"kubernetes.io/projected/ad457817-57db-426b-8191-f35e1d82d266-kube-api-access-bdwzx\") on node \"crc\" DevicePath \"\"" Jan 21 12:01:32 crc kubenswrapper[4824]: I0121 12:01:32.039587 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ae1325f602a0dbb91e07fcc9773e501ad40e003b6336ac64860e84a0cb84054f" Jan 21 12:01:32 crc kubenswrapper[4824]: I0121 12:01:32.039667 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-87fjw/crc-debug-7b2zl" Jan 21 12:01:32 crc kubenswrapper[4824]: I0121 12:01:32.073725 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ad457817-57db-426b-8191-f35e1d82d266" path="/var/lib/kubelet/pods/ad457817-57db-426b-8191-f35e1d82d266/volumes" Jan 21 12:01:34 crc kubenswrapper[4824]: I0121 12:01:34.050548 4824 scope.go:117] "RemoveContainer" containerID="753267d7f07800b4c7939c46da56af707c5831b8615d9de85460b5d273fc051a" Jan 21 12:01:34 crc kubenswrapper[4824]: E0121 12:01:34.051541 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 12:01:35 crc kubenswrapper[4824]: I0121 12:01:35.213499 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-7754f76f8b-xjqs5_1575ab9d-016f-403a-990b-3b4c4d53f6f1/nmstate-console-plugin/0.log" Jan 21 12:01:35 crc kubenswrapper[4824]: I0121 12:01:35.238527 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-qjpv9_a01ecff0-9051-4ddb-8c8e-a4f379b6c823/nmstate-handler/0.log" Jan 21 12:01:35 crc kubenswrapper[4824]: I0121 12:01:35.251800 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-54757c584b-w2j6n_30b70686-bf05-4798-b5dc-3c7dc9cee56a/nmstate-metrics/0.log" Jan 21 12:01:35 crc kubenswrapper[4824]: I0121 12:01:35.257456 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-54757c584b-w2j6n_30b70686-bf05-4798-b5dc-3c7dc9cee56a/kube-rbac-proxy/0.log" Jan 21 12:01:35 crc kubenswrapper[4824]: I0121 12:01:35.273024 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-646758c888-zmc4k_df60ca53-0573-4a6a-9235-86e0a1c17f91/nmstate-operator/0.log" Jan 21 12:01:35 crc kubenswrapper[4824]: I0121 12:01:35.291562 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-8474b5b9d8-h6kkh_edf54909-b53e-44de-90bc-355e23bbc767/nmstate-webhook/0.log" Jan 21 12:01:43 crc kubenswrapper[4824]: I0121 12:01:43.221356 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-6968d8fdc4-bbv7p_d30695b8-6f6a-48f0-88ed-9181484634b8/controller/0.log" Jan 21 12:01:43 crc kubenswrapper[4824]: I0121 12:01:43.226091 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-6968d8fdc4-bbv7p_d30695b8-6f6a-48f0-88ed-9181484634b8/kube-rbac-proxy/0.log" Jan 21 12:01:43 crc kubenswrapper[4824]: I0121 12:01:43.242505 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9q9wj_e38576f0-c423-497a-a30d-60c2e8edd93a/controller/0.log" Jan 21 12:01:44 crc kubenswrapper[4824]: I0121 12:01:44.400325 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9q9wj_e38576f0-c423-497a-a30d-60c2e8edd93a/frr/0.log" Jan 21 12:01:44 crc kubenswrapper[4824]: I0121 12:01:44.412242 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9q9wj_e38576f0-c423-497a-a30d-60c2e8edd93a/reloader/0.log" Jan 21 12:01:44 crc kubenswrapper[4824]: I0121 12:01:44.417328 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9q9wj_e38576f0-c423-497a-a30d-60c2e8edd93a/frr-metrics/0.log" Jan 21 12:01:44 crc kubenswrapper[4824]: I0121 12:01:44.422836 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9q9wj_e38576f0-c423-497a-a30d-60c2e8edd93a/kube-rbac-proxy/0.log" Jan 21 12:01:44 crc kubenswrapper[4824]: I0121 12:01:44.428295 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9q9wj_e38576f0-c423-497a-a30d-60c2e8edd93a/kube-rbac-proxy-frr/0.log" Jan 21 12:01:44 crc kubenswrapper[4824]: I0121 12:01:44.436054 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9q9wj_e38576f0-c423-497a-a30d-60c2e8edd93a/cp-frr-files/0.log" Jan 21 12:01:44 crc kubenswrapper[4824]: I0121 12:01:44.439443 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9q9wj_e38576f0-c423-497a-a30d-60c2e8edd93a/cp-reloader/0.log" Jan 21 12:01:44 crc kubenswrapper[4824]: I0121 12:01:44.445618 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9q9wj_e38576f0-c423-497a-a30d-60c2e8edd93a/cp-metrics/0.log" Jan 21 12:01:44 crc kubenswrapper[4824]: I0121 12:01:44.454484 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7df86c4f6c-fdblb_d71f993c-b2c7-47af-884f-4cbe6dcf3c1a/frr-k8s-webhook-server/0.log" Jan 21 12:01:44 crc kubenswrapper[4824]: I0121 12:01:44.473539 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-5c58bcfcff-69lct_714d2131-afc8-4828-840f-e2f0ce727d5d/manager/0.log" Jan 21 12:01:44 crc kubenswrapper[4824]: I0121 12:01:44.479924 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-68bc9bfcb6-hzmp5_81fed5ab-ec3d-46f0-8998-199393bffba0/webhook-server/0.log" Jan 21 12:01:44 crc kubenswrapper[4824]: I0121 12:01:44.768475 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-8tsb9_baf5db51-d3ac-4b97-9c5f-afc598d59ab7/speaker/0.log" Jan 21 12:01:44 crc kubenswrapper[4824]: I0121 12:01:44.775246 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-8tsb9_baf5db51-d3ac-4b97-9c5f-afc598d59ab7/kube-rbac-proxy/0.log" Jan 21 12:01:47 crc kubenswrapper[4824]: I0121 12:01:47.505194 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcwqkxq_d9005d95-8246-435d-aaea-5c60d58175c6/extract/0.log" Jan 21 12:01:47 crc kubenswrapper[4824]: I0121 12:01:47.514143 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcwqkxq_d9005d95-8246-435d-aaea-5c60d58175c6/util/0.log" Jan 21 12:01:47 crc kubenswrapper[4824]: I0121 12:01:47.521256 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcwqkxq_d9005d95-8246-435d-aaea-5c60d58175c6/pull/0.log" Jan 21 12:01:47 crc kubenswrapper[4824]: I0121 12:01:47.529505 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713g7gj4_d2dd714a-8d35-49a1-b6c2-7bec802eed41/extract/0.log" Jan 21 12:01:47 crc kubenswrapper[4824]: I0121 12:01:47.537776 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713g7gj4_d2dd714a-8d35-49a1-b6c2-7bec802eed41/util/0.log" Jan 21 12:01:47 crc kubenswrapper[4824]: I0121 12:01:47.544123 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713g7gj4_d2dd714a-8d35-49a1-b6c2-7bec802eed41/pull/0.log" Jan 21 12:01:47 crc kubenswrapper[4824]: I0121 12:01:47.868402 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-rgl7w_c321fe0f-f966-4865-a619-6f5f26e9bb50/registry-server/0.log" Jan 21 12:01:47 crc kubenswrapper[4824]: I0121 12:01:47.873301 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-rgl7w_c321fe0f-f966-4865-a619-6f5f26e9bb50/extract-utilities/0.log" Jan 21 12:01:47 crc kubenswrapper[4824]: I0121 12:01:47.881060 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-rgl7w_c321fe0f-f966-4865-a619-6f5f26e9bb50/extract-content/0.log" Jan 21 12:01:48 crc kubenswrapper[4824]: I0121 12:01:48.054237 4824 scope.go:117] "RemoveContainer" containerID="753267d7f07800b4c7939c46da56af707c5831b8615d9de85460b5d273fc051a" Jan 21 12:01:48 crc kubenswrapper[4824]: E0121 12:01:48.054591 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 12:01:48 crc kubenswrapper[4824]: I0121 12:01:48.320057 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-m4srg_53e1053f-f40b-4afc-ad9c-b07d46fc1f94/registry-server/0.log" Jan 21 12:01:48 crc kubenswrapper[4824]: I0121 12:01:48.325123 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-m4srg_53e1053f-f40b-4afc-ad9c-b07d46fc1f94/extract-utilities/0.log" Jan 21 12:01:48 crc kubenswrapper[4824]: I0121 12:01:48.331470 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-m4srg_53e1053f-f40b-4afc-ad9c-b07d46fc1f94/extract-content/0.log" Jan 21 12:01:48 crc kubenswrapper[4824]: I0121 12:01:48.347829 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-xcxs9_b4c29122-d9b2-4f85-8b29-c0e7d888667f/marketplace-operator/0.log" Jan 21 12:01:48 crc kubenswrapper[4824]: I0121 12:01:48.456228 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-k2t4f_46423d25-afc7-4a50-ba21-5f7c7b7012f2/registry-server/0.log" Jan 21 12:01:48 crc kubenswrapper[4824]: I0121 12:01:48.460487 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-k2t4f_46423d25-afc7-4a50-ba21-5f7c7b7012f2/extract-utilities/0.log" Jan 21 12:01:48 crc kubenswrapper[4824]: I0121 12:01:48.465951 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-k2t4f_46423d25-afc7-4a50-ba21-5f7c7b7012f2/extract-content/0.log" Jan 21 12:01:48 crc kubenswrapper[4824]: I0121 12:01:48.924605 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-cmbbh_3447df50-807b-487c-9acf-2a2250780bdc/registry-server/0.log" Jan 21 12:01:48 crc kubenswrapper[4824]: I0121 12:01:48.929259 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-cmbbh_3447df50-807b-487c-9acf-2a2250780bdc/extract-utilities/0.log" Jan 21 12:01:48 crc kubenswrapper[4824]: I0121 12:01:48.936054 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-cmbbh_3447df50-807b-487c-9acf-2a2250780bdc/extract-content/0.log" Jan 21 12:02:02 crc kubenswrapper[4824]: I0121 12:02:02.049885 4824 scope.go:117] "RemoveContainer" containerID="753267d7f07800b4c7939c46da56af707c5831b8615d9de85460b5d273fc051a" Jan 21 12:02:02 crc kubenswrapper[4824]: E0121 12:02:02.050850 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 12:02:16 crc kubenswrapper[4824]: I0121 12:02:16.049239 4824 scope.go:117] "RemoveContainer" containerID="753267d7f07800b4c7939c46da56af707c5831b8615d9de85460b5d273fc051a" Jan 21 12:02:16 crc kubenswrapper[4824]: E0121 12:02:16.050104 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 12:02:29 crc kubenswrapper[4824]: I0121 12:02:29.049402 4824 scope.go:117] "RemoveContainer" containerID="753267d7f07800b4c7939c46da56af707c5831b8615d9de85460b5d273fc051a" Jan 21 12:02:29 crc kubenswrapper[4824]: E0121 12:02:29.050416 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 12:02:42 crc kubenswrapper[4824]: I0121 12:02:42.050030 4824 scope.go:117] "RemoveContainer" containerID="753267d7f07800b4c7939c46da56af707c5831b8615d9de85460b5d273fc051a" Jan 21 12:02:42 crc kubenswrapper[4824]: E0121 12:02:42.050706 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 12:02:47 crc kubenswrapper[4824]: I0121 12:02:47.977836 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-6968d8fdc4-bbv7p_d30695b8-6f6a-48f0-88ed-9181484634b8/controller/0.log" Jan 21 12:02:47 crc kubenswrapper[4824]: I0121 12:02:47.984043 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-6968d8fdc4-bbv7p_d30695b8-6f6a-48f0-88ed-9181484634b8/kube-rbac-proxy/0.log" Jan 21 12:02:47 crc kubenswrapper[4824]: I0121 12:02:47.999206 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9q9wj_e38576f0-c423-497a-a30d-60c2e8edd93a/controller/0.log" Jan 21 12:02:48 crc kubenswrapper[4824]: I0121 12:02:48.072728 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-858654f9db-qhlzr_b9405067-2dfa-46f9-9471-fe456b3ca5aa/cert-manager-controller/0.log" Jan 21 12:02:48 crc kubenswrapper[4824]: I0121 12:02:48.089884 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-cf98fcc89-xjwkp_d783bb77-baab-4e1c-99e9-ddba7ef78fc2/cert-manager-cainjector/0.log" Jan 21 12:02:48 crc kubenswrapper[4824]: I0121 12:02:48.098307 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-687f57d79b-r6qj7_b12640f9-bf75-4056-9c20-a75d63a0e177/cert-manager-webhook/0.log" Jan 21 12:02:49 crc kubenswrapper[4824]: I0121 12:02:49.079146 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7ddb5c749-bvptq_8ed6bb2c-2770-4785-bf38-9b6fd0bd8472/manager/0.log" Jan 21 12:02:49 crc kubenswrapper[4824]: I0121 12:02:49.142035 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-9b68f5989-24r54_a25733d4-a5f8-4c97-a897-8d96e637c253/manager/0.log" Jan 21 12:02:49 crc kubenswrapper[4824]: I0121 12:02:49.151939 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_db4b9aacd6f51b5aff2d66474158631eb1e58fb4d1358e39c8aaaca1bdn7cxr_72ab26cb-149e-4825-9587-3acfa66a368f/extract/0.log" Jan 21 12:02:49 crc kubenswrapper[4824]: I0121 12:02:49.164762 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_db4b9aacd6f51b5aff2d66474158631eb1e58fb4d1358e39c8aaaca1bdn7cxr_72ab26cb-149e-4825-9587-3acfa66a368f/util/0.log" Jan 21 12:02:49 crc kubenswrapper[4824]: I0121 12:02:49.178102 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_db4b9aacd6f51b5aff2d66474158631eb1e58fb4d1358e39c8aaaca1bdn7cxr_72ab26cb-149e-4825-9587-3acfa66a368f/pull/0.log" Jan 21 12:02:49 crc kubenswrapper[4824]: I0121 12:02:49.195640 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-9f958b845-wtfj4_14dd1f12-3943-42e4-be1b-c9e37e49d21b/manager/0.log" Jan 21 12:02:49 crc kubenswrapper[4824]: I0121 12:02:49.306470 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-c6994669c-hvq6s_b7053a68-ce89-48b6-9b5d-1c6dac813a75/manager/0.log" Jan 21 12:02:49 crc kubenswrapper[4824]: I0121 12:02:49.384378 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5fbc59b5b6-hgmf4_944c3123-9063-4731-9581-5b683741edc4/manager/0.log" Jan 21 12:02:49 crc kubenswrapper[4824]: I0121 12:02:49.392119 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-77d5c5b54f-mp5ml_2463afb1-0271-4ab3-ad1a-9329ecaeec4d/manager/0.log" Jan 21 12:02:49 crc kubenswrapper[4824]: I0121 12:02:49.532220 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9q9wj_e38576f0-c423-497a-a30d-60c2e8edd93a/frr/0.log" Jan 21 12:02:49 crc kubenswrapper[4824]: I0121 12:02:49.546064 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9q9wj_e38576f0-c423-497a-a30d-60c2e8edd93a/reloader/0.log" Jan 21 12:02:49 crc kubenswrapper[4824]: I0121 12:02:49.551994 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9q9wj_e38576f0-c423-497a-a30d-60c2e8edd93a/frr-metrics/0.log" Jan 21 12:02:49 crc kubenswrapper[4824]: I0121 12:02:49.561569 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9q9wj_e38576f0-c423-497a-a30d-60c2e8edd93a/kube-rbac-proxy/0.log" Jan 21 12:02:49 crc kubenswrapper[4824]: I0121 12:02:49.570214 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9q9wj_e38576f0-c423-497a-a30d-60c2e8edd93a/kube-rbac-proxy-frr/0.log" Jan 21 12:02:49 crc kubenswrapper[4824]: I0121 12:02:49.577929 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9q9wj_e38576f0-c423-497a-a30d-60c2e8edd93a/cp-frr-files/0.log" Jan 21 12:02:49 crc kubenswrapper[4824]: I0121 12:02:49.584410 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9q9wj_e38576f0-c423-497a-a30d-60c2e8edd93a/cp-reloader/0.log" Jan 21 12:02:49 crc kubenswrapper[4824]: I0121 12:02:49.600248 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9q9wj_e38576f0-c423-497a-a30d-60c2e8edd93a/cp-metrics/0.log" Jan 21 12:02:49 crc kubenswrapper[4824]: I0121 12:02:49.606333 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7df86c4f6c-fdblb_d71f993c-b2c7-47af-884f-4cbe6dcf3c1a/frr-k8s-webhook-server/0.log" Jan 21 12:02:49 crc kubenswrapper[4824]: I0121 12:02:49.639345 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-5c58bcfcff-69lct_714d2131-afc8-4828-840f-e2f0ce727d5d/manager/0.log" Jan 21 12:02:49 crc kubenswrapper[4824]: I0121 12:02:49.647385 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-68bc9bfcb6-hzmp5_81fed5ab-ec3d-46f0-8998-199393bffba0/webhook-server/0.log" Jan 21 12:02:49 crc kubenswrapper[4824]: I0121 12:02:49.717967 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-77c48c7859-dzn7w_57e1fe13-b2c8-4711-9efe-1ebbc7e9d12e/manager/0.log" Jan 21 12:02:49 crc kubenswrapper[4824]: I0121 12:02:49.733345 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-78757b4889-gkwjs_fc911701-7c15-4a20-b89e-709521bdd3a1/manager/0.log" Jan 21 12:02:49 crc kubenswrapper[4824]: I0121 12:02:49.831224 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-767fdc4f47-5mhmd_2e5efa9b-40bc-4860-967d-aa37e1838f81/manager/0.log" Jan 21 12:02:49 crc kubenswrapper[4824]: I0121 12:02:49.841673 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-864f6b75bf-c7vsf_18017a19-1676-4fe9-a97e-ae878f423847/manager/0.log" Jan 21 12:02:49 crc kubenswrapper[4824]: I0121 12:02:49.882737 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-c87fff755-d5hn9_19b1c150-f5ed-4432-bfc2-717b78eebc7e/manager/0.log" Jan 21 12:02:49 crc kubenswrapper[4824]: I0121 12:02:49.941247 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-cb4666565-7w8p4_8c74d3da-d870-45e9-96b9-70eb2395c977/manager/0.log" Jan 21 12:02:50 crc kubenswrapper[4824]: I0121 12:02:50.070277 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-65849867d6-8ngx6_b38ac60b-1643-44d9-a1fe-5f75129247dc/manager/0.log" Jan 21 12:02:50 crc kubenswrapper[4824]: I0121 12:02:50.079792 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-7fc9b76cf6-tncbx_5429da27-a618-44da-8aeb-12b917af8734/manager/0.log" Jan 21 12:02:50 crc kubenswrapper[4824]: I0121 12:02:50.093470 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-6b68b8b854zxcbl_3fa95966-3147-4e50-8ed8-f702a87992c4/manager/0.log" Jan 21 12:02:50 crc kubenswrapper[4824]: I0121 12:02:50.177377 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-8tsb9_baf5db51-d3ac-4b97-9c5f-afc598d59ab7/speaker/0.log" Jan 21 12:02:50 crc kubenswrapper[4824]: I0121 12:02:50.189323 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-8tsb9_baf5db51-d3ac-4b97-9c5f-afc598d59ab7/kube-rbac-proxy/0.log" Jan 21 12:02:50 crc kubenswrapper[4824]: I0121 12:02:50.273038 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-init-dbdbd6855-gkm5b_35334861-8e16-4b16-8cf5-167a049836af/operator/0.log" Jan 21 12:02:51 crc kubenswrapper[4824]: I0121 12:02:51.125814 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-858654f9db-qhlzr_b9405067-2dfa-46f9-9471-fe456b3ca5aa/cert-manager-controller/0.log" Jan 21 12:02:51 crc kubenswrapper[4824]: I0121 12:02:51.142338 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-cf98fcc89-xjwkp_d783bb77-baab-4e1c-99e9-ddba7ef78fc2/cert-manager-cainjector/0.log" Jan 21 12:02:51 crc kubenswrapper[4824]: I0121 12:02:51.148943 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-687f57d79b-r6qj7_b12640f9-bf75-4056-9c20-a75d63a0e177/cert-manager-webhook/0.log" Jan 21 12:02:51 crc kubenswrapper[4824]: I0121 12:02:51.399860 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-74d84f758c-vjgps_c1128e01-8b7c-4742-b1da-dbf06aa8b4d8/manager/0.log" Jan 21 12:02:51 crc kubenswrapper[4824]: I0121 12:02:51.413512 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-zggxm_78eb1906-d9e6-49f7-a512-ed3cc0a07da7/registry-server/0.log" Jan 21 12:02:51 crc kubenswrapper[4824]: I0121 12:02:51.468970 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-55db956ddc-78z2q_f5639f17-3a5b-4ef4-bd6e-6eeb3274636a/manager/0.log" Jan 21 12:02:51 crc kubenswrapper[4824]: I0121 12:02:51.494743 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-686df47fcb-c8xvs_d8b36106-8f21-4873-890d-c1879ea49068/manager/0.log" Jan 21 12:02:51 crc kubenswrapper[4824]: I0121 12:02:51.515692 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-94lqx_782093ab-751c-4e38-bba3-c5f76308c82e/operator/0.log" Jan 21 12:02:51 crc kubenswrapper[4824]: I0121 12:02:51.539930 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-85dd56d4cc-wlrdt_b4cb80e6-c720-4875-9848-7d9597a354e4/manager/0.log" Jan 21 12:02:51 crc kubenswrapper[4824]: I0121 12:02:51.595290 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-5f8f495fcf-6rz5h_db84b172-59d8-4995-815b-50b492283bee/manager/0.log" Jan 21 12:02:51 crc kubenswrapper[4824]: I0121 12:02:51.604565 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-7cd8bc9dbb-xkp8x_52491e7b-cd8c-403a-942c-312debd6f1c6/manager/0.log" Jan 21 12:02:51 crc kubenswrapper[4824]: I0121 12:02:51.617377 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-64cd966744-9w4ch_0715aa91-2ae5-4e07-8388-c37def4c7743/manager/0.log" Jan 21 12:02:51 crc kubenswrapper[4824]: I0121 12:02:51.804864 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-r65dm_c117483a-fe5c-423b-995e-9c2e7f59566a/control-plane-machine-set-operator/0.log" Jan 21 12:02:51 crc kubenswrapper[4824]: I0121 12:02:51.815813 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-7grz2_fb8bdb2b-bdc7-4433-be10-296fc67dc1c3/kube-rbac-proxy/0.log" Jan 21 12:02:51 crc kubenswrapper[4824]: I0121 12:02:51.823341 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-7grz2_fb8bdb2b-bdc7-4433-be10-296fc67dc1c3/machine-api-operator/0.log" Jan 21 12:02:52 crc kubenswrapper[4824]: I0121 12:02:52.548436 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7ddb5c749-bvptq_8ed6bb2c-2770-4785-bf38-9b6fd0bd8472/manager/0.log" Jan 21 12:02:52 crc kubenswrapper[4824]: I0121 12:02:52.578673 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-9b68f5989-24r54_a25733d4-a5f8-4c97-a897-8d96e637c253/manager/0.log" Jan 21 12:02:52 crc kubenswrapper[4824]: I0121 12:02:52.586872 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_db4b9aacd6f51b5aff2d66474158631eb1e58fb4d1358e39c8aaaca1bdn7cxr_72ab26cb-149e-4825-9587-3acfa66a368f/extract/0.log" Jan 21 12:02:52 crc kubenswrapper[4824]: I0121 12:02:52.593107 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_db4b9aacd6f51b5aff2d66474158631eb1e58fb4d1358e39c8aaaca1bdn7cxr_72ab26cb-149e-4825-9587-3acfa66a368f/util/0.log" Jan 21 12:02:52 crc kubenswrapper[4824]: I0121 12:02:52.601511 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_db4b9aacd6f51b5aff2d66474158631eb1e58fb4d1358e39c8aaaca1bdn7cxr_72ab26cb-149e-4825-9587-3acfa66a368f/pull/0.log" Jan 21 12:02:52 crc kubenswrapper[4824]: I0121 12:02:52.613141 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-9f958b845-wtfj4_14dd1f12-3943-42e4-be1b-c9e37e49d21b/manager/0.log" Jan 21 12:02:52 crc kubenswrapper[4824]: I0121 12:02:52.703207 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-c6994669c-hvq6s_b7053a68-ce89-48b6-9b5d-1c6dac813a75/manager/0.log" Jan 21 12:02:52 crc kubenswrapper[4824]: I0121 12:02:52.790599 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5fbc59b5b6-hgmf4_944c3123-9063-4731-9581-5b683741edc4/manager/0.log" Jan 21 12:02:52 crc kubenswrapper[4824]: I0121 12:02:52.800750 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-77d5c5b54f-mp5ml_2463afb1-0271-4ab3-ad1a-9329ecaeec4d/manager/0.log" Jan 21 12:02:52 crc kubenswrapper[4824]: I0121 12:02:52.801650 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-7754f76f8b-xjqs5_1575ab9d-016f-403a-990b-3b4c4d53f6f1/nmstate-console-plugin/0.log" Jan 21 12:02:52 crc kubenswrapper[4824]: I0121 12:02:52.824517 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-qjpv9_a01ecff0-9051-4ddb-8c8e-a4f379b6c823/nmstate-handler/0.log" Jan 21 12:02:52 crc kubenswrapper[4824]: I0121 12:02:52.833467 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-54757c584b-w2j6n_30b70686-bf05-4798-b5dc-3c7dc9cee56a/nmstate-metrics/0.log" Jan 21 12:02:52 crc kubenswrapper[4824]: I0121 12:02:52.840689 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-54757c584b-w2j6n_30b70686-bf05-4798-b5dc-3c7dc9cee56a/kube-rbac-proxy/0.log" Jan 21 12:02:52 crc kubenswrapper[4824]: I0121 12:02:52.855586 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-646758c888-zmc4k_df60ca53-0573-4a6a-9235-86e0a1c17f91/nmstate-operator/0.log" Jan 21 12:02:52 crc kubenswrapper[4824]: I0121 12:02:52.875637 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-8474b5b9d8-h6kkh_edf54909-b53e-44de-90bc-355e23bbc767/nmstate-webhook/0.log" Jan 21 12:02:53 crc kubenswrapper[4824]: I0121 12:02:53.057451 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-77c48c7859-dzn7w_57e1fe13-b2c8-4711-9efe-1ebbc7e9d12e/manager/0.log" Jan 21 12:02:53 crc kubenswrapper[4824]: I0121 12:02:53.068490 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-78757b4889-gkwjs_fc911701-7c15-4a20-b89e-709521bdd3a1/manager/0.log" Jan 21 12:02:53 crc kubenswrapper[4824]: I0121 12:02:53.126023 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-767fdc4f47-5mhmd_2e5efa9b-40bc-4860-967d-aa37e1838f81/manager/0.log" Jan 21 12:02:53 crc kubenswrapper[4824]: I0121 12:02:53.137350 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-864f6b75bf-c7vsf_18017a19-1676-4fe9-a97e-ae878f423847/manager/0.log" Jan 21 12:02:53 crc kubenswrapper[4824]: I0121 12:02:53.170507 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-c87fff755-d5hn9_19b1c150-f5ed-4432-bfc2-717b78eebc7e/manager/0.log" Jan 21 12:02:53 crc kubenswrapper[4824]: I0121 12:02:53.208261 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-cb4666565-7w8p4_8c74d3da-d870-45e9-96b9-70eb2395c977/manager/0.log" Jan 21 12:02:53 crc kubenswrapper[4824]: I0121 12:02:53.283860 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-65849867d6-8ngx6_b38ac60b-1643-44d9-a1fe-5f75129247dc/manager/0.log" Jan 21 12:02:53 crc kubenswrapper[4824]: I0121 12:02:53.296612 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-7fc9b76cf6-tncbx_5429da27-a618-44da-8aeb-12b917af8734/manager/0.log" Jan 21 12:02:53 crc kubenswrapper[4824]: I0121 12:02:53.312344 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-6b68b8b854zxcbl_3fa95966-3147-4e50-8ed8-f702a87992c4/manager/0.log" Jan 21 12:02:53 crc kubenswrapper[4824]: I0121 12:02:53.470838 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-init-dbdbd6855-gkm5b_35334861-8e16-4b16-8cf5-167a049836af/operator/0.log" Jan 21 12:02:54 crc kubenswrapper[4824]: I0121 12:02:54.061771 4824 scope.go:117] "RemoveContainer" containerID="753267d7f07800b4c7939c46da56af707c5831b8615d9de85460b5d273fc051a" Jan 21 12:02:54 crc kubenswrapper[4824]: E0121 12:02:54.062422 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 12:02:54 crc kubenswrapper[4824]: I0121 12:02:54.720502 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-74d84f758c-vjgps_c1128e01-8b7c-4742-b1da-dbf06aa8b4d8/manager/0.log" Jan 21 12:02:54 crc kubenswrapper[4824]: I0121 12:02:54.734072 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-zggxm_78eb1906-d9e6-49f7-a512-ed3cc0a07da7/registry-server/0.log" Jan 21 12:02:54 crc kubenswrapper[4824]: I0121 12:02:54.789840 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-55db956ddc-78z2q_f5639f17-3a5b-4ef4-bd6e-6eeb3274636a/manager/0.log" Jan 21 12:02:54 crc kubenswrapper[4824]: I0121 12:02:54.812317 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-686df47fcb-c8xvs_d8b36106-8f21-4873-890d-c1879ea49068/manager/0.log" Jan 21 12:02:54 crc kubenswrapper[4824]: I0121 12:02:54.829579 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-94lqx_782093ab-751c-4e38-bba3-c5f76308c82e/operator/0.log" Jan 21 12:02:54 crc kubenswrapper[4824]: I0121 12:02:54.850330 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-85dd56d4cc-wlrdt_b4cb80e6-c720-4875-9848-7d9597a354e4/manager/0.log" Jan 21 12:02:54 crc kubenswrapper[4824]: I0121 12:02:54.908385 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-5f8f495fcf-6rz5h_db84b172-59d8-4995-815b-50b492283bee/manager/0.log" Jan 21 12:02:54 crc kubenswrapper[4824]: I0121 12:02:54.919569 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-7cd8bc9dbb-xkp8x_52491e7b-cd8c-403a-942c-312debd6f1c6/manager/0.log" Jan 21 12:02:54 crc kubenswrapper[4824]: I0121 12:02:54.928319 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-64cd966744-9w4ch_0715aa91-2ae5-4e07-8388-c37def4c7743/manager/0.log" Jan 21 12:02:56 crc kubenswrapper[4824]: I0121 12:02:56.325700 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-additional-cni-plugins-gjtnh_3a8d2cac-2d0f-4894-9bc0-2abd81699c47/kube-multus-additional-cni-plugins/0.log" Jan 21 12:02:56 crc kubenswrapper[4824]: I0121 12:02:56.332998 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-additional-cni-plugins-gjtnh_3a8d2cac-2d0f-4894-9bc0-2abd81699c47/egress-router-binary-copy/0.log" Jan 21 12:02:56 crc kubenswrapper[4824]: I0121 12:02:56.340184 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-additional-cni-plugins-gjtnh_3a8d2cac-2d0f-4894-9bc0-2abd81699c47/cni-plugins/0.log" Jan 21 12:02:56 crc kubenswrapper[4824]: I0121 12:02:56.344861 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-additional-cni-plugins-gjtnh_3a8d2cac-2d0f-4894-9bc0-2abd81699c47/bond-cni-plugin/0.log" Jan 21 12:02:56 crc kubenswrapper[4824]: I0121 12:02:56.349947 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-additional-cni-plugins-gjtnh_3a8d2cac-2d0f-4894-9bc0-2abd81699c47/routeoverride-cni/0.log" Jan 21 12:02:56 crc kubenswrapper[4824]: I0121 12:02:56.355371 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-additional-cni-plugins-gjtnh_3a8d2cac-2d0f-4894-9bc0-2abd81699c47/whereabouts-cni-bincopy/0.log" Jan 21 12:02:56 crc kubenswrapper[4824]: I0121 12:02:56.361608 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-additional-cni-plugins-gjtnh_3a8d2cac-2d0f-4894-9bc0-2abd81699c47/whereabouts-cni/0.log" Jan 21 12:02:56 crc kubenswrapper[4824]: I0121 12:02:56.391859 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-admission-controller-857f4d67dd-27d2q_7352d717-b2a3-4cd2-9e35-3ae2fe2d7b34/multus-admission-controller/0.log" Jan 21 12:02:56 crc kubenswrapper[4824]: I0121 12:02:56.395772 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-admission-controller-857f4d67dd-27d2q_7352d717-b2a3-4cd2-9e35-3ae2fe2d7b34/kube-rbac-proxy/0.log" Jan 21 12:02:56 crc kubenswrapper[4824]: I0121 12:02:56.441794 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-x94nb_6b9588d1-5876-4c2c-8de5-e5c0ad46f04e/kube-multus/2.log" Jan 21 12:02:56 crc kubenswrapper[4824]: I0121 12:02:56.534438 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-x94nb_6b9588d1-5876-4c2c-8de5-e5c0ad46f04e/kube-multus/3.log" Jan 21 12:02:56 crc kubenswrapper[4824]: I0121 12:02:56.569925 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_network-metrics-daemon-9rmnw_8f9bd12b-22ee-44ed-a8b9-c100d043f691/network-metrics-daemon/0.log" Jan 21 12:02:56 crc kubenswrapper[4824]: I0121 12:02:56.575263 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_network-metrics-daemon-9rmnw_8f9bd12b-22ee-44ed-a8b9-c100d043f691/kube-rbac-proxy/0.log" Jan 21 12:03:08 crc kubenswrapper[4824]: I0121 12:03:08.056229 4824 scope.go:117] "RemoveContainer" containerID="753267d7f07800b4c7939c46da56af707c5831b8615d9de85460b5d273fc051a" Jan 21 12:03:08 crc kubenswrapper[4824]: E0121 12:03:08.057375 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 12:03:22 crc kubenswrapper[4824]: I0121 12:03:22.049888 4824 scope.go:117] "RemoveContainer" containerID="753267d7f07800b4c7939c46da56af707c5831b8615d9de85460b5d273fc051a" Jan 21 12:03:22 crc kubenswrapper[4824]: E0121 12:03:22.050849 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 12:03:37 crc kubenswrapper[4824]: I0121 12:03:37.048900 4824 scope.go:117] "RemoveContainer" containerID="753267d7f07800b4c7939c46da56af707c5831b8615d9de85460b5d273fc051a" Jan 21 12:03:37 crc kubenswrapper[4824]: E0121 12:03:37.049698 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 12:03:40 crc kubenswrapper[4824]: I0121 12:03:40.179810 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-rv2rv"] Jan 21 12:03:40 crc kubenswrapper[4824]: E0121 12:03:40.180672 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad457817-57db-426b-8191-f35e1d82d266" containerName="container-00" Jan 21 12:03:40 crc kubenswrapper[4824]: I0121 12:03:40.180685 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad457817-57db-426b-8191-f35e1d82d266" containerName="container-00" Jan 21 12:03:40 crc kubenswrapper[4824]: I0121 12:03:40.180884 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="ad457817-57db-426b-8191-f35e1d82d266" containerName="container-00" Jan 21 12:03:40 crc kubenswrapper[4824]: I0121 12:03:40.186533 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rv2rv" Jan 21 12:03:40 crc kubenswrapper[4824]: I0121 12:03:40.195091 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-rv2rv"] Jan 21 12:03:40 crc kubenswrapper[4824]: I0121 12:03:40.368500 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f22ac107-20fa-4273-9c3a-0d0d41b26c8f-catalog-content\") pod \"certified-operators-rv2rv\" (UID: \"f22ac107-20fa-4273-9c3a-0d0d41b26c8f\") " pod="openshift-marketplace/certified-operators-rv2rv" Jan 21 12:03:40 crc kubenswrapper[4824]: I0121 12:03:40.368584 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f22ac107-20fa-4273-9c3a-0d0d41b26c8f-utilities\") pod \"certified-operators-rv2rv\" (UID: \"f22ac107-20fa-4273-9c3a-0d0d41b26c8f\") " pod="openshift-marketplace/certified-operators-rv2rv" Jan 21 12:03:40 crc kubenswrapper[4824]: I0121 12:03:40.369070 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nxf8c\" (UniqueName: \"kubernetes.io/projected/f22ac107-20fa-4273-9c3a-0d0d41b26c8f-kube-api-access-nxf8c\") pod \"certified-operators-rv2rv\" (UID: \"f22ac107-20fa-4273-9c3a-0d0d41b26c8f\") " pod="openshift-marketplace/certified-operators-rv2rv" Jan 21 12:03:40 crc kubenswrapper[4824]: I0121 12:03:40.471656 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nxf8c\" (UniqueName: \"kubernetes.io/projected/f22ac107-20fa-4273-9c3a-0d0d41b26c8f-kube-api-access-nxf8c\") pod \"certified-operators-rv2rv\" (UID: \"f22ac107-20fa-4273-9c3a-0d0d41b26c8f\") " pod="openshift-marketplace/certified-operators-rv2rv" Jan 21 12:03:40 crc kubenswrapper[4824]: I0121 12:03:40.471812 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f22ac107-20fa-4273-9c3a-0d0d41b26c8f-catalog-content\") pod \"certified-operators-rv2rv\" (UID: \"f22ac107-20fa-4273-9c3a-0d0d41b26c8f\") " pod="openshift-marketplace/certified-operators-rv2rv" Jan 21 12:03:40 crc kubenswrapper[4824]: I0121 12:03:40.471842 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f22ac107-20fa-4273-9c3a-0d0d41b26c8f-utilities\") pod \"certified-operators-rv2rv\" (UID: \"f22ac107-20fa-4273-9c3a-0d0d41b26c8f\") " pod="openshift-marketplace/certified-operators-rv2rv" Jan 21 12:03:40 crc kubenswrapper[4824]: I0121 12:03:40.472371 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f22ac107-20fa-4273-9c3a-0d0d41b26c8f-catalog-content\") pod \"certified-operators-rv2rv\" (UID: \"f22ac107-20fa-4273-9c3a-0d0d41b26c8f\") " pod="openshift-marketplace/certified-operators-rv2rv" Jan 21 12:03:40 crc kubenswrapper[4824]: I0121 12:03:40.472398 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f22ac107-20fa-4273-9c3a-0d0d41b26c8f-utilities\") pod \"certified-operators-rv2rv\" (UID: \"f22ac107-20fa-4273-9c3a-0d0d41b26c8f\") " pod="openshift-marketplace/certified-operators-rv2rv" Jan 21 12:03:40 crc kubenswrapper[4824]: I0121 12:03:40.492852 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nxf8c\" (UniqueName: \"kubernetes.io/projected/f22ac107-20fa-4273-9c3a-0d0d41b26c8f-kube-api-access-nxf8c\") pod \"certified-operators-rv2rv\" (UID: \"f22ac107-20fa-4273-9c3a-0d0d41b26c8f\") " pod="openshift-marketplace/certified-operators-rv2rv" Jan 21 12:03:40 crc kubenswrapper[4824]: I0121 12:03:40.510933 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rv2rv" Jan 21 12:03:41 crc kubenswrapper[4824]: I0121 12:03:41.066125 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-rv2rv"] Jan 21 12:03:41 crc kubenswrapper[4824]: I0121 12:03:41.089643 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rv2rv" event={"ID":"f22ac107-20fa-4273-9c3a-0d0d41b26c8f","Type":"ContainerStarted","Data":"6c39dd25e15891de0d35d968c22d1ed84039fb4585c5ab11eeb1d6baac169fee"} Jan 21 12:03:42 crc kubenswrapper[4824]: I0121 12:03:42.100254 4824 generic.go:334] "Generic (PLEG): container finished" podID="f22ac107-20fa-4273-9c3a-0d0d41b26c8f" containerID="717d8985b272c834d9fa000fa009eff0aa3da93a8eddabe335385e612d96acff" exitCode=0 Jan 21 12:03:42 crc kubenswrapper[4824]: I0121 12:03:42.100489 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rv2rv" event={"ID":"f22ac107-20fa-4273-9c3a-0d0d41b26c8f","Type":"ContainerDied","Data":"717d8985b272c834d9fa000fa009eff0aa3da93a8eddabe335385e612d96acff"} Jan 21 12:03:42 crc kubenswrapper[4824]: I0121 12:03:42.103252 4824 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 21 12:03:44 crc kubenswrapper[4824]: I0121 12:03:44.117362 4824 generic.go:334] "Generic (PLEG): container finished" podID="f22ac107-20fa-4273-9c3a-0d0d41b26c8f" containerID="3fc94aae5daf0492b01258d4b4f2021d184ea31ea13b186581b62f9bef9930e0" exitCode=0 Jan 21 12:03:44 crc kubenswrapper[4824]: I0121 12:03:44.117542 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rv2rv" event={"ID":"f22ac107-20fa-4273-9c3a-0d0d41b26c8f","Type":"ContainerDied","Data":"3fc94aae5daf0492b01258d4b4f2021d184ea31ea13b186581b62f9bef9930e0"} Jan 21 12:03:45 crc kubenswrapper[4824]: I0121 12:03:45.130612 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rv2rv" event={"ID":"f22ac107-20fa-4273-9c3a-0d0d41b26c8f","Type":"ContainerStarted","Data":"15aed7bba17be639243e74d722c208c8755f606cf6e0571fa9fcf764974f4e84"} Jan 21 12:03:45 crc kubenswrapper[4824]: I0121 12:03:45.171393 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-rv2rv" podStartSLOduration=2.616691652 podStartE2EDuration="5.171364388s" podCreationTimestamp="2026-01-21 12:03:40 +0000 UTC" firstStartedPulling="2026-01-21 12:03:42.10291993 +0000 UTC m=+3164.395949223" lastFinishedPulling="2026-01-21 12:03:44.657592667 +0000 UTC m=+3166.950621959" observedRunningTime="2026-01-21 12:03:45.169929714 +0000 UTC m=+3167.462959006" watchObservedRunningTime="2026-01-21 12:03:45.171364388 +0000 UTC m=+3167.464393671" Jan 21 12:03:50 crc kubenswrapper[4824]: I0121 12:03:50.048927 4824 scope.go:117] "RemoveContainer" containerID="753267d7f07800b4c7939c46da56af707c5831b8615d9de85460b5d273fc051a" Jan 21 12:03:50 crc kubenswrapper[4824]: E0121 12:03:50.050572 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 12:03:50 crc kubenswrapper[4824]: I0121 12:03:50.511303 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-rv2rv" Jan 21 12:03:50 crc kubenswrapper[4824]: I0121 12:03:50.511668 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-rv2rv" Jan 21 12:03:50 crc kubenswrapper[4824]: I0121 12:03:50.549330 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-rv2rv" Jan 21 12:03:51 crc kubenswrapper[4824]: I0121 12:03:51.233744 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-rv2rv" Jan 21 12:03:51 crc kubenswrapper[4824]: I0121 12:03:51.280156 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-rv2rv"] Jan 21 12:03:53 crc kubenswrapper[4824]: I0121 12:03:53.214717 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-rv2rv" podUID="f22ac107-20fa-4273-9c3a-0d0d41b26c8f" containerName="registry-server" containerID="cri-o://15aed7bba17be639243e74d722c208c8755f606cf6e0571fa9fcf764974f4e84" gracePeriod=2 Jan 21 12:03:53 crc kubenswrapper[4824]: I0121 12:03:53.644044 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rv2rv" Jan 21 12:03:53 crc kubenswrapper[4824]: I0121 12:03:53.821047 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nxf8c\" (UniqueName: \"kubernetes.io/projected/f22ac107-20fa-4273-9c3a-0d0d41b26c8f-kube-api-access-nxf8c\") pod \"f22ac107-20fa-4273-9c3a-0d0d41b26c8f\" (UID: \"f22ac107-20fa-4273-9c3a-0d0d41b26c8f\") " Jan 21 12:03:53 crc kubenswrapper[4824]: I0121 12:03:53.821142 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f22ac107-20fa-4273-9c3a-0d0d41b26c8f-utilities\") pod \"f22ac107-20fa-4273-9c3a-0d0d41b26c8f\" (UID: \"f22ac107-20fa-4273-9c3a-0d0d41b26c8f\") " Jan 21 12:03:53 crc kubenswrapper[4824]: I0121 12:03:53.821176 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f22ac107-20fa-4273-9c3a-0d0d41b26c8f-catalog-content\") pod \"f22ac107-20fa-4273-9c3a-0d0d41b26c8f\" (UID: \"f22ac107-20fa-4273-9c3a-0d0d41b26c8f\") " Jan 21 12:03:53 crc kubenswrapper[4824]: I0121 12:03:53.821810 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f22ac107-20fa-4273-9c3a-0d0d41b26c8f-utilities" (OuterVolumeSpecName: "utilities") pod "f22ac107-20fa-4273-9c3a-0d0d41b26c8f" (UID: "f22ac107-20fa-4273-9c3a-0d0d41b26c8f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 12:03:53 crc kubenswrapper[4824]: I0121 12:03:53.822511 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f22ac107-20fa-4273-9c3a-0d0d41b26c8f-utilities\") on node \"crc\" DevicePath \"\"" Jan 21 12:03:53 crc kubenswrapper[4824]: I0121 12:03:53.827867 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f22ac107-20fa-4273-9c3a-0d0d41b26c8f-kube-api-access-nxf8c" (OuterVolumeSpecName: "kube-api-access-nxf8c") pod "f22ac107-20fa-4273-9c3a-0d0d41b26c8f" (UID: "f22ac107-20fa-4273-9c3a-0d0d41b26c8f"). InnerVolumeSpecName "kube-api-access-nxf8c". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 12:03:53 crc kubenswrapper[4824]: I0121 12:03:53.857947 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f22ac107-20fa-4273-9c3a-0d0d41b26c8f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f22ac107-20fa-4273-9c3a-0d0d41b26c8f" (UID: "f22ac107-20fa-4273-9c3a-0d0d41b26c8f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 12:03:53 crc kubenswrapper[4824]: I0121 12:03:53.925640 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nxf8c\" (UniqueName: \"kubernetes.io/projected/f22ac107-20fa-4273-9c3a-0d0d41b26c8f-kube-api-access-nxf8c\") on node \"crc\" DevicePath \"\"" Jan 21 12:03:53 crc kubenswrapper[4824]: I0121 12:03:53.925756 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f22ac107-20fa-4273-9c3a-0d0d41b26c8f-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 21 12:03:54 crc kubenswrapper[4824]: I0121 12:03:54.224136 4824 generic.go:334] "Generic (PLEG): container finished" podID="f22ac107-20fa-4273-9c3a-0d0d41b26c8f" containerID="15aed7bba17be639243e74d722c208c8755f606cf6e0571fa9fcf764974f4e84" exitCode=0 Jan 21 12:03:54 crc kubenswrapper[4824]: I0121 12:03:54.224182 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rv2rv" event={"ID":"f22ac107-20fa-4273-9c3a-0d0d41b26c8f","Type":"ContainerDied","Data":"15aed7bba17be639243e74d722c208c8755f606cf6e0571fa9fcf764974f4e84"} Jan 21 12:03:54 crc kubenswrapper[4824]: I0121 12:03:54.224198 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rv2rv" Jan 21 12:03:54 crc kubenswrapper[4824]: I0121 12:03:54.224222 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rv2rv" event={"ID":"f22ac107-20fa-4273-9c3a-0d0d41b26c8f","Type":"ContainerDied","Data":"6c39dd25e15891de0d35d968c22d1ed84039fb4585c5ab11eeb1d6baac169fee"} Jan 21 12:03:54 crc kubenswrapper[4824]: I0121 12:03:54.224248 4824 scope.go:117] "RemoveContainer" containerID="15aed7bba17be639243e74d722c208c8755f606cf6e0571fa9fcf764974f4e84" Jan 21 12:03:54 crc kubenswrapper[4824]: I0121 12:03:54.251761 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-rv2rv"] Jan 21 12:03:54 crc kubenswrapper[4824]: I0121 12:03:54.252436 4824 scope.go:117] "RemoveContainer" containerID="3fc94aae5daf0492b01258d4b4f2021d184ea31ea13b186581b62f9bef9930e0" Jan 21 12:03:54 crc kubenswrapper[4824]: I0121 12:03:54.260926 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-rv2rv"] Jan 21 12:03:54 crc kubenswrapper[4824]: I0121 12:03:54.269995 4824 scope.go:117] "RemoveContainer" containerID="717d8985b272c834d9fa000fa009eff0aa3da93a8eddabe335385e612d96acff" Jan 21 12:03:54 crc kubenswrapper[4824]: I0121 12:03:54.307137 4824 scope.go:117] "RemoveContainer" containerID="15aed7bba17be639243e74d722c208c8755f606cf6e0571fa9fcf764974f4e84" Jan 21 12:03:54 crc kubenswrapper[4824]: E0121 12:03:54.307421 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"15aed7bba17be639243e74d722c208c8755f606cf6e0571fa9fcf764974f4e84\": container with ID starting with 15aed7bba17be639243e74d722c208c8755f606cf6e0571fa9fcf764974f4e84 not found: ID does not exist" containerID="15aed7bba17be639243e74d722c208c8755f606cf6e0571fa9fcf764974f4e84" Jan 21 12:03:54 crc kubenswrapper[4824]: I0121 12:03:54.307458 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"15aed7bba17be639243e74d722c208c8755f606cf6e0571fa9fcf764974f4e84"} err="failed to get container status \"15aed7bba17be639243e74d722c208c8755f606cf6e0571fa9fcf764974f4e84\": rpc error: code = NotFound desc = could not find container \"15aed7bba17be639243e74d722c208c8755f606cf6e0571fa9fcf764974f4e84\": container with ID starting with 15aed7bba17be639243e74d722c208c8755f606cf6e0571fa9fcf764974f4e84 not found: ID does not exist" Jan 21 12:03:54 crc kubenswrapper[4824]: I0121 12:03:54.307484 4824 scope.go:117] "RemoveContainer" containerID="3fc94aae5daf0492b01258d4b4f2021d184ea31ea13b186581b62f9bef9930e0" Jan 21 12:03:54 crc kubenswrapper[4824]: E0121 12:03:54.307742 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3fc94aae5daf0492b01258d4b4f2021d184ea31ea13b186581b62f9bef9930e0\": container with ID starting with 3fc94aae5daf0492b01258d4b4f2021d184ea31ea13b186581b62f9bef9930e0 not found: ID does not exist" containerID="3fc94aae5daf0492b01258d4b4f2021d184ea31ea13b186581b62f9bef9930e0" Jan 21 12:03:54 crc kubenswrapper[4824]: I0121 12:03:54.307764 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3fc94aae5daf0492b01258d4b4f2021d184ea31ea13b186581b62f9bef9930e0"} err="failed to get container status \"3fc94aae5daf0492b01258d4b4f2021d184ea31ea13b186581b62f9bef9930e0\": rpc error: code = NotFound desc = could not find container \"3fc94aae5daf0492b01258d4b4f2021d184ea31ea13b186581b62f9bef9930e0\": container with ID starting with 3fc94aae5daf0492b01258d4b4f2021d184ea31ea13b186581b62f9bef9930e0 not found: ID does not exist" Jan 21 12:03:54 crc kubenswrapper[4824]: I0121 12:03:54.307779 4824 scope.go:117] "RemoveContainer" containerID="717d8985b272c834d9fa000fa009eff0aa3da93a8eddabe335385e612d96acff" Jan 21 12:03:54 crc kubenswrapper[4824]: E0121 12:03:54.308030 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"717d8985b272c834d9fa000fa009eff0aa3da93a8eddabe335385e612d96acff\": container with ID starting with 717d8985b272c834d9fa000fa009eff0aa3da93a8eddabe335385e612d96acff not found: ID does not exist" containerID="717d8985b272c834d9fa000fa009eff0aa3da93a8eddabe335385e612d96acff" Jan 21 12:03:54 crc kubenswrapper[4824]: I0121 12:03:54.308060 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"717d8985b272c834d9fa000fa009eff0aa3da93a8eddabe335385e612d96acff"} err="failed to get container status \"717d8985b272c834d9fa000fa009eff0aa3da93a8eddabe335385e612d96acff\": rpc error: code = NotFound desc = could not find container \"717d8985b272c834d9fa000fa009eff0aa3da93a8eddabe335385e612d96acff\": container with ID starting with 717d8985b272c834d9fa000fa009eff0aa3da93a8eddabe335385e612d96acff not found: ID does not exist" Jan 21 12:03:56 crc kubenswrapper[4824]: I0121 12:03:56.060226 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f22ac107-20fa-4273-9c3a-0d0d41b26c8f" path="/var/lib/kubelet/pods/f22ac107-20fa-4273-9c3a-0d0d41b26c8f/volumes" Jan 21 12:04:05 crc kubenswrapper[4824]: I0121 12:04:05.050216 4824 scope.go:117] "RemoveContainer" containerID="753267d7f07800b4c7939c46da56af707c5831b8615d9de85460b5d273fc051a" Jan 21 12:04:05 crc kubenswrapper[4824]: E0121 12:04:05.051387 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 12:04:17 crc kubenswrapper[4824]: I0121 12:04:17.051047 4824 scope.go:117] "RemoveContainer" containerID="753267d7f07800b4c7939c46da56af707c5831b8615d9de85460b5d273fc051a" Jan 21 12:04:17 crc kubenswrapper[4824]: I0121 12:04:17.424743 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" event={"ID":"33f3d922-4ffe-409b-a49a-d88c85898260","Type":"ContainerStarted","Data":"253e918881780a912353868ed3027165c77d110c4811895ff906677458c0004d"} Jan 21 12:04:43 crc kubenswrapper[4824]: I0121 12:04:43.134761 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-hwgw5"] Jan 21 12:04:43 crc kubenswrapper[4824]: E0121 12:04:43.136315 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f22ac107-20fa-4273-9c3a-0d0d41b26c8f" containerName="registry-server" Jan 21 12:04:43 crc kubenswrapper[4824]: I0121 12:04:43.136336 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="f22ac107-20fa-4273-9c3a-0d0d41b26c8f" containerName="registry-server" Jan 21 12:04:43 crc kubenswrapper[4824]: E0121 12:04:43.136379 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f22ac107-20fa-4273-9c3a-0d0d41b26c8f" containerName="extract-utilities" Jan 21 12:04:43 crc kubenswrapper[4824]: I0121 12:04:43.136386 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="f22ac107-20fa-4273-9c3a-0d0d41b26c8f" containerName="extract-utilities" Jan 21 12:04:43 crc kubenswrapper[4824]: E0121 12:04:43.136399 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f22ac107-20fa-4273-9c3a-0d0d41b26c8f" containerName="extract-content" Jan 21 12:04:43 crc kubenswrapper[4824]: I0121 12:04:43.136405 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="f22ac107-20fa-4273-9c3a-0d0d41b26c8f" containerName="extract-content" Jan 21 12:04:43 crc kubenswrapper[4824]: I0121 12:04:43.136724 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="f22ac107-20fa-4273-9c3a-0d0d41b26c8f" containerName="registry-server" Jan 21 12:04:43 crc kubenswrapper[4824]: I0121 12:04:43.138977 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hwgw5" Jan 21 12:04:43 crc kubenswrapper[4824]: I0121 12:04:43.151590 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-hwgw5"] Jan 21 12:04:43 crc kubenswrapper[4824]: I0121 12:04:43.229541 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/92367417-426c-47d7-ba64-b54f204fa7eb-utilities\") pod \"redhat-operators-hwgw5\" (UID: \"92367417-426c-47d7-ba64-b54f204fa7eb\") " pod="openshift-marketplace/redhat-operators-hwgw5" Jan 21 12:04:43 crc kubenswrapper[4824]: I0121 12:04:43.229744 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8bkcw\" (UniqueName: \"kubernetes.io/projected/92367417-426c-47d7-ba64-b54f204fa7eb-kube-api-access-8bkcw\") pod \"redhat-operators-hwgw5\" (UID: \"92367417-426c-47d7-ba64-b54f204fa7eb\") " pod="openshift-marketplace/redhat-operators-hwgw5" Jan 21 12:04:43 crc kubenswrapper[4824]: I0121 12:04:43.229800 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/92367417-426c-47d7-ba64-b54f204fa7eb-catalog-content\") pod \"redhat-operators-hwgw5\" (UID: \"92367417-426c-47d7-ba64-b54f204fa7eb\") " pod="openshift-marketplace/redhat-operators-hwgw5" Jan 21 12:04:43 crc kubenswrapper[4824]: I0121 12:04:43.331708 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/92367417-426c-47d7-ba64-b54f204fa7eb-utilities\") pod \"redhat-operators-hwgw5\" (UID: \"92367417-426c-47d7-ba64-b54f204fa7eb\") " pod="openshift-marketplace/redhat-operators-hwgw5" Jan 21 12:04:43 crc kubenswrapper[4824]: I0121 12:04:43.331812 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8bkcw\" (UniqueName: \"kubernetes.io/projected/92367417-426c-47d7-ba64-b54f204fa7eb-kube-api-access-8bkcw\") pod \"redhat-operators-hwgw5\" (UID: \"92367417-426c-47d7-ba64-b54f204fa7eb\") " pod="openshift-marketplace/redhat-operators-hwgw5" Jan 21 12:04:43 crc kubenswrapper[4824]: I0121 12:04:43.331835 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/92367417-426c-47d7-ba64-b54f204fa7eb-catalog-content\") pod \"redhat-operators-hwgw5\" (UID: \"92367417-426c-47d7-ba64-b54f204fa7eb\") " pod="openshift-marketplace/redhat-operators-hwgw5" Jan 21 12:04:43 crc kubenswrapper[4824]: I0121 12:04:43.332299 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/92367417-426c-47d7-ba64-b54f204fa7eb-catalog-content\") pod \"redhat-operators-hwgw5\" (UID: \"92367417-426c-47d7-ba64-b54f204fa7eb\") " pod="openshift-marketplace/redhat-operators-hwgw5" Jan 21 12:04:43 crc kubenswrapper[4824]: I0121 12:04:43.332297 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/92367417-426c-47d7-ba64-b54f204fa7eb-utilities\") pod \"redhat-operators-hwgw5\" (UID: \"92367417-426c-47d7-ba64-b54f204fa7eb\") " pod="openshift-marketplace/redhat-operators-hwgw5" Jan 21 12:04:43 crc kubenswrapper[4824]: I0121 12:04:43.356848 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8bkcw\" (UniqueName: \"kubernetes.io/projected/92367417-426c-47d7-ba64-b54f204fa7eb-kube-api-access-8bkcw\") pod \"redhat-operators-hwgw5\" (UID: \"92367417-426c-47d7-ba64-b54f204fa7eb\") " pod="openshift-marketplace/redhat-operators-hwgw5" Jan 21 12:04:43 crc kubenswrapper[4824]: I0121 12:04:43.461514 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hwgw5" Jan 21 12:04:43 crc kubenswrapper[4824]: I0121 12:04:43.917690 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-hwgw5"] Jan 21 12:04:44 crc kubenswrapper[4824]: I0121 12:04:44.688036 4824 generic.go:334] "Generic (PLEG): container finished" podID="92367417-426c-47d7-ba64-b54f204fa7eb" containerID="bc811765056abef25fcd9c3d043eafc95537e1fc2d0196c69ab520c5f78f8a2a" exitCode=0 Jan 21 12:04:44 crc kubenswrapper[4824]: I0121 12:04:44.688156 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hwgw5" event={"ID":"92367417-426c-47d7-ba64-b54f204fa7eb","Type":"ContainerDied","Data":"bc811765056abef25fcd9c3d043eafc95537e1fc2d0196c69ab520c5f78f8a2a"} Jan 21 12:04:44 crc kubenswrapper[4824]: I0121 12:04:44.688426 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hwgw5" event={"ID":"92367417-426c-47d7-ba64-b54f204fa7eb","Type":"ContainerStarted","Data":"a7db3ead5f33a027665a2a78595a34e4bf12954ed323508aab571a54c6300caf"} Jan 21 12:04:45 crc kubenswrapper[4824]: I0121 12:04:45.698796 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hwgw5" event={"ID":"92367417-426c-47d7-ba64-b54f204fa7eb","Type":"ContainerStarted","Data":"c7077a590a8ba136b68618edff9dd3a84f819a8211f27e06004c0cea21dc802a"} Jan 21 12:04:48 crc kubenswrapper[4824]: I0121 12:04:48.731127 4824 generic.go:334] "Generic (PLEG): container finished" podID="92367417-426c-47d7-ba64-b54f204fa7eb" containerID="c7077a590a8ba136b68618edff9dd3a84f819a8211f27e06004c0cea21dc802a" exitCode=0 Jan 21 12:04:48 crc kubenswrapper[4824]: I0121 12:04:48.731321 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hwgw5" event={"ID":"92367417-426c-47d7-ba64-b54f204fa7eb","Type":"ContainerDied","Data":"c7077a590a8ba136b68618edff9dd3a84f819a8211f27e06004c0cea21dc802a"} Jan 21 12:04:49 crc kubenswrapper[4824]: I0121 12:04:49.744256 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hwgw5" event={"ID":"92367417-426c-47d7-ba64-b54f204fa7eb","Type":"ContainerStarted","Data":"09074918ab9646d24744b237b7d0c5b8e1d15db20d259b572120ffd50235b4d7"} Jan 21 12:04:49 crc kubenswrapper[4824]: I0121 12:04:49.770308 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-hwgw5" podStartSLOduration=2.078753089 podStartE2EDuration="6.770284038s" podCreationTimestamp="2026-01-21 12:04:43 +0000 UTC" firstStartedPulling="2026-01-21 12:04:44.690179002 +0000 UTC m=+3226.983208294" lastFinishedPulling="2026-01-21 12:04:49.38170995 +0000 UTC m=+3231.674739243" observedRunningTime="2026-01-21 12:04:49.760778578 +0000 UTC m=+3232.053807869" watchObservedRunningTime="2026-01-21 12:04:49.770284038 +0000 UTC m=+3232.063313330" Jan 21 12:04:53 crc kubenswrapper[4824]: I0121 12:04:53.462063 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-hwgw5" Jan 21 12:04:53 crc kubenswrapper[4824]: I0121 12:04:53.462782 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-hwgw5" Jan 21 12:04:54 crc kubenswrapper[4824]: I0121 12:04:54.496659 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-hwgw5" podUID="92367417-426c-47d7-ba64-b54f204fa7eb" containerName="registry-server" probeResult="failure" output=< Jan 21 12:04:54 crc kubenswrapper[4824]: timeout: failed to connect service ":50051" within 1s Jan 21 12:04:54 crc kubenswrapper[4824]: > Jan 21 12:05:03 crc kubenswrapper[4824]: I0121 12:05:03.504561 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-hwgw5" Jan 21 12:05:03 crc kubenswrapper[4824]: I0121 12:05:03.549614 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-hwgw5" Jan 21 12:05:03 crc kubenswrapper[4824]: I0121 12:05:03.746281 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-hwgw5"] Jan 21 12:05:04 crc kubenswrapper[4824]: I0121 12:05:04.870262 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-hwgw5" podUID="92367417-426c-47d7-ba64-b54f204fa7eb" containerName="registry-server" containerID="cri-o://09074918ab9646d24744b237b7d0c5b8e1d15db20d259b572120ffd50235b4d7" gracePeriod=2 Jan 21 12:05:05 crc kubenswrapper[4824]: I0121 12:05:05.301393 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hwgw5" Jan 21 12:05:05 crc kubenswrapper[4824]: I0121 12:05:05.495252 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8bkcw\" (UniqueName: \"kubernetes.io/projected/92367417-426c-47d7-ba64-b54f204fa7eb-kube-api-access-8bkcw\") pod \"92367417-426c-47d7-ba64-b54f204fa7eb\" (UID: \"92367417-426c-47d7-ba64-b54f204fa7eb\") " Jan 21 12:05:05 crc kubenswrapper[4824]: I0121 12:05:05.495355 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/92367417-426c-47d7-ba64-b54f204fa7eb-catalog-content\") pod \"92367417-426c-47d7-ba64-b54f204fa7eb\" (UID: \"92367417-426c-47d7-ba64-b54f204fa7eb\") " Jan 21 12:05:05 crc kubenswrapper[4824]: I0121 12:05:05.495388 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/92367417-426c-47d7-ba64-b54f204fa7eb-utilities\") pod \"92367417-426c-47d7-ba64-b54f204fa7eb\" (UID: \"92367417-426c-47d7-ba64-b54f204fa7eb\") " Jan 21 12:05:05 crc kubenswrapper[4824]: I0121 12:05:05.496262 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/92367417-426c-47d7-ba64-b54f204fa7eb-utilities" (OuterVolumeSpecName: "utilities") pod "92367417-426c-47d7-ba64-b54f204fa7eb" (UID: "92367417-426c-47d7-ba64-b54f204fa7eb"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 12:05:05 crc kubenswrapper[4824]: I0121 12:05:05.504037 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/92367417-426c-47d7-ba64-b54f204fa7eb-kube-api-access-8bkcw" (OuterVolumeSpecName: "kube-api-access-8bkcw") pod "92367417-426c-47d7-ba64-b54f204fa7eb" (UID: "92367417-426c-47d7-ba64-b54f204fa7eb"). InnerVolumeSpecName "kube-api-access-8bkcw". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 12:05:05 crc kubenswrapper[4824]: I0121 12:05:05.594475 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/92367417-426c-47d7-ba64-b54f204fa7eb-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "92367417-426c-47d7-ba64-b54f204fa7eb" (UID: "92367417-426c-47d7-ba64-b54f204fa7eb"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 12:05:05 crc kubenswrapper[4824]: I0121 12:05:05.598593 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8bkcw\" (UniqueName: \"kubernetes.io/projected/92367417-426c-47d7-ba64-b54f204fa7eb-kube-api-access-8bkcw\") on node \"crc\" DevicePath \"\"" Jan 21 12:05:05 crc kubenswrapper[4824]: I0121 12:05:05.598635 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/92367417-426c-47d7-ba64-b54f204fa7eb-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 21 12:05:05 crc kubenswrapper[4824]: I0121 12:05:05.598648 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/92367417-426c-47d7-ba64-b54f204fa7eb-utilities\") on node \"crc\" DevicePath \"\"" Jan 21 12:05:05 crc kubenswrapper[4824]: I0121 12:05:05.881414 4824 generic.go:334] "Generic (PLEG): container finished" podID="92367417-426c-47d7-ba64-b54f204fa7eb" containerID="09074918ab9646d24744b237b7d0c5b8e1d15db20d259b572120ffd50235b4d7" exitCode=0 Jan 21 12:05:05 crc kubenswrapper[4824]: I0121 12:05:05.881456 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hwgw5" event={"ID":"92367417-426c-47d7-ba64-b54f204fa7eb","Type":"ContainerDied","Data":"09074918ab9646d24744b237b7d0c5b8e1d15db20d259b572120ffd50235b4d7"} Jan 21 12:05:05 crc kubenswrapper[4824]: I0121 12:05:05.881523 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hwgw5" Jan 21 12:05:05 crc kubenswrapper[4824]: I0121 12:05:05.881550 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hwgw5" event={"ID":"92367417-426c-47d7-ba64-b54f204fa7eb","Type":"ContainerDied","Data":"a7db3ead5f33a027665a2a78595a34e4bf12954ed323508aab571a54c6300caf"} Jan 21 12:05:05 crc kubenswrapper[4824]: I0121 12:05:05.881600 4824 scope.go:117] "RemoveContainer" containerID="09074918ab9646d24744b237b7d0c5b8e1d15db20d259b572120ffd50235b4d7" Jan 21 12:05:05 crc kubenswrapper[4824]: I0121 12:05:05.934051 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-hwgw5"] Jan 21 12:05:05 crc kubenswrapper[4824]: I0121 12:05:05.936166 4824 scope.go:117] "RemoveContainer" containerID="c7077a590a8ba136b68618edff9dd3a84f819a8211f27e06004c0cea21dc802a" Jan 21 12:05:05 crc kubenswrapper[4824]: I0121 12:05:05.938098 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-hwgw5"] Jan 21 12:05:05 crc kubenswrapper[4824]: I0121 12:05:05.956148 4824 scope.go:117] "RemoveContainer" containerID="bc811765056abef25fcd9c3d043eafc95537e1fc2d0196c69ab520c5f78f8a2a" Jan 21 12:05:05 crc kubenswrapper[4824]: I0121 12:05:05.987473 4824 scope.go:117] "RemoveContainer" containerID="09074918ab9646d24744b237b7d0c5b8e1d15db20d259b572120ffd50235b4d7" Jan 21 12:05:05 crc kubenswrapper[4824]: E0121 12:05:05.987784 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"09074918ab9646d24744b237b7d0c5b8e1d15db20d259b572120ffd50235b4d7\": container with ID starting with 09074918ab9646d24744b237b7d0c5b8e1d15db20d259b572120ffd50235b4d7 not found: ID does not exist" containerID="09074918ab9646d24744b237b7d0c5b8e1d15db20d259b572120ffd50235b4d7" Jan 21 12:05:05 crc kubenswrapper[4824]: I0121 12:05:05.987814 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"09074918ab9646d24744b237b7d0c5b8e1d15db20d259b572120ffd50235b4d7"} err="failed to get container status \"09074918ab9646d24744b237b7d0c5b8e1d15db20d259b572120ffd50235b4d7\": rpc error: code = NotFound desc = could not find container \"09074918ab9646d24744b237b7d0c5b8e1d15db20d259b572120ffd50235b4d7\": container with ID starting with 09074918ab9646d24744b237b7d0c5b8e1d15db20d259b572120ffd50235b4d7 not found: ID does not exist" Jan 21 12:05:05 crc kubenswrapper[4824]: I0121 12:05:05.987837 4824 scope.go:117] "RemoveContainer" containerID="c7077a590a8ba136b68618edff9dd3a84f819a8211f27e06004c0cea21dc802a" Jan 21 12:05:05 crc kubenswrapper[4824]: E0121 12:05:05.988145 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c7077a590a8ba136b68618edff9dd3a84f819a8211f27e06004c0cea21dc802a\": container with ID starting with c7077a590a8ba136b68618edff9dd3a84f819a8211f27e06004c0cea21dc802a not found: ID does not exist" containerID="c7077a590a8ba136b68618edff9dd3a84f819a8211f27e06004c0cea21dc802a" Jan 21 12:05:05 crc kubenswrapper[4824]: I0121 12:05:05.988209 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c7077a590a8ba136b68618edff9dd3a84f819a8211f27e06004c0cea21dc802a"} err="failed to get container status \"c7077a590a8ba136b68618edff9dd3a84f819a8211f27e06004c0cea21dc802a\": rpc error: code = NotFound desc = could not find container \"c7077a590a8ba136b68618edff9dd3a84f819a8211f27e06004c0cea21dc802a\": container with ID starting with c7077a590a8ba136b68618edff9dd3a84f819a8211f27e06004c0cea21dc802a not found: ID does not exist" Jan 21 12:05:05 crc kubenswrapper[4824]: I0121 12:05:05.988256 4824 scope.go:117] "RemoveContainer" containerID="bc811765056abef25fcd9c3d043eafc95537e1fc2d0196c69ab520c5f78f8a2a" Jan 21 12:05:05 crc kubenswrapper[4824]: E0121 12:05:05.988522 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bc811765056abef25fcd9c3d043eafc95537e1fc2d0196c69ab520c5f78f8a2a\": container with ID starting with bc811765056abef25fcd9c3d043eafc95537e1fc2d0196c69ab520c5f78f8a2a not found: ID does not exist" containerID="bc811765056abef25fcd9c3d043eafc95537e1fc2d0196c69ab520c5f78f8a2a" Jan 21 12:05:05 crc kubenswrapper[4824]: I0121 12:05:05.988557 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bc811765056abef25fcd9c3d043eafc95537e1fc2d0196c69ab520c5f78f8a2a"} err="failed to get container status \"bc811765056abef25fcd9c3d043eafc95537e1fc2d0196c69ab520c5f78f8a2a\": rpc error: code = NotFound desc = could not find container \"bc811765056abef25fcd9c3d043eafc95537e1fc2d0196c69ab520c5f78f8a2a\": container with ID starting with bc811765056abef25fcd9c3d043eafc95537e1fc2d0196c69ab520c5f78f8a2a not found: ID does not exist" Jan 21 12:05:06 crc kubenswrapper[4824]: I0121 12:05:06.061058 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="92367417-426c-47d7-ba64-b54f204fa7eb" path="/var/lib/kubelet/pods/92367417-426c-47d7-ba64-b54f204fa7eb/volumes" Jan 21 12:06:46 crc kubenswrapper[4824]: I0121 12:06:46.065345 4824 patch_prober.go:28] interesting pod/machine-config-daemon-zqjg9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 21 12:06:46 crc kubenswrapper[4824]: I0121 12:06:46.067990 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 21 12:07:01 crc kubenswrapper[4824]: I0121 12:07:01.324702 4824 scope.go:117] "RemoveContainer" containerID="c3e2158290b1bee0a6e844d1cfc3b1ea1c8ca3e18fabfc983d0a39b05fe5af6d" Jan 21 12:07:16 crc kubenswrapper[4824]: I0121 12:07:16.064751 4824 patch_prober.go:28] interesting pod/machine-config-daemon-zqjg9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 21 12:07:16 crc kubenswrapper[4824]: I0121 12:07:16.065477 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 21 12:07:46 crc kubenswrapper[4824]: I0121 12:07:46.065263 4824 patch_prober.go:28] interesting pod/machine-config-daemon-zqjg9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 21 12:07:46 crc kubenswrapper[4824]: I0121 12:07:46.065835 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 21 12:07:46 crc kubenswrapper[4824]: I0121 12:07:46.065904 4824 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" Jan 21 12:07:46 crc kubenswrapper[4824]: I0121 12:07:46.066806 4824 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"253e918881780a912353868ed3027165c77d110c4811895ff906677458c0004d"} pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 21 12:07:46 crc kubenswrapper[4824]: I0121 12:07:46.066875 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" containerName="machine-config-daemon" containerID="cri-o://253e918881780a912353868ed3027165c77d110c4811895ff906677458c0004d" gracePeriod=600 Jan 21 12:07:46 crc kubenswrapper[4824]: I0121 12:07:46.357577 4824 generic.go:334] "Generic (PLEG): container finished" podID="33f3d922-4ffe-409b-a49a-d88c85898260" containerID="253e918881780a912353868ed3027165c77d110c4811895ff906677458c0004d" exitCode=0 Jan 21 12:07:46 crc kubenswrapper[4824]: I0121 12:07:46.357658 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" event={"ID":"33f3d922-4ffe-409b-a49a-d88c85898260","Type":"ContainerDied","Data":"253e918881780a912353868ed3027165c77d110c4811895ff906677458c0004d"} Jan 21 12:07:46 crc kubenswrapper[4824]: I0121 12:07:46.357891 4824 scope.go:117] "RemoveContainer" containerID="753267d7f07800b4c7939c46da56af707c5831b8615d9de85460b5d273fc051a" Jan 21 12:07:47 crc kubenswrapper[4824]: I0121 12:07:47.368122 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" event={"ID":"33f3d922-4ffe-409b-a49a-d88c85898260","Type":"ContainerStarted","Data":"719b0a497b4be25b481cf2d3456b6c515017599e2aeafd1f30d22bffc35f79d4"} Jan 21 12:07:52 crc kubenswrapper[4824]: I0121 12:07:52.145040 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-vznz5"] Jan 21 12:07:52 crc kubenswrapper[4824]: E0121 12:07:52.146500 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92367417-426c-47d7-ba64-b54f204fa7eb" containerName="extract-content" Jan 21 12:07:52 crc kubenswrapper[4824]: I0121 12:07:52.146521 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="92367417-426c-47d7-ba64-b54f204fa7eb" containerName="extract-content" Jan 21 12:07:52 crc kubenswrapper[4824]: E0121 12:07:52.146585 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92367417-426c-47d7-ba64-b54f204fa7eb" containerName="extract-utilities" Jan 21 12:07:52 crc kubenswrapper[4824]: I0121 12:07:52.146593 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="92367417-426c-47d7-ba64-b54f204fa7eb" containerName="extract-utilities" Jan 21 12:07:52 crc kubenswrapper[4824]: E0121 12:07:52.146614 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92367417-426c-47d7-ba64-b54f204fa7eb" containerName="registry-server" Jan 21 12:07:52 crc kubenswrapper[4824]: I0121 12:07:52.146624 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="92367417-426c-47d7-ba64-b54f204fa7eb" containerName="registry-server" Jan 21 12:07:52 crc kubenswrapper[4824]: I0121 12:07:52.147598 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="92367417-426c-47d7-ba64-b54f204fa7eb" containerName="registry-server" Jan 21 12:07:52 crc kubenswrapper[4824]: I0121 12:07:52.150057 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vznz5" Jan 21 12:07:52 crc kubenswrapper[4824]: I0121 12:07:52.155905 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-vznz5"] Jan 21 12:07:52 crc kubenswrapper[4824]: I0121 12:07:52.199265 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7d34ec22-f51b-4776-8342-80b16e3c41e0-catalog-content\") pod \"community-operators-vznz5\" (UID: \"7d34ec22-f51b-4776-8342-80b16e3c41e0\") " pod="openshift-marketplace/community-operators-vznz5" Jan 21 12:07:52 crc kubenswrapper[4824]: I0121 12:07:52.199935 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7d34ec22-f51b-4776-8342-80b16e3c41e0-utilities\") pod \"community-operators-vznz5\" (UID: \"7d34ec22-f51b-4776-8342-80b16e3c41e0\") " pod="openshift-marketplace/community-operators-vznz5" Jan 21 12:07:52 crc kubenswrapper[4824]: I0121 12:07:52.200204 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wf6dz\" (UniqueName: \"kubernetes.io/projected/7d34ec22-f51b-4776-8342-80b16e3c41e0-kube-api-access-wf6dz\") pod \"community-operators-vznz5\" (UID: \"7d34ec22-f51b-4776-8342-80b16e3c41e0\") " pod="openshift-marketplace/community-operators-vznz5" Jan 21 12:07:52 crc kubenswrapper[4824]: I0121 12:07:52.301591 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7d34ec22-f51b-4776-8342-80b16e3c41e0-utilities\") pod \"community-operators-vznz5\" (UID: \"7d34ec22-f51b-4776-8342-80b16e3c41e0\") " pod="openshift-marketplace/community-operators-vznz5" Jan 21 12:07:52 crc kubenswrapper[4824]: I0121 12:07:52.301937 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wf6dz\" (UniqueName: \"kubernetes.io/projected/7d34ec22-f51b-4776-8342-80b16e3c41e0-kube-api-access-wf6dz\") pod \"community-operators-vznz5\" (UID: \"7d34ec22-f51b-4776-8342-80b16e3c41e0\") " pod="openshift-marketplace/community-operators-vznz5" Jan 21 12:07:52 crc kubenswrapper[4824]: I0121 12:07:52.302315 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7d34ec22-f51b-4776-8342-80b16e3c41e0-catalog-content\") pod \"community-operators-vznz5\" (UID: \"7d34ec22-f51b-4776-8342-80b16e3c41e0\") " pod="openshift-marketplace/community-operators-vznz5" Jan 21 12:07:52 crc kubenswrapper[4824]: I0121 12:07:52.302400 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7d34ec22-f51b-4776-8342-80b16e3c41e0-utilities\") pod \"community-operators-vznz5\" (UID: \"7d34ec22-f51b-4776-8342-80b16e3c41e0\") " pod="openshift-marketplace/community-operators-vznz5" Jan 21 12:07:52 crc kubenswrapper[4824]: I0121 12:07:52.302637 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7d34ec22-f51b-4776-8342-80b16e3c41e0-catalog-content\") pod \"community-operators-vznz5\" (UID: \"7d34ec22-f51b-4776-8342-80b16e3c41e0\") " pod="openshift-marketplace/community-operators-vznz5" Jan 21 12:07:52 crc kubenswrapper[4824]: I0121 12:07:52.323336 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wf6dz\" (UniqueName: \"kubernetes.io/projected/7d34ec22-f51b-4776-8342-80b16e3c41e0-kube-api-access-wf6dz\") pod \"community-operators-vznz5\" (UID: \"7d34ec22-f51b-4776-8342-80b16e3c41e0\") " pod="openshift-marketplace/community-operators-vznz5" Jan 21 12:07:52 crc kubenswrapper[4824]: I0121 12:07:52.470811 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vznz5" Jan 21 12:07:52 crc kubenswrapper[4824]: I0121 12:07:52.895678 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-vznz5"] Jan 21 12:07:53 crc kubenswrapper[4824]: I0121 12:07:53.426011 4824 generic.go:334] "Generic (PLEG): container finished" podID="7d34ec22-f51b-4776-8342-80b16e3c41e0" containerID="5c31284c4e517abb0c2ed3aad2846f4e895b48e7107df0de48f76ce606ec9a59" exitCode=0 Jan 21 12:07:53 crc kubenswrapper[4824]: I0121 12:07:53.426206 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vznz5" event={"ID":"7d34ec22-f51b-4776-8342-80b16e3c41e0","Type":"ContainerDied","Data":"5c31284c4e517abb0c2ed3aad2846f4e895b48e7107df0de48f76ce606ec9a59"} Jan 21 12:07:53 crc kubenswrapper[4824]: I0121 12:07:53.427298 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vznz5" event={"ID":"7d34ec22-f51b-4776-8342-80b16e3c41e0","Type":"ContainerStarted","Data":"736f0aba8b28cb81d6e6ab52fa2cf14cf8d30b070cc2c19d4e02723a2f9bfa89"} Jan 21 12:07:54 crc kubenswrapper[4824]: I0121 12:07:54.441519 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vznz5" event={"ID":"7d34ec22-f51b-4776-8342-80b16e3c41e0","Type":"ContainerStarted","Data":"23eb202bbcc162d4c110f855d0420222e9ba57029df847331b07595e4396c59a"} Jan 21 12:07:55 crc kubenswrapper[4824]: I0121 12:07:55.459083 4824 generic.go:334] "Generic (PLEG): container finished" podID="7d34ec22-f51b-4776-8342-80b16e3c41e0" containerID="23eb202bbcc162d4c110f855d0420222e9ba57029df847331b07595e4396c59a" exitCode=0 Jan 21 12:07:55 crc kubenswrapper[4824]: I0121 12:07:55.459282 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vznz5" event={"ID":"7d34ec22-f51b-4776-8342-80b16e3c41e0","Type":"ContainerDied","Data":"23eb202bbcc162d4c110f855d0420222e9ba57029df847331b07595e4396c59a"} Jan 21 12:07:56 crc kubenswrapper[4824]: I0121 12:07:56.470320 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vznz5" event={"ID":"7d34ec22-f51b-4776-8342-80b16e3c41e0","Type":"ContainerStarted","Data":"b21e44a96fa872806119c076184a66ebdeea97873f525518035d31b0fd37b6dd"} Jan 21 12:07:56 crc kubenswrapper[4824]: I0121 12:07:56.496205 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-vznz5" podStartSLOduration=1.980979414 podStartE2EDuration="4.496188228s" podCreationTimestamp="2026-01-21 12:07:52 +0000 UTC" firstStartedPulling="2026-01-21 12:07:53.42980227 +0000 UTC m=+3415.722831562" lastFinishedPulling="2026-01-21 12:07:55.945011085 +0000 UTC m=+3418.238040376" observedRunningTime="2026-01-21 12:07:56.490314173 +0000 UTC m=+3418.783343465" watchObservedRunningTime="2026-01-21 12:07:56.496188228 +0000 UTC m=+3418.789217520" Jan 21 12:08:01 crc kubenswrapper[4824]: I0121 12:08:01.406456 4824 scope.go:117] "RemoveContainer" containerID="c6793d88c2985e5e3d28f1236121b047a23b50c741fdeb44a8e3cc182cf06683" Jan 21 12:08:01 crc kubenswrapper[4824]: I0121 12:08:01.433743 4824 scope.go:117] "RemoveContainer" containerID="6e4471e366c5d457dbdb31676ebe2e10113437d05155002808726047c10699c0" Jan 21 12:08:02 crc kubenswrapper[4824]: I0121 12:08:02.471781 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-vznz5" Jan 21 12:08:02 crc kubenswrapper[4824]: I0121 12:08:02.472094 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-vznz5" Jan 21 12:08:02 crc kubenswrapper[4824]: I0121 12:08:02.509612 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-vznz5" Jan 21 12:08:02 crc kubenswrapper[4824]: I0121 12:08:02.575568 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-vznz5" Jan 21 12:08:02 crc kubenswrapper[4824]: I0121 12:08:02.747611 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-vznz5"] Jan 21 12:08:04 crc kubenswrapper[4824]: I0121 12:08:04.560205 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-vznz5" podUID="7d34ec22-f51b-4776-8342-80b16e3c41e0" containerName="registry-server" containerID="cri-o://b21e44a96fa872806119c076184a66ebdeea97873f525518035d31b0fd37b6dd" gracePeriod=2 Jan 21 12:08:04 crc kubenswrapper[4824]: I0121 12:08:04.962220 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vznz5" Jan 21 12:08:05 crc kubenswrapper[4824]: I0121 12:08:05.063324 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7d34ec22-f51b-4776-8342-80b16e3c41e0-utilities\") pod \"7d34ec22-f51b-4776-8342-80b16e3c41e0\" (UID: \"7d34ec22-f51b-4776-8342-80b16e3c41e0\") " Jan 21 12:08:05 crc kubenswrapper[4824]: I0121 12:08:05.064124 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7d34ec22-f51b-4776-8342-80b16e3c41e0-utilities" (OuterVolumeSpecName: "utilities") pod "7d34ec22-f51b-4776-8342-80b16e3c41e0" (UID: "7d34ec22-f51b-4776-8342-80b16e3c41e0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 12:08:05 crc kubenswrapper[4824]: I0121 12:08:05.082199 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7d34ec22-f51b-4776-8342-80b16e3c41e0-utilities\") on node \"crc\" DevicePath \"\"" Jan 21 12:08:05 crc kubenswrapper[4824]: I0121 12:08:05.183240 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wf6dz\" (UniqueName: \"kubernetes.io/projected/7d34ec22-f51b-4776-8342-80b16e3c41e0-kube-api-access-wf6dz\") pod \"7d34ec22-f51b-4776-8342-80b16e3c41e0\" (UID: \"7d34ec22-f51b-4776-8342-80b16e3c41e0\") " Jan 21 12:08:05 crc kubenswrapper[4824]: I0121 12:08:05.183388 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7d34ec22-f51b-4776-8342-80b16e3c41e0-catalog-content\") pod \"7d34ec22-f51b-4776-8342-80b16e3c41e0\" (UID: \"7d34ec22-f51b-4776-8342-80b16e3c41e0\") " Jan 21 12:08:05 crc kubenswrapper[4824]: I0121 12:08:05.193114 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7d34ec22-f51b-4776-8342-80b16e3c41e0-kube-api-access-wf6dz" (OuterVolumeSpecName: "kube-api-access-wf6dz") pod "7d34ec22-f51b-4776-8342-80b16e3c41e0" (UID: "7d34ec22-f51b-4776-8342-80b16e3c41e0"). InnerVolumeSpecName "kube-api-access-wf6dz". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 12:08:05 crc kubenswrapper[4824]: I0121 12:08:05.230409 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7d34ec22-f51b-4776-8342-80b16e3c41e0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7d34ec22-f51b-4776-8342-80b16e3c41e0" (UID: "7d34ec22-f51b-4776-8342-80b16e3c41e0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 12:08:05 crc kubenswrapper[4824]: I0121 12:08:05.288004 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wf6dz\" (UniqueName: \"kubernetes.io/projected/7d34ec22-f51b-4776-8342-80b16e3c41e0-kube-api-access-wf6dz\") on node \"crc\" DevicePath \"\"" Jan 21 12:08:05 crc kubenswrapper[4824]: I0121 12:08:05.288039 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7d34ec22-f51b-4776-8342-80b16e3c41e0-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 21 12:08:05 crc kubenswrapper[4824]: I0121 12:08:05.573793 4824 generic.go:334] "Generic (PLEG): container finished" podID="7d34ec22-f51b-4776-8342-80b16e3c41e0" containerID="b21e44a96fa872806119c076184a66ebdeea97873f525518035d31b0fd37b6dd" exitCode=0 Jan 21 12:08:05 crc kubenswrapper[4824]: I0121 12:08:05.573885 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vznz5" event={"ID":"7d34ec22-f51b-4776-8342-80b16e3c41e0","Type":"ContainerDied","Data":"b21e44a96fa872806119c076184a66ebdeea97873f525518035d31b0fd37b6dd"} Jan 21 12:08:05 crc kubenswrapper[4824]: I0121 12:08:05.573925 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vznz5" Jan 21 12:08:05 crc kubenswrapper[4824]: I0121 12:08:05.573967 4824 scope.go:117] "RemoveContainer" containerID="b21e44a96fa872806119c076184a66ebdeea97873f525518035d31b0fd37b6dd" Jan 21 12:08:05 crc kubenswrapper[4824]: I0121 12:08:05.573936 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vznz5" event={"ID":"7d34ec22-f51b-4776-8342-80b16e3c41e0","Type":"ContainerDied","Data":"736f0aba8b28cb81d6e6ab52fa2cf14cf8d30b070cc2c19d4e02723a2f9bfa89"} Jan 21 12:08:05 crc kubenswrapper[4824]: I0121 12:08:05.594453 4824 scope.go:117] "RemoveContainer" containerID="23eb202bbcc162d4c110f855d0420222e9ba57029df847331b07595e4396c59a" Jan 21 12:08:05 crc kubenswrapper[4824]: I0121 12:08:05.608229 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-vznz5"] Jan 21 12:08:05 crc kubenswrapper[4824]: I0121 12:08:05.615436 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-vznz5"] Jan 21 12:08:05 crc kubenswrapper[4824]: I0121 12:08:05.626129 4824 scope.go:117] "RemoveContainer" containerID="5c31284c4e517abb0c2ed3aad2846f4e895b48e7107df0de48f76ce606ec9a59" Jan 21 12:08:05 crc kubenswrapper[4824]: I0121 12:08:05.646463 4824 scope.go:117] "RemoveContainer" containerID="b21e44a96fa872806119c076184a66ebdeea97873f525518035d31b0fd37b6dd" Jan 21 12:08:05 crc kubenswrapper[4824]: E0121 12:08:05.646879 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b21e44a96fa872806119c076184a66ebdeea97873f525518035d31b0fd37b6dd\": container with ID starting with b21e44a96fa872806119c076184a66ebdeea97873f525518035d31b0fd37b6dd not found: ID does not exist" containerID="b21e44a96fa872806119c076184a66ebdeea97873f525518035d31b0fd37b6dd" Jan 21 12:08:05 crc kubenswrapper[4824]: I0121 12:08:05.646928 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b21e44a96fa872806119c076184a66ebdeea97873f525518035d31b0fd37b6dd"} err="failed to get container status \"b21e44a96fa872806119c076184a66ebdeea97873f525518035d31b0fd37b6dd\": rpc error: code = NotFound desc = could not find container \"b21e44a96fa872806119c076184a66ebdeea97873f525518035d31b0fd37b6dd\": container with ID starting with b21e44a96fa872806119c076184a66ebdeea97873f525518035d31b0fd37b6dd not found: ID does not exist" Jan 21 12:08:05 crc kubenswrapper[4824]: I0121 12:08:05.646973 4824 scope.go:117] "RemoveContainer" containerID="23eb202bbcc162d4c110f855d0420222e9ba57029df847331b07595e4396c59a" Jan 21 12:08:05 crc kubenswrapper[4824]: E0121 12:08:05.647284 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"23eb202bbcc162d4c110f855d0420222e9ba57029df847331b07595e4396c59a\": container with ID starting with 23eb202bbcc162d4c110f855d0420222e9ba57029df847331b07595e4396c59a not found: ID does not exist" containerID="23eb202bbcc162d4c110f855d0420222e9ba57029df847331b07595e4396c59a" Jan 21 12:08:05 crc kubenswrapper[4824]: I0121 12:08:05.647317 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"23eb202bbcc162d4c110f855d0420222e9ba57029df847331b07595e4396c59a"} err="failed to get container status \"23eb202bbcc162d4c110f855d0420222e9ba57029df847331b07595e4396c59a\": rpc error: code = NotFound desc = could not find container \"23eb202bbcc162d4c110f855d0420222e9ba57029df847331b07595e4396c59a\": container with ID starting with 23eb202bbcc162d4c110f855d0420222e9ba57029df847331b07595e4396c59a not found: ID does not exist" Jan 21 12:08:05 crc kubenswrapper[4824]: I0121 12:08:05.647343 4824 scope.go:117] "RemoveContainer" containerID="5c31284c4e517abb0c2ed3aad2846f4e895b48e7107df0de48f76ce606ec9a59" Jan 21 12:08:05 crc kubenswrapper[4824]: E0121 12:08:05.647836 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5c31284c4e517abb0c2ed3aad2846f4e895b48e7107df0de48f76ce606ec9a59\": container with ID starting with 5c31284c4e517abb0c2ed3aad2846f4e895b48e7107df0de48f76ce606ec9a59 not found: ID does not exist" containerID="5c31284c4e517abb0c2ed3aad2846f4e895b48e7107df0de48f76ce606ec9a59" Jan 21 12:08:05 crc kubenswrapper[4824]: I0121 12:08:05.647864 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5c31284c4e517abb0c2ed3aad2846f4e895b48e7107df0de48f76ce606ec9a59"} err="failed to get container status \"5c31284c4e517abb0c2ed3aad2846f4e895b48e7107df0de48f76ce606ec9a59\": rpc error: code = NotFound desc = could not find container \"5c31284c4e517abb0c2ed3aad2846f4e895b48e7107df0de48f76ce606ec9a59\": container with ID starting with 5c31284c4e517abb0c2ed3aad2846f4e895b48e7107df0de48f76ce606ec9a59 not found: ID does not exist" Jan 21 12:08:06 crc kubenswrapper[4824]: I0121 12:08:06.061846 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7d34ec22-f51b-4776-8342-80b16e3c41e0" path="/var/lib/kubelet/pods/7d34ec22-f51b-4776-8342-80b16e3c41e0/volumes" Jan 21 12:08:57 crc kubenswrapper[4824]: I0121 12:08:57.644891 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-45vjg"] Jan 21 12:08:57 crc kubenswrapper[4824]: E0121 12:08:57.645730 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d34ec22-f51b-4776-8342-80b16e3c41e0" containerName="extract-content" Jan 21 12:08:57 crc kubenswrapper[4824]: I0121 12:08:57.645746 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d34ec22-f51b-4776-8342-80b16e3c41e0" containerName="extract-content" Jan 21 12:08:57 crc kubenswrapper[4824]: E0121 12:08:57.645767 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d34ec22-f51b-4776-8342-80b16e3c41e0" containerName="extract-utilities" Jan 21 12:08:57 crc kubenswrapper[4824]: I0121 12:08:57.645774 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d34ec22-f51b-4776-8342-80b16e3c41e0" containerName="extract-utilities" Jan 21 12:08:57 crc kubenswrapper[4824]: E0121 12:08:57.645784 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d34ec22-f51b-4776-8342-80b16e3c41e0" containerName="registry-server" Jan 21 12:08:57 crc kubenswrapper[4824]: I0121 12:08:57.645790 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d34ec22-f51b-4776-8342-80b16e3c41e0" containerName="registry-server" Jan 21 12:08:57 crc kubenswrapper[4824]: I0121 12:08:57.646010 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="7d34ec22-f51b-4776-8342-80b16e3c41e0" containerName="registry-server" Jan 21 12:08:57 crc kubenswrapper[4824]: I0121 12:08:57.647292 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-45vjg" Jan 21 12:08:57 crc kubenswrapper[4824]: I0121 12:08:57.663315 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-45vjg"] Jan 21 12:08:57 crc kubenswrapper[4824]: I0121 12:08:57.692043 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rgzqh\" (UniqueName: \"kubernetes.io/projected/c37315cd-e029-42bf-bfeb-4e5c497833e4-kube-api-access-rgzqh\") pod \"redhat-marketplace-45vjg\" (UID: \"c37315cd-e029-42bf-bfeb-4e5c497833e4\") " pod="openshift-marketplace/redhat-marketplace-45vjg" Jan 21 12:08:57 crc kubenswrapper[4824]: I0121 12:08:57.692285 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c37315cd-e029-42bf-bfeb-4e5c497833e4-utilities\") pod \"redhat-marketplace-45vjg\" (UID: \"c37315cd-e029-42bf-bfeb-4e5c497833e4\") " pod="openshift-marketplace/redhat-marketplace-45vjg" Jan 21 12:08:57 crc kubenswrapper[4824]: I0121 12:08:57.692505 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c37315cd-e029-42bf-bfeb-4e5c497833e4-catalog-content\") pod \"redhat-marketplace-45vjg\" (UID: \"c37315cd-e029-42bf-bfeb-4e5c497833e4\") " pod="openshift-marketplace/redhat-marketplace-45vjg" Jan 21 12:08:57 crc kubenswrapper[4824]: I0121 12:08:57.794480 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rgzqh\" (UniqueName: \"kubernetes.io/projected/c37315cd-e029-42bf-bfeb-4e5c497833e4-kube-api-access-rgzqh\") pod \"redhat-marketplace-45vjg\" (UID: \"c37315cd-e029-42bf-bfeb-4e5c497833e4\") " pod="openshift-marketplace/redhat-marketplace-45vjg" Jan 21 12:08:57 crc kubenswrapper[4824]: I0121 12:08:57.794596 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c37315cd-e029-42bf-bfeb-4e5c497833e4-utilities\") pod \"redhat-marketplace-45vjg\" (UID: \"c37315cd-e029-42bf-bfeb-4e5c497833e4\") " pod="openshift-marketplace/redhat-marketplace-45vjg" Jan 21 12:08:57 crc kubenswrapper[4824]: I0121 12:08:57.794731 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c37315cd-e029-42bf-bfeb-4e5c497833e4-catalog-content\") pod \"redhat-marketplace-45vjg\" (UID: \"c37315cd-e029-42bf-bfeb-4e5c497833e4\") " pod="openshift-marketplace/redhat-marketplace-45vjg" Jan 21 12:08:57 crc kubenswrapper[4824]: I0121 12:08:57.795260 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c37315cd-e029-42bf-bfeb-4e5c497833e4-catalog-content\") pod \"redhat-marketplace-45vjg\" (UID: \"c37315cd-e029-42bf-bfeb-4e5c497833e4\") " pod="openshift-marketplace/redhat-marketplace-45vjg" Jan 21 12:08:57 crc kubenswrapper[4824]: I0121 12:08:57.795280 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c37315cd-e029-42bf-bfeb-4e5c497833e4-utilities\") pod \"redhat-marketplace-45vjg\" (UID: \"c37315cd-e029-42bf-bfeb-4e5c497833e4\") " pod="openshift-marketplace/redhat-marketplace-45vjg" Jan 21 12:08:57 crc kubenswrapper[4824]: I0121 12:08:57.811194 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rgzqh\" (UniqueName: \"kubernetes.io/projected/c37315cd-e029-42bf-bfeb-4e5c497833e4-kube-api-access-rgzqh\") pod \"redhat-marketplace-45vjg\" (UID: \"c37315cd-e029-42bf-bfeb-4e5c497833e4\") " pod="openshift-marketplace/redhat-marketplace-45vjg" Jan 21 12:08:57 crc kubenswrapper[4824]: I0121 12:08:57.969315 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-45vjg" Jan 21 12:08:58 crc kubenswrapper[4824]: I0121 12:08:58.445561 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-45vjg"] Jan 21 12:08:59 crc kubenswrapper[4824]: I0121 12:08:59.122104 4824 generic.go:334] "Generic (PLEG): container finished" podID="c37315cd-e029-42bf-bfeb-4e5c497833e4" containerID="f65363c6a723f7b36c886e60cfe89af054e651ee5287c8dcbfb73f6bcd5fec91" exitCode=0 Jan 21 12:08:59 crc kubenswrapper[4824]: I0121 12:08:59.122217 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-45vjg" event={"ID":"c37315cd-e029-42bf-bfeb-4e5c497833e4","Type":"ContainerDied","Data":"f65363c6a723f7b36c886e60cfe89af054e651ee5287c8dcbfb73f6bcd5fec91"} Jan 21 12:08:59 crc kubenswrapper[4824]: I0121 12:08:59.122498 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-45vjg" event={"ID":"c37315cd-e029-42bf-bfeb-4e5c497833e4","Type":"ContainerStarted","Data":"1d304bfa846d8ce7ddcda5a95d370eb81f9bbae489b9d7c8996d46294146e282"} Jan 21 12:08:59 crc kubenswrapper[4824]: I0121 12:08:59.125162 4824 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 21 12:08:59 crc kubenswrapper[4824]: I0121 12:08:59.125234 4824 generic.go:334] "Generic (PLEG): container finished" podID="2245f836-eda4-4bf5-aa8a-5b18de2927b2" containerID="a6e5e2a199134ec83203b77006aebfa54870b884b3f5b127c99a1be886340ffe" exitCode=0 Jan 21 12:08:59 crc kubenswrapper[4824]: I0121 12:08:59.125293 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-87fjw/must-gather-z5tfn" event={"ID":"2245f836-eda4-4bf5-aa8a-5b18de2927b2","Type":"ContainerDied","Data":"a6e5e2a199134ec83203b77006aebfa54870b884b3f5b127c99a1be886340ffe"} Jan 21 12:08:59 crc kubenswrapper[4824]: I0121 12:08:59.125835 4824 scope.go:117] "RemoveContainer" containerID="a6e5e2a199134ec83203b77006aebfa54870b884b3f5b127c99a1be886340ffe" Jan 21 12:08:59 crc kubenswrapper[4824]: I0121 12:08:59.664643 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-87fjw_must-gather-z5tfn_2245f836-eda4-4bf5-aa8a-5b18de2927b2/gather/0.log" Jan 21 12:09:00 crc kubenswrapper[4824]: I0121 12:09:00.151989 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-45vjg" event={"ID":"c37315cd-e029-42bf-bfeb-4e5c497833e4","Type":"ContainerStarted","Data":"9f0fb9cdda33e041182c2ae4cabcc86e02386278a06914d1a83dbb0ca2921abb"} Jan 21 12:09:01 crc kubenswrapper[4824]: I0121 12:09:01.162680 4824 generic.go:334] "Generic (PLEG): container finished" podID="c37315cd-e029-42bf-bfeb-4e5c497833e4" containerID="9f0fb9cdda33e041182c2ae4cabcc86e02386278a06914d1a83dbb0ca2921abb" exitCode=0 Jan 21 12:09:01 crc kubenswrapper[4824]: I0121 12:09:01.162776 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-45vjg" event={"ID":"c37315cd-e029-42bf-bfeb-4e5c497833e4","Type":"ContainerDied","Data":"9f0fb9cdda33e041182c2ae4cabcc86e02386278a06914d1a83dbb0ca2921abb"} Jan 21 12:09:02 crc kubenswrapper[4824]: I0121 12:09:02.173028 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-45vjg" event={"ID":"c37315cd-e029-42bf-bfeb-4e5c497833e4","Type":"ContainerStarted","Data":"703beb222722d1b6f5c66cca863d8d016718bb21d316dd2c21514c550c466cce"} Jan 21 12:09:02 crc kubenswrapper[4824]: I0121 12:09:02.196525 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-45vjg" podStartSLOduration=2.6812069149999997 podStartE2EDuration="5.196506392s" podCreationTimestamp="2026-01-21 12:08:57 +0000 UTC" firstStartedPulling="2026-01-21 12:08:59.124322213 +0000 UTC m=+3481.417351494" lastFinishedPulling="2026-01-21 12:09:01.639621679 +0000 UTC m=+3483.932650971" observedRunningTime="2026-01-21 12:09:02.187758589 +0000 UTC m=+3484.480787882" watchObservedRunningTime="2026-01-21 12:09:02.196506392 +0000 UTC m=+3484.489535684" Jan 21 12:09:07 crc kubenswrapper[4824]: I0121 12:09:07.332025 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-87fjw/must-gather-z5tfn"] Jan 21 12:09:07 crc kubenswrapper[4824]: I0121 12:09:07.333144 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-87fjw/must-gather-z5tfn" podUID="2245f836-eda4-4bf5-aa8a-5b18de2927b2" containerName="copy" containerID="cri-o://a7da4d1f02d49f551acb1d97262ed02dfab6791465333cc45925d14ffc0a4320" gracePeriod=2 Jan 21 12:09:07 crc kubenswrapper[4824]: I0121 12:09:07.342583 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-87fjw/must-gather-z5tfn"] Jan 21 12:09:07 crc kubenswrapper[4824]: I0121 12:09:07.743685 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-87fjw_must-gather-z5tfn_2245f836-eda4-4bf5-aa8a-5b18de2927b2/copy/0.log" Jan 21 12:09:07 crc kubenswrapper[4824]: I0121 12:09:07.744413 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-87fjw/must-gather-z5tfn" Jan 21 12:09:07 crc kubenswrapper[4824]: I0121 12:09:07.759925 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qjtv2\" (UniqueName: \"kubernetes.io/projected/2245f836-eda4-4bf5-aa8a-5b18de2927b2-kube-api-access-qjtv2\") pod \"2245f836-eda4-4bf5-aa8a-5b18de2927b2\" (UID: \"2245f836-eda4-4bf5-aa8a-5b18de2927b2\") " Jan 21 12:09:07 crc kubenswrapper[4824]: I0121 12:09:07.760280 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/2245f836-eda4-4bf5-aa8a-5b18de2927b2-must-gather-output\") pod \"2245f836-eda4-4bf5-aa8a-5b18de2927b2\" (UID: \"2245f836-eda4-4bf5-aa8a-5b18de2927b2\") " Jan 21 12:09:07 crc kubenswrapper[4824]: I0121 12:09:07.765744 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2245f836-eda4-4bf5-aa8a-5b18de2927b2-kube-api-access-qjtv2" (OuterVolumeSpecName: "kube-api-access-qjtv2") pod "2245f836-eda4-4bf5-aa8a-5b18de2927b2" (UID: "2245f836-eda4-4bf5-aa8a-5b18de2927b2"). InnerVolumeSpecName "kube-api-access-qjtv2". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 12:09:07 crc kubenswrapper[4824]: I0121 12:09:07.862626 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qjtv2\" (UniqueName: \"kubernetes.io/projected/2245f836-eda4-4bf5-aa8a-5b18de2927b2-kube-api-access-qjtv2\") on node \"crc\" DevicePath \"\"" Jan 21 12:09:07 crc kubenswrapper[4824]: I0121 12:09:07.912950 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2245f836-eda4-4bf5-aa8a-5b18de2927b2-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "2245f836-eda4-4bf5-aa8a-5b18de2927b2" (UID: "2245f836-eda4-4bf5-aa8a-5b18de2927b2"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 12:09:07 crc kubenswrapper[4824]: I0121 12:09:07.966757 4824 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/2245f836-eda4-4bf5-aa8a-5b18de2927b2-must-gather-output\") on node \"crc\" DevicePath \"\"" Jan 21 12:09:07 crc kubenswrapper[4824]: I0121 12:09:07.969500 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-45vjg" Jan 21 12:09:07 crc kubenswrapper[4824]: I0121 12:09:07.970152 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-45vjg" Jan 21 12:09:08 crc kubenswrapper[4824]: I0121 12:09:08.015071 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-45vjg" Jan 21 12:09:08 crc kubenswrapper[4824]: I0121 12:09:08.059376 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2245f836-eda4-4bf5-aa8a-5b18de2927b2" path="/var/lib/kubelet/pods/2245f836-eda4-4bf5-aa8a-5b18de2927b2/volumes" Jan 21 12:09:08 crc kubenswrapper[4824]: I0121 12:09:08.224882 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-87fjw_must-gather-z5tfn_2245f836-eda4-4bf5-aa8a-5b18de2927b2/copy/0.log" Jan 21 12:09:08 crc kubenswrapper[4824]: I0121 12:09:08.225212 4824 generic.go:334] "Generic (PLEG): container finished" podID="2245f836-eda4-4bf5-aa8a-5b18de2927b2" containerID="a7da4d1f02d49f551acb1d97262ed02dfab6791465333cc45925d14ffc0a4320" exitCode=143 Jan 21 12:09:08 crc kubenswrapper[4824]: I0121 12:09:08.225933 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-87fjw/must-gather-z5tfn" Jan 21 12:09:08 crc kubenswrapper[4824]: I0121 12:09:08.226315 4824 scope.go:117] "RemoveContainer" containerID="a7da4d1f02d49f551acb1d97262ed02dfab6791465333cc45925d14ffc0a4320" Jan 21 12:09:08 crc kubenswrapper[4824]: I0121 12:09:08.252555 4824 scope.go:117] "RemoveContainer" containerID="a6e5e2a199134ec83203b77006aebfa54870b884b3f5b127c99a1be886340ffe" Jan 21 12:09:08 crc kubenswrapper[4824]: I0121 12:09:08.269752 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-45vjg" Jan 21 12:09:08 crc kubenswrapper[4824]: I0121 12:09:08.315277 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-45vjg"] Jan 21 12:09:08 crc kubenswrapper[4824]: I0121 12:09:08.315372 4824 scope.go:117] "RemoveContainer" containerID="a7da4d1f02d49f551acb1d97262ed02dfab6791465333cc45925d14ffc0a4320" Jan 21 12:09:08 crc kubenswrapper[4824]: E0121 12:09:08.315784 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a7da4d1f02d49f551acb1d97262ed02dfab6791465333cc45925d14ffc0a4320\": container with ID starting with a7da4d1f02d49f551acb1d97262ed02dfab6791465333cc45925d14ffc0a4320 not found: ID does not exist" containerID="a7da4d1f02d49f551acb1d97262ed02dfab6791465333cc45925d14ffc0a4320" Jan 21 12:09:08 crc kubenswrapper[4824]: I0121 12:09:08.315813 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a7da4d1f02d49f551acb1d97262ed02dfab6791465333cc45925d14ffc0a4320"} err="failed to get container status \"a7da4d1f02d49f551acb1d97262ed02dfab6791465333cc45925d14ffc0a4320\": rpc error: code = NotFound desc = could not find container \"a7da4d1f02d49f551acb1d97262ed02dfab6791465333cc45925d14ffc0a4320\": container with ID starting with a7da4d1f02d49f551acb1d97262ed02dfab6791465333cc45925d14ffc0a4320 not found: ID does not exist" Jan 21 12:09:08 crc kubenswrapper[4824]: I0121 12:09:08.315833 4824 scope.go:117] "RemoveContainer" containerID="a6e5e2a199134ec83203b77006aebfa54870b884b3f5b127c99a1be886340ffe" Jan 21 12:09:08 crc kubenswrapper[4824]: E0121 12:09:08.316305 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a6e5e2a199134ec83203b77006aebfa54870b884b3f5b127c99a1be886340ffe\": container with ID starting with a6e5e2a199134ec83203b77006aebfa54870b884b3f5b127c99a1be886340ffe not found: ID does not exist" containerID="a6e5e2a199134ec83203b77006aebfa54870b884b3f5b127c99a1be886340ffe" Jan 21 12:09:08 crc kubenswrapper[4824]: I0121 12:09:08.316332 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a6e5e2a199134ec83203b77006aebfa54870b884b3f5b127c99a1be886340ffe"} err="failed to get container status \"a6e5e2a199134ec83203b77006aebfa54870b884b3f5b127c99a1be886340ffe\": rpc error: code = NotFound desc = could not find container \"a6e5e2a199134ec83203b77006aebfa54870b884b3f5b127c99a1be886340ffe\": container with ID starting with a6e5e2a199134ec83203b77006aebfa54870b884b3f5b127c99a1be886340ffe not found: ID does not exist" Jan 21 12:09:10 crc kubenswrapper[4824]: I0121 12:09:10.247896 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-45vjg" podUID="c37315cd-e029-42bf-bfeb-4e5c497833e4" containerName="registry-server" containerID="cri-o://703beb222722d1b6f5c66cca863d8d016718bb21d316dd2c21514c550c466cce" gracePeriod=2 Jan 21 12:09:10 crc kubenswrapper[4824]: I0121 12:09:10.620809 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-45vjg" Jan 21 12:09:10 crc kubenswrapper[4824]: I0121 12:09:10.722415 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c37315cd-e029-42bf-bfeb-4e5c497833e4-utilities\") pod \"c37315cd-e029-42bf-bfeb-4e5c497833e4\" (UID: \"c37315cd-e029-42bf-bfeb-4e5c497833e4\") " Jan 21 12:09:10 crc kubenswrapper[4824]: I0121 12:09:10.722659 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rgzqh\" (UniqueName: \"kubernetes.io/projected/c37315cd-e029-42bf-bfeb-4e5c497833e4-kube-api-access-rgzqh\") pod \"c37315cd-e029-42bf-bfeb-4e5c497833e4\" (UID: \"c37315cd-e029-42bf-bfeb-4e5c497833e4\") " Jan 21 12:09:10 crc kubenswrapper[4824]: I0121 12:09:10.722735 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c37315cd-e029-42bf-bfeb-4e5c497833e4-catalog-content\") pod \"c37315cd-e029-42bf-bfeb-4e5c497833e4\" (UID: \"c37315cd-e029-42bf-bfeb-4e5c497833e4\") " Jan 21 12:09:10 crc kubenswrapper[4824]: I0121 12:09:10.723202 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c37315cd-e029-42bf-bfeb-4e5c497833e4-utilities" (OuterVolumeSpecName: "utilities") pod "c37315cd-e029-42bf-bfeb-4e5c497833e4" (UID: "c37315cd-e029-42bf-bfeb-4e5c497833e4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 12:09:10 crc kubenswrapper[4824]: I0121 12:09:10.728098 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c37315cd-e029-42bf-bfeb-4e5c497833e4-kube-api-access-rgzqh" (OuterVolumeSpecName: "kube-api-access-rgzqh") pod "c37315cd-e029-42bf-bfeb-4e5c497833e4" (UID: "c37315cd-e029-42bf-bfeb-4e5c497833e4"). InnerVolumeSpecName "kube-api-access-rgzqh". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 12:09:10 crc kubenswrapper[4824]: I0121 12:09:10.740607 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c37315cd-e029-42bf-bfeb-4e5c497833e4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c37315cd-e029-42bf-bfeb-4e5c497833e4" (UID: "c37315cd-e029-42bf-bfeb-4e5c497833e4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 12:09:10 crc kubenswrapper[4824]: I0121 12:09:10.826184 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c37315cd-e029-42bf-bfeb-4e5c497833e4-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 21 12:09:10 crc kubenswrapper[4824]: I0121 12:09:10.826591 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c37315cd-e029-42bf-bfeb-4e5c497833e4-utilities\") on node \"crc\" DevicePath \"\"" Jan 21 12:09:10 crc kubenswrapper[4824]: I0121 12:09:10.826607 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rgzqh\" (UniqueName: \"kubernetes.io/projected/c37315cd-e029-42bf-bfeb-4e5c497833e4-kube-api-access-rgzqh\") on node \"crc\" DevicePath \"\"" Jan 21 12:09:11 crc kubenswrapper[4824]: I0121 12:09:11.259362 4824 generic.go:334] "Generic (PLEG): container finished" podID="c37315cd-e029-42bf-bfeb-4e5c497833e4" containerID="703beb222722d1b6f5c66cca863d8d016718bb21d316dd2c21514c550c466cce" exitCode=0 Jan 21 12:09:11 crc kubenswrapper[4824]: I0121 12:09:11.259428 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-45vjg" event={"ID":"c37315cd-e029-42bf-bfeb-4e5c497833e4","Type":"ContainerDied","Data":"703beb222722d1b6f5c66cca863d8d016718bb21d316dd2c21514c550c466cce"} Jan 21 12:09:11 crc kubenswrapper[4824]: I0121 12:09:11.259476 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-45vjg" event={"ID":"c37315cd-e029-42bf-bfeb-4e5c497833e4","Type":"ContainerDied","Data":"1d304bfa846d8ce7ddcda5a95d370eb81f9bbae489b9d7c8996d46294146e282"} Jan 21 12:09:11 crc kubenswrapper[4824]: I0121 12:09:11.259478 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-45vjg" Jan 21 12:09:11 crc kubenswrapper[4824]: I0121 12:09:11.259500 4824 scope.go:117] "RemoveContainer" containerID="703beb222722d1b6f5c66cca863d8d016718bb21d316dd2c21514c550c466cce" Jan 21 12:09:11 crc kubenswrapper[4824]: I0121 12:09:11.291635 4824 scope.go:117] "RemoveContainer" containerID="9f0fb9cdda33e041182c2ae4cabcc86e02386278a06914d1a83dbb0ca2921abb" Jan 21 12:09:11 crc kubenswrapper[4824]: I0121 12:09:11.295732 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-45vjg"] Jan 21 12:09:11 crc kubenswrapper[4824]: I0121 12:09:11.314492 4824 scope.go:117] "RemoveContainer" containerID="f65363c6a723f7b36c886e60cfe89af054e651ee5287c8dcbfb73f6bcd5fec91" Jan 21 12:09:11 crc kubenswrapper[4824]: I0121 12:09:11.317826 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-45vjg"] Jan 21 12:09:11 crc kubenswrapper[4824]: I0121 12:09:11.353215 4824 scope.go:117] "RemoveContainer" containerID="703beb222722d1b6f5c66cca863d8d016718bb21d316dd2c21514c550c466cce" Jan 21 12:09:11 crc kubenswrapper[4824]: E0121 12:09:11.353723 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"703beb222722d1b6f5c66cca863d8d016718bb21d316dd2c21514c550c466cce\": container with ID starting with 703beb222722d1b6f5c66cca863d8d016718bb21d316dd2c21514c550c466cce not found: ID does not exist" containerID="703beb222722d1b6f5c66cca863d8d016718bb21d316dd2c21514c550c466cce" Jan 21 12:09:11 crc kubenswrapper[4824]: I0121 12:09:11.353770 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"703beb222722d1b6f5c66cca863d8d016718bb21d316dd2c21514c550c466cce"} err="failed to get container status \"703beb222722d1b6f5c66cca863d8d016718bb21d316dd2c21514c550c466cce\": rpc error: code = NotFound desc = could not find container \"703beb222722d1b6f5c66cca863d8d016718bb21d316dd2c21514c550c466cce\": container with ID starting with 703beb222722d1b6f5c66cca863d8d016718bb21d316dd2c21514c550c466cce not found: ID does not exist" Jan 21 12:09:11 crc kubenswrapper[4824]: I0121 12:09:11.353802 4824 scope.go:117] "RemoveContainer" containerID="9f0fb9cdda33e041182c2ae4cabcc86e02386278a06914d1a83dbb0ca2921abb" Jan 21 12:09:11 crc kubenswrapper[4824]: E0121 12:09:11.354370 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9f0fb9cdda33e041182c2ae4cabcc86e02386278a06914d1a83dbb0ca2921abb\": container with ID starting with 9f0fb9cdda33e041182c2ae4cabcc86e02386278a06914d1a83dbb0ca2921abb not found: ID does not exist" containerID="9f0fb9cdda33e041182c2ae4cabcc86e02386278a06914d1a83dbb0ca2921abb" Jan 21 12:09:11 crc kubenswrapper[4824]: I0121 12:09:11.354415 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9f0fb9cdda33e041182c2ae4cabcc86e02386278a06914d1a83dbb0ca2921abb"} err="failed to get container status \"9f0fb9cdda33e041182c2ae4cabcc86e02386278a06914d1a83dbb0ca2921abb\": rpc error: code = NotFound desc = could not find container \"9f0fb9cdda33e041182c2ae4cabcc86e02386278a06914d1a83dbb0ca2921abb\": container with ID starting with 9f0fb9cdda33e041182c2ae4cabcc86e02386278a06914d1a83dbb0ca2921abb not found: ID does not exist" Jan 21 12:09:11 crc kubenswrapper[4824]: I0121 12:09:11.354454 4824 scope.go:117] "RemoveContainer" containerID="f65363c6a723f7b36c886e60cfe89af054e651ee5287c8dcbfb73f6bcd5fec91" Jan 21 12:09:11 crc kubenswrapper[4824]: E0121 12:09:11.354771 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f65363c6a723f7b36c886e60cfe89af054e651ee5287c8dcbfb73f6bcd5fec91\": container with ID starting with f65363c6a723f7b36c886e60cfe89af054e651ee5287c8dcbfb73f6bcd5fec91 not found: ID does not exist" containerID="f65363c6a723f7b36c886e60cfe89af054e651ee5287c8dcbfb73f6bcd5fec91" Jan 21 12:09:11 crc kubenswrapper[4824]: I0121 12:09:11.354797 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f65363c6a723f7b36c886e60cfe89af054e651ee5287c8dcbfb73f6bcd5fec91"} err="failed to get container status \"f65363c6a723f7b36c886e60cfe89af054e651ee5287c8dcbfb73f6bcd5fec91\": rpc error: code = NotFound desc = could not find container \"f65363c6a723f7b36c886e60cfe89af054e651ee5287c8dcbfb73f6bcd5fec91\": container with ID starting with f65363c6a723f7b36c886e60cfe89af054e651ee5287c8dcbfb73f6bcd5fec91 not found: ID does not exist" Jan 21 12:09:12 crc kubenswrapper[4824]: I0121 12:09:12.059714 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c37315cd-e029-42bf-bfeb-4e5c497833e4" path="/var/lib/kubelet/pods/c37315cd-e029-42bf-bfeb-4e5c497833e4/volumes" Jan 21 12:09:46 crc kubenswrapper[4824]: I0121 12:09:46.064975 4824 patch_prober.go:28] interesting pod/machine-config-daemon-zqjg9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 21 12:09:46 crc kubenswrapper[4824]: I0121 12:09:46.065694 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 21 12:10:16 crc kubenswrapper[4824]: I0121 12:10:16.065768 4824 patch_prober.go:28] interesting pod/machine-config-daemon-zqjg9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 21 12:10:16 crc kubenswrapper[4824]: I0121 12:10:16.066461 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 21 12:10:35 crc kubenswrapper[4824]: I0121 12:10:35.101234 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-n5vql/must-gather-jkhmz"] Jan 21 12:10:35 crc kubenswrapper[4824]: E0121 12:10:35.102236 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2245f836-eda4-4bf5-aa8a-5b18de2927b2" containerName="gather" Jan 21 12:10:35 crc kubenswrapper[4824]: I0121 12:10:35.102251 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="2245f836-eda4-4bf5-aa8a-5b18de2927b2" containerName="gather" Jan 21 12:10:35 crc kubenswrapper[4824]: E0121 12:10:35.102261 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c37315cd-e029-42bf-bfeb-4e5c497833e4" containerName="registry-server" Jan 21 12:10:35 crc kubenswrapper[4824]: I0121 12:10:35.102267 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="c37315cd-e029-42bf-bfeb-4e5c497833e4" containerName="registry-server" Jan 21 12:10:35 crc kubenswrapper[4824]: E0121 12:10:35.102278 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2245f836-eda4-4bf5-aa8a-5b18de2927b2" containerName="copy" Jan 21 12:10:35 crc kubenswrapper[4824]: I0121 12:10:35.102284 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="2245f836-eda4-4bf5-aa8a-5b18de2927b2" containerName="copy" Jan 21 12:10:35 crc kubenswrapper[4824]: E0121 12:10:35.102297 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c37315cd-e029-42bf-bfeb-4e5c497833e4" containerName="extract-utilities" Jan 21 12:10:35 crc kubenswrapper[4824]: I0121 12:10:35.102303 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="c37315cd-e029-42bf-bfeb-4e5c497833e4" containerName="extract-utilities" Jan 21 12:10:35 crc kubenswrapper[4824]: E0121 12:10:35.102316 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c37315cd-e029-42bf-bfeb-4e5c497833e4" containerName="extract-content" Jan 21 12:10:35 crc kubenswrapper[4824]: I0121 12:10:35.102321 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="c37315cd-e029-42bf-bfeb-4e5c497833e4" containerName="extract-content" Jan 21 12:10:35 crc kubenswrapper[4824]: I0121 12:10:35.102511 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="c37315cd-e029-42bf-bfeb-4e5c497833e4" containerName="registry-server" Jan 21 12:10:35 crc kubenswrapper[4824]: I0121 12:10:35.102535 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="2245f836-eda4-4bf5-aa8a-5b18de2927b2" containerName="copy" Jan 21 12:10:35 crc kubenswrapper[4824]: I0121 12:10:35.102544 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="2245f836-eda4-4bf5-aa8a-5b18de2927b2" containerName="gather" Jan 21 12:10:35 crc kubenswrapper[4824]: I0121 12:10:35.103536 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-n5vql/must-gather-jkhmz" Jan 21 12:10:35 crc kubenswrapper[4824]: I0121 12:10:35.105433 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-n5vql"/"openshift-service-ca.crt" Jan 21 12:10:35 crc kubenswrapper[4824]: I0121 12:10:35.105686 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-n5vql"/"default-dockercfg-2zcwc" Jan 21 12:10:35 crc kubenswrapper[4824]: I0121 12:10:35.105815 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-n5vql"/"kube-root-ca.crt" Jan 21 12:10:35 crc kubenswrapper[4824]: I0121 12:10:35.119290 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-n5vql/must-gather-jkhmz"] Jan 21 12:10:35 crc kubenswrapper[4824]: I0121 12:10:35.170267 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fdgjw\" (UniqueName: \"kubernetes.io/projected/e1c80411-6f38-43b2-8656-639978632741-kube-api-access-fdgjw\") pod \"must-gather-jkhmz\" (UID: \"e1c80411-6f38-43b2-8656-639978632741\") " pod="openshift-must-gather-n5vql/must-gather-jkhmz" Jan 21 12:10:35 crc kubenswrapper[4824]: I0121 12:10:35.170378 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/e1c80411-6f38-43b2-8656-639978632741-must-gather-output\") pod \"must-gather-jkhmz\" (UID: \"e1c80411-6f38-43b2-8656-639978632741\") " pod="openshift-must-gather-n5vql/must-gather-jkhmz" Jan 21 12:10:35 crc kubenswrapper[4824]: I0121 12:10:35.272892 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/e1c80411-6f38-43b2-8656-639978632741-must-gather-output\") pod \"must-gather-jkhmz\" (UID: \"e1c80411-6f38-43b2-8656-639978632741\") " pod="openshift-must-gather-n5vql/must-gather-jkhmz" Jan 21 12:10:35 crc kubenswrapper[4824]: I0121 12:10:35.273133 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fdgjw\" (UniqueName: \"kubernetes.io/projected/e1c80411-6f38-43b2-8656-639978632741-kube-api-access-fdgjw\") pod \"must-gather-jkhmz\" (UID: \"e1c80411-6f38-43b2-8656-639978632741\") " pod="openshift-must-gather-n5vql/must-gather-jkhmz" Jan 21 12:10:35 crc kubenswrapper[4824]: I0121 12:10:35.273384 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/e1c80411-6f38-43b2-8656-639978632741-must-gather-output\") pod \"must-gather-jkhmz\" (UID: \"e1c80411-6f38-43b2-8656-639978632741\") " pod="openshift-must-gather-n5vql/must-gather-jkhmz" Jan 21 12:10:35 crc kubenswrapper[4824]: I0121 12:10:35.290707 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fdgjw\" (UniqueName: \"kubernetes.io/projected/e1c80411-6f38-43b2-8656-639978632741-kube-api-access-fdgjw\") pod \"must-gather-jkhmz\" (UID: \"e1c80411-6f38-43b2-8656-639978632741\") " pod="openshift-must-gather-n5vql/must-gather-jkhmz" Jan 21 12:10:35 crc kubenswrapper[4824]: I0121 12:10:35.426776 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-n5vql/must-gather-jkhmz" Jan 21 12:10:35 crc kubenswrapper[4824]: I0121 12:10:35.837233 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-n5vql/must-gather-jkhmz"] Jan 21 12:10:36 crc kubenswrapper[4824]: I0121 12:10:36.040611 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-n5vql/must-gather-jkhmz" event={"ID":"e1c80411-6f38-43b2-8656-639978632741","Type":"ContainerStarted","Data":"65ed37028631bc528541d931c3708c861369ab00cb45d0d01989f2e346501aa0"} Jan 21 12:10:37 crc kubenswrapper[4824]: I0121 12:10:37.052566 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-n5vql/must-gather-jkhmz" event={"ID":"e1c80411-6f38-43b2-8656-639978632741","Type":"ContainerStarted","Data":"d0b75fbd713ed74fa81844f4c6ccb4b25946f8456765e9861e694df4f9f52b15"} Jan 21 12:10:37 crc kubenswrapper[4824]: I0121 12:10:37.052991 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-n5vql/must-gather-jkhmz" event={"ID":"e1c80411-6f38-43b2-8656-639978632741","Type":"ContainerStarted","Data":"0937d1d47c67d983a6ee92637b82dba42aee39048f14d63afc09202819ad4750"} Jan 21 12:10:37 crc kubenswrapper[4824]: I0121 12:10:37.072979 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-n5vql/must-gather-jkhmz" podStartSLOduration=2.072946813 podStartE2EDuration="2.072946813s" podCreationTimestamp="2026-01-21 12:10:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 12:10:37.070699456 +0000 UTC m=+3579.363728748" watchObservedRunningTime="2026-01-21 12:10:37.072946813 +0000 UTC m=+3579.365976095" Jan 21 12:10:39 crc kubenswrapper[4824]: I0121 12:10:39.572088 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-n5vql/crc-debug-9n2zc"] Jan 21 12:10:39 crc kubenswrapper[4824]: I0121 12:10:39.573826 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-n5vql/crc-debug-9n2zc" Jan 21 12:10:39 crc kubenswrapper[4824]: I0121 12:10:39.765131 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9rbp4\" (UniqueName: \"kubernetes.io/projected/d2b727cf-5339-4da5-be88-b46b2bcf5dd1-kube-api-access-9rbp4\") pod \"crc-debug-9n2zc\" (UID: \"d2b727cf-5339-4da5-be88-b46b2bcf5dd1\") " pod="openshift-must-gather-n5vql/crc-debug-9n2zc" Jan 21 12:10:39 crc kubenswrapper[4824]: I0121 12:10:39.765335 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d2b727cf-5339-4da5-be88-b46b2bcf5dd1-host\") pod \"crc-debug-9n2zc\" (UID: \"d2b727cf-5339-4da5-be88-b46b2bcf5dd1\") " pod="openshift-must-gather-n5vql/crc-debug-9n2zc" Jan 21 12:10:39 crc kubenswrapper[4824]: I0121 12:10:39.868151 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9rbp4\" (UniqueName: \"kubernetes.io/projected/d2b727cf-5339-4da5-be88-b46b2bcf5dd1-kube-api-access-9rbp4\") pod \"crc-debug-9n2zc\" (UID: \"d2b727cf-5339-4da5-be88-b46b2bcf5dd1\") " pod="openshift-must-gather-n5vql/crc-debug-9n2zc" Jan 21 12:10:39 crc kubenswrapper[4824]: I0121 12:10:39.868469 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d2b727cf-5339-4da5-be88-b46b2bcf5dd1-host\") pod \"crc-debug-9n2zc\" (UID: \"d2b727cf-5339-4da5-be88-b46b2bcf5dd1\") " pod="openshift-must-gather-n5vql/crc-debug-9n2zc" Jan 21 12:10:39 crc kubenswrapper[4824]: I0121 12:10:39.868604 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d2b727cf-5339-4da5-be88-b46b2bcf5dd1-host\") pod \"crc-debug-9n2zc\" (UID: \"d2b727cf-5339-4da5-be88-b46b2bcf5dd1\") " pod="openshift-must-gather-n5vql/crc-debug-9n2zc" Jan 21 12:10:39 crc kubenswrapper[4824]: I0121 12:10:39.888455 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9rbp4\" (UniqueName: \"kubernetes.io/projected/d2b727cf-5339-4da5-be88-b46b2bcf5dd1-kube-api-access-9rbp4\") pod \"crc-debug-9n2zc\" (UID: \"d2b727cf-5339-4da5-be88-b46b2bcf5dd1\") " pod="openshift-must-gather-n5vql/crc-debug-9n2zc" Jan 21 12:10:39 crc kubenswrapper[4824]: I0121 12:10:39.893829 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-n5vql/crc-debug-9n2zc" Jan 21 12:10:39 crc kubenswrapper[4824]: W0121 12:10:39.924735 4824 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd2b727cf_5339_4da5_be88_b46b2bcf5dd1.slice/crio-2730b37ace706deb236dc24c607e69f539af5cbe65b803a5f377d1ca122f5af3 WatchSource:0}: Error finding container 2730b37ace706deb236dc24c607e69f539af5cbe65b803a5f377d1ca122f5af3: Status 404 returned error can't find the container with id 2730b37ace706deb236dc24c607e69f539af5cbe65b803a5f377d1ca122f5af3 Jan 21 12:10:40 crc kubenswrapper[4824]: I0121 12:10:40.099909 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-n5vql/crc-debug-9n2zc" event={"ID":"d2b727cf-5339-4da5-be88-b46b2bcf5dd1","Type":"ContainerStarted","Data":"2730b37ace706deb236dc24c607e69f539af5cbe65b803a5f377d1ca122f5af3"} Jan 21 12:10:41 crc kubenswrapper[4824]: I0121 12:10:41.109269 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-n5vql/crc-debug-9n2zc" event={"ID":"d2b727cf-5339-4da5-be88-b46b2bcf5dd1","Type":"ContainerStarted","Data":"2719d0b32f33e24c79ae317ca212bc0f5429a69ed705aafd8970f2d93f3e60b9"} Jan 21 12:10:41 crc kubenswrapper[4824]: I0121 12:10:41.128647 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-n5vql/crc-debug-9n2zc" podStartSLOduration=2.128633542 podStartE2EDuration="2.128633542s" podCreationTimestamp="2026-01-21 12:10:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-21 12:10:41.120545735 +0000 UTC m=+3583.413575026" watchObservedRunningTime="2026-01-21 12:10:41.128633542 +0000 UTC m=+3583.421662834" Jan 21 12:10:41 crc kubenswrapper[4824]: I0121 12:10:41.892470 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-6c7f965fb4-8slr2_3f7f97b5-8abd-4351-8396-cf1b76c376a7/barbican-api-log/0.log" Jan 21 12:10:41 crc kubenswrapper[4824]: I0121 12:10:41.903888 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-6c7f965fb4-8slr2_3f7f97b5-8abd-4351-8396-cf1b76c376a7/barbican-api/0.log" Jan 21 12:10:41 crc kubenswrapper[4824]: I0121 12:10:41.960221 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-79b7cfb8cd-ktpp5_8b84edbe-4485-4027-bfc4-cebbb1b10543/barbican-keystone-listener-log/0.log" Jan 21 12:10:41 crc kubenswrapper[4824]: I0121 12:10:41.966424 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-79b7cfb8cd-ktpp5_8b84edbe-4485-4027-bfc4-cebbb1b10543/barbican-keystone-listener/0.log" Jan 21 12:10:41 crc kubenswrapper[4824]: I0121 12:10:41.982738 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-7cf777545f-f8drb_283e6982-7444-48c7-b7ac-ed4210329a08/barbican-worker-log/0.log" Jan 21 12:10:41 crc kubenswrapper[4824]: I0121 12:10:41.989709 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-7cf777545f-f8drb_283e6982-7444-48c7-b7ac-ed4210329a08/barbican-worker/0.log" Jan 21 12:10:42 crc kubenswrapper[4824]: I0121 12:10:42.025571 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-9z7z6_300a037e-ac25-4a9e-9339-6cc707969f7c/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Jan 21 12:10:42 crc kubenswrapper[4824]: I0121 12:10:42.052187 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_adc98d07-2b32-4f30-bdb7-40923468389e/ceilometer-central-agent/0.log" Jan 21 12:10:42 crc kubenswrapper[4824]: I0121 12:10:42.074700 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_adc98d07-2b32-4f30-bdb7-40923468389e/ceilometer-notification-agent/0.log" Jan 21 12:10:42 crc kubenswrapper[4824]: I0121 12:10:42.078458 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_adc98d07-2b32-4f30-bdb7-40923468389e/sg-core/0.log" Jan 21 12:10:42 crc kubenswrapper[4824]: I0121 12:10:42.088883 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_adc98d07-2b32-4f30-bdb7-40923468389e/proxy-httpd/0.log" Jan 21 12:10:42 crc kubenswrapper[4824]: I0121 12:10:42.105023 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_ff9de537-ecd2-469d-b90e-5857bbf64c2d/cinder-api-log/0.log" Jan 21 12:10:42 crc kubenswrapper[4824]: I0121 12:10:42.155316 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_ff9de537-ecd2-469d-b90e-5857bbf64c2d/cinder-api/0.log" Jan 21 12:10:42 crc kubenswrapper[4824]: I0121 12:10:42.202215 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_2e0f0267-58d2-4690-9187-8ddf03ef082e/cinder-scheduler/0.log" Jan 21 12:10:42 crc kubenswrapper[4824]: I0121 12:10:42.231228 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_2e0f0267-58d2-4690-9187-8ddf03ef082e/probe/0.log" Jan 21 12:10:42 crc kubenswrapper[4824]: I0121 12:10:42.251133 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-wqzrz_b8c7ec7b-ee3f-46bc-95f8-4406713ba43a/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Jan 21 12:10:42 crc kubenswrapper[4824]: I0121 12:10:42.269972 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-4bd5n_8d46d1eb-4230-49c7-8376-698e7689ad2b/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Jan 21 12:10:42 crc kubenswrapper[4824]: I0121 12:10:42.315367 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-6f6df4f56c-26d8b_976a6446-e79c-42db-be45-43147767f1da/dnsmasq-dns/0.log" Jan 21 12:10:42 crc kubenswrapper[4824]: I0121 12:10:42.318472 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-6f6df4f56c-26d8b_976a6446-e79c-42db-be45-43147767f1da/init/0.log" Jan 21 12:10:42 crc kubenswrapper[4824]: I0121 12:10:42.338536 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-dpgt9_8dbaf72e-72a7-4073-a4c1-95cbdabaa64b/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Jan 21 12:10:42 crc kubenswrapper[4824]: I0121 12:10:42.348643 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_219725d0-9bef-4e42-8b78-cbc74a8ac9db/glance-log/0.log" Jan 21 12:10:42 crc kubenswrapper[4824]: I0121 12:10:42.369743 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_219725d0-9bef-4e42-8b78-cbc74a8ac9db/glance-httpd/0.log" Jan 21 12:10:42 crc kubenswrapper[4824]: I0121 12:10:42.385868 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_7419ae4a-e8ea-4384-a9b8-3a3e02e27f25/glance-log/0.log" Jan 21 12:10:42 crc kubenswrapper[4824]: I0121 12:10:42.407747 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_7419ae4a-e8ea-4384-a9b8-3a3e02e27f25/glance-httpd/0.log" Jan 21 12:10:42 crc kubenswrapper[4824]: I0121 12:10:42.844183 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-api-549445f5fc-fwxlz_8655cad5-2680-4cf6-b50a-0745adeeb469/heat-api/0.log" Jan 21 12:10:43 crc kubenswrapper[4824]: I0121 12:10:43.127934 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-cfnapi-68b8cf894c-ws5zd_a1aefb21-aac8-4f8f-a402-16006c1f336a/heat-cfnapi/0.log" Jan 21 12:10:43 crc kubenswrapper[4824]: I0121 12:10:43.232428 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-engine-5bc6fc99dd-nv45v_ca8530d7-bcfb-45fa-b0ff-a2702c081bda/heat-engine/0.log" Jan 21 12:10:43 crc kubenswrapper[4824]: I0121 12:10:43.248420 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-bfzb2_b821ea8d-9327-4623-b5be-3da7d3872cab/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Jan 21 12:10:43 crc kubenswrapper[4824]: I0121 12:10:43.273435 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-b74gh_60fedbf7-07e6-41be-ada3-2a06f28835e3/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Jan 21 12:10:43 crc kubenswrapper[4824]: I0121 12:10:43.638773 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-6c5fc769c9-spmzr_7d943b5e-d05f-445a-888d-7ebe340468b9/keystone-api/0.log" Jan 21 12:10:43 crc kubenswrapper[4824]: I0121 12:10:43.651348 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29483281-bz8hk_c92d6f21-1fe2-41a2-a516-a0d0740ab36f/keystone-cron/0.log" Jan 21 12:10:43 crc kubenswrapper[4824]: I0121 12:10:43.662187 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_d42233a9-cfd9-486d-b314-f17d6a9d3b03/kube-state-metrics/0.log" Jan 21 12:10:43 crc kubenswrapper[4824]: I0121 12:10:43.685084 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-pgjpn_fdde8991-c204-45a0-b344-dcdc41a9a275/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Jan 21 12:10:46 crc kubenswrapper[4824]: I0121 12:10:46.065105 4824 patch_prober.go:28] interesting pod/machine-config-daemon-zqjg9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 21 12:10:46 crc kubenswrapper[4824]: I0121 12:10:46.065623 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 21 12:10:46 crc kubenswrapper[4824]: I0121 12:10:46.070590 4824 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" Jan 21 12:10:46 crc kubenswrapper[4824]: I0121 12:10:46.071026 4824 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"719b0a497b4be25b481cf2d3456b6c515017599e2aeafd1f30d22bffc35f79d4"} pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 21 12:10:46 crc kubenswrapper[4824]: I0121 12:10:46.071091 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" containerName="machine-config-daemon" containerID="cri-o://719b0a497b4be25b481cf2d3456b6c515017599e2aeafd1f30d22bffc35f79d4" gracePeriod=600 Jan 21 12:10:46 crc kubenswrapper[4824]: E0121 12:10:46.191379 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 12:10:47 crc kubenswrapper[4824]: I0121 12:10:47.160097 4824 generic.go:334] "Generic (PLEG): container finished" podID="33f3d922-4ffe-409b-a49a-d88c85898260" containerID="719b0a497b4be25b481cf2d3456b6c515017599e2aeafd1f30d22bffc35f79d4" exitCode=0 Jan 21 12:10:47 crc kubenswrapper[4824]: I0121 12:10:47.160166 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" event={"ID":"33f3d922-4ffe-409b-a49a-d88c85898260","Type":"ContainerDied","Data":"719b0a497b4be25b481cf2d3456b6c515017599e2aeafd1f30d22bffc35f79d4"} Jan 21 12:10:47 crc kubenswrapper[4824]: I0121 12:10:47.160379 4824 scope.go:117] "RemoveContainer" containerID="253e918881780a912353868ed3027165c77d110c4811895ff906677458c0004d" Jan 21 12:10:47 crc kubenswrapper[4824]: I0121 12:10:47.161311 4824 scope.go:117] "RemoveContainer" containerID="719b0a497b4be25b481cf2d3456b6c515017599e2aeafd1f30d22bffc35f79d4" Jan 21 12:10:47 crc kubenswrapper[4824]: E0121 12:10:47.161675 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 12:11:00 crc kubenswrapper[4824]: I0121 12:11:00.050267 4824 scope.go:117] "RemoveContainer" containerID="719b0a497b4be25b481cf2d3456b6c515017599e2aeafd1f30d22bffc35f79d4" Jan 21 12:11:00 crc kubenswrapper[4824]: E0121 12:11:00.050750 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 12:11:00 crc kubenswrapper[4824]: I0121 12:11:00.070115 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_dba0050b-8a73-4355-a1e0-7c9a03557ead/memcached/0.log" Jan 21 12:11:00 crc kubenswrapper[4824]: I0121 12:11:00.171266 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-659565f76f-jdg2q_b6f2ba03-746e-465c-bde2-c917129fb125/neutron-api/0.log" Jan 21 12:11:00 crc kubenswrapper[4824]: I0121 12:11:00.229881 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-659565f76f-jdg2q_b6f2ba03-746e-465c-bde2-c917129fb125/neutron-httpd/0.log" Jan 21 12:11:00 crc kubenswrapper[4824]: I0121 12:11:00.251307 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-rf47l_4f3c3a58-5d6b-4cc2-9707-d088b2fd463d/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Jan 21 12:11:00 crc kubenswrapper[4824]: I0121 12:11:00.523095 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_55525bd3-ae2b-494d-9a33-dd7d00c576b2/nova-api-log/0.log" Jan 21 12:11:00 crc kubenswrapper[4824]: I0121 12:11:00.980640 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_55525bd3-ae2b-494d-9a33-dd7d00c576b2/nova-api-api/0.log" Jan 21 12:11:01 crc kubenswrapper[4824]: I0121 12:11:01.122349 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_8effdacf-8fec-40b4-8f61-8856ac0f232c/nova-cell0-conductor-conductor/0.log" Jan 21 12:11:01 crc kubenswrapper[4824]: I0121 12:11:01.243525 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_c398d1e9-13e1-400c-a609-bceea49aea51/nova-cell1-conductor-conductor/0.log" Jan 21 12:11:01 crc kubenswrapper[4824]: I0121 12:11:01.348506 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_8e42e9d6-485f-4b97-b049-316cd47af30e/nova-cell1-novncproxy-novncproxy/0.log" Jan 21 12:11:01 crc kubenswrapper[4824]: I0121 12:11:01.402490 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-7crw8_7355a8fc-8c18-4de1-8baf-f6294ea9e11c/nova-edpm-deployment-openstack-edpm-ipam/0.log" Jan 21 12:11:01 crc kubenswrapper[4824]: I0121 12:11:01.497599 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_654a41d0-b44a-4194-a514-a3c60126d0d9/nova-metadata-log/0.log" Jan 21 12:11:02 crc kubenswrapper[4824]: I0121 12:11:02.522445 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_654a41d0-b44a-4194-a514-a3c60126d0d9/nova-metadata-metadata/0.log" Jan 21 12:11:02 crc kubenswrapper[4824]: I0121 12:11:02.678159 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_7860ffbb-b435-4736-b2d3-f8871e70dc7a/nova-scheduler-scheduler/0.log" Jan 21 12:11:02 crc kubenswrapper[4824]: I0121 12:11:02.697342 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_7cd8d62b-ab74-4a1b-95f5-7253ed6c0346/galera/0.log" Jan 21 12:11:02 crc kubenswrapper[4824]: I0121 12:11:02.720168 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_7cd8d62b-ab74-4a1b-95f5-7253ed6c0346/mysql-bootstrap/0.log" Jan 21 12:11:02 crc kubenswrapper[4824]: I0121 12:11:02.901714 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_218aca21-4c37-4862-b891-fc228baa5348/galera/0.log" Jan 21 12:11:02 crc kubenswrapper[4824]: I0121 12:11:02.911937 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_218aca21-4c37-4862-b891-fc228baa5348/mysql-bootstrap/0.log" Jan 21 12:11:02 crc kubenswrapper[4824]: I0121 12:11:02.931199 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_1af94d08-6d1e-4a97-8f82-27a1755d17a1/openstackclient/0.log" Jan 21 12:11:02 crc kubenswrapper[4824]: I0121 12:11:02.946428 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-htwd6_511f2b6d-a08d-49f8-b393-ab222219d4a7/ovn-controller/0.log" Jan 21 12:11:02 crc kubenswrapper[4824]: I0121 12:11:02.953140 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-vrpmw_8652f3f6-1ad7-49d9-9048-827be17a3dd0/openstack-network-exporter/0.log" Jan 21 12:11:02 crc kubenswrapper[4824]: I0121 12:11:02.963285 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-gr9gj_b27b3f77-8476-4f3e-a946-f4f26fc66cb5/ovsdb-server/0.log" Jan 21 12:11:02 crc kubenswrapper[4824]: I0121 12:11:02.972175 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-gr9gj_b27b3f77-8476-4f3e-a946-f4f26fc66cb5/ovs-vswitchd/0.log" Jan 21 12:11:02 crc kubenswrapper[4824]: I0121 12:11:02.976717 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-gr9gj_b27b3f77-8476-4f3e-a946-f4f26fc66cb5/ovsdb-server-init/0.log" Jan 21 12:11:03 crc kubenswrapper[4824]: I0121 12:11:03.005683 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-gf4zq_18e65eb9-3b96-48cd-88b2-9f7fb0e54b6f/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Jan 21 12:11:03 crc kubenswrapper[4824]: I0121 12:11:03.016373 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_8b9c25b1-d0b9-430e-88cf-0aeaf6785461/ovn-northd/0.log" Jan 21 12:11:03 crc kubenswrapper[4824]: I0121 12:11:03.022359 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_8b9c25b1-d0b9-430e-88cf-0aeaf6785461/openstack-network-exporter/0.log" Jan 21 12:11:03 crc kubenswrapper[4824]: I0121 12:11:03.040695 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_5c0b5876-da04-49e6-8946-4fc6060a64ec/ovsdbserver-nb/0.log" Jan 21 12:11:03 crc kubenswrapper[4824]: I0121 12:11:03.046109 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_5c0b5876-da04-49e6-8946-4fc6060a64ec/openstack-network-exporter/0.log" Jan 21 12:11:03 crc kubenswrapper[4824]: I0121 12:11:03.063301 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_e86f23d4-099d-4e8b-b3a4-d2fc7152c1ef/ovsdbserver-sb/0.log" Jan 21 12:11:03 crc kubenswrapper[4824]: I0121 12:11:03.067562 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_e86f23d4-099d-4e8b-b3a4-d2fc7152c1ef/openstack-network-exporter/0.log" Jan 21 12:11:03 crc kubenswrapper[4824]: I0121 12:11:03.154323 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-8654bbf998-p8crm_c23477ba-5a41-44da-a8b7-6606595b9726/placement-log/0.log" Jan 21 12:11:03 crc kubenswrapper[4824]: I0121 12:11:03.223685 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-8654bbf998-p8crm_c23477ba-5a41-44da-a8b7-6606595b9726/placement-api/0.log" Jan 21 12:11:03 crc kubenswrapper[4824]: I0121 12:11:03.244086 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_5d8e256a-aa1b-4ccb-8e8c-c86be1e31762/rabbitmq/0.log" Jan 21 12:11:03 crc kubenswrapper[4824]: I0121 12:11:03.247409 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_5d8e256a-aa1b-4ccb-8e8c-c86be1e31762/setup-container/0.log" Jan 21 12:11:03 crc kubenswrapper[4824]: I0121 12:11:03.270850 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_e4a4d725-1a49-4292-af74-6ab09ab15cdb/rabbitmq/0.log" Jan 21 12:11:03 crc kubenswrapper[4824]: I0121 12:11:03.274912 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_e4a4d725-1a49-4292-af74-6ab09ab15cdb/setup-container/0.log" Jan 21 12:11:03 crc kubenswrapper[4824]: I0121 12:11:03.289701 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-8w57w_890acdab-9f31-4afc-80e4-c1df308ac5a6/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Jan 21 12:11:03 crc kubenswrapper[4824]: I0121 12:11:03.297327 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-2flzn_34c68ecf-cf29-44cb-9880-a3d45f625454/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Jan 21 12:11:03 crc kubenswrapper[4824]: I0121 12:11:03.307345 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-2mbcf_e3664d03-d53d-4a49-8cab-c4266f2d1426/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Jan 21 12:11:03 crc kubenswrapper[4824]: I0121 12:11:03.317256 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-pn4sw_ad533443-01b7-4860-9920-93ed67f6b52f/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Jan 21 12:11:03 crc kubenswrapper[4824]: I0121 12:11:03.332490 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-wbrwb_4529344f-3dfd-4858-83f1-4abffe47f2ad/ssh-known-hosts-edpm-deployment/0.log" Jan 21 12:11:03 crc kubenswrapper[4824]: I0121 12:11:03.474432 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-79f4c948b7-4hz4p_01d4ce4d-3728-4f4c-9fbf-ce58793b6b7d/proxy-httpd/0.log" Jan 21 12:11:03 crc kubenswrapper[4824]: I0121 12:11:03.504393 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-79f4c948b7-4hz4p_01d4ce4d-3728-4f4c-9fbf-ce58793b6b7d/proxy-server/0.log" Jan 21 12:11:03 crc kubenswrapper[4824]: I0121 12:11:03.515608 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-nc6vf_29473359-3517-4b86-bd5e-80e25706ff27/swift-ring-rebalance/0.log" Jan 21 12:11:03 crc kubenswrapper[4824]: I0121 12:11:03.559703 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_f980e493-ca74-47f1-8f6f-aca9c9011bf5/account-server/0.log" Jan 21 12:11:03 crc kubenswrapper[4824]: I0121 12:11:03.574383 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-6968d8fdc4-bbv7p_d30695b8-6f6a-48f0-88ed-9181484634b8/controller/0.log" Jan 21 12:11:03 crc kubenswrapper[4824]: I0121 12:11:03.580147 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-6968d8fdc4-bbv7p_d30695b8-6f6a-48f0-88ed-9181484634b8/kube-rbac-proxy/0.log" Jan 21 12:11:03 crc kubenswrapper[4824]: I0121 12:11:03.588466 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_f980e493-ca74-47f1-8f6f-aca9c9011bf5/account-replicator/0.log" Jan 21 12:11:03 crc kubenswrapper[4824]: I0121 12:11:03.593038 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_f980e493-ca74-47f1-8f6f-aca9c9011bf5/account-auditor/0.log" Jan 21 12:11:03 crc kubenswrapper[4824]: I0121 12:11:03.602458 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_f980e493-ca74-47f1-8f6f-aca9c9011bf5/account-reaper/0.log" Jan 21 12:11:03 crc kubenswrapper[4824]: I0121 12:11:03.603945 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9q9wj_e38576f0-c423-497a-a30d-60c2e8edd93a/controller/0.log" Jan 21 12:11:03 crc kubenswrapper[4824]: I0121 12:11:03.613727 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_f980e493-ca74-47f1-8f6f-aca9c9011bf5/container-server/0.log" Jan 21 12:11:03 crc kubenswrapper[4824]: I0121 12:11:03.647457 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_f980e493-ca74-47f1-8f6f-aca9c9011bf5/container-replicator/0.log" Jan 21 12:11:03 crc kubenswrapper[4824]: I0121 12:11:03.652840 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_f980e493-ca74-47f1-8f6f-aca9c9011bf5/container-auditor/0.log" Jan 21 12:11:03 crc kubenswrapper[4824]: I0121 12:11:03.664293 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_f980e493-ca74-47f1-8f6f-aca9c9011bf5/container-updater/0.log" Jan 21 12:11:03 crc kubenswrapper[4824]: I0121 12:11:03.673724 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_f980e493-ca74-47f1-8f6f-aca9c9011bf5/object-server/0.log" Jan 21 12:11:03 crc kubenswrapper[4824]: I0121 12:11:03.703695 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_f980e493-ca74-47f1-8f6f-aca9c9011bf5/object-replicator/0.log" Jan 21 12:11:03 crc kubenswrapper[4824]: I0121 12:11:03.731306 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_f980e493-ca74-47f1-8f6f-aca9c9011bf5/object-auditor/0.log" Jan 21 12:11:03 crc kubenswrapper[4824]: I0121 12:11:03.737626 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_f980e493-ca74-47f1-8f6f-aca9c9011bf5/object-updater/0.log" Jan 21 12:11:03 crc kubenswrapper[4824]: I0121 12:11:03.750929 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_f980e493-ca74-47f1-8f6f-aca9c9011bf5/object-expirer/0.log" Jan 21 12:11:03 crc kubenswrapper[4824]: I0121 12:11:03.766525 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_f980e493-ca74-47f1-8f6f-aca9c9011bf5/rsync/0.log" Jan 21 12:11:03 crc kubenswrapper[4824]: I0121 12:11:03.781445 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_f980e493-ca74-47f1-8f6f-aca9c9011bf5/swift-recon-cron/0.log" Jan 21 12:11:03 crc kubenswrapper[4824]: I0121 12:11:03.887614 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-g8gx6_49b2a136-e5c3-4741-80c6-4c545d4a82a0/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Jan 21 12:11:03 crc kubenswrapper[4824]: I0121 12:11:03.942686 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_58d30951-ee56-4217-a614-a04951369c8b/tempest-tests-tempest-tests-runner/0.log" Jan 21 12:11:03 crc kubenswrapper[4824]: I0121 12:11:03.949836 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_333160e2-3c05-401f-af89-abae8c0cb9d9/test-operator-logs-container/0.log" Jan 21 12:11:03 crc kubenswrapper[4824]: I0121 12:11:03.985546 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-5b6wz_d672f153-6ef2-432d-829a-db63c8daab69/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Jan 21 12:11:05 crc kubenswrapper[4824]: I0121 12:11:05.009262 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7ddb5c749-bvptq_8ed6bb2c-2770-4785-bf38-9b6fd0bd8472/manager/0.log" Jan 21 12:11:05 crc kubenswrapper[4824]: I0121 12:11:05.064031 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-9b68f5989-24r54_a25733d4-a5f8-4c97-a897-8d96e637c253/manager/0.log" Jan 21 12:11:05 crc kubenswrapper[4824]: I0121 12:11:05.075317 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_db4b9aacd6f51b5aff2d66474158631eb1e58fb4d1358e39c8aaaca1bdn7cxr_72ab26cb-149e-4825-9587-3acfa66a368f/extract/0.log" Jan 21 12:11:05 crc kubenswrapper[4824]: I0121 12:11:05.090178 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_db4b9aacd6f51b5aff2d66474158631eb1e58fb4d1358e39c8aaaca1bdn7cxr_72ab26cb-149e-4825-9587-3acfa66a368f/util/0.log" Jan 21 12:11:05 crc kubenswrapper[4824]: I0121 12:11:05.095270 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_db4b9aacd6f51b5aff2d66474158631eb1e58fb4d1358e39c8aaaca1bdn7cxr_72ab26cb-149e-4825-9587-3acfa66a368f/pull/0.log" Jan 21 12:11:05 crc kubenswrapper[4824]: I0121 12:11:05.108383 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-9f958b845-wtfj4_14dd1f12-3943-42e4-be1b-c9e37e49d21b/manager/0.log" Jan 21 12:11:05 crc kubenswrapper[4824]: I0121 12:11:05.229048 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-c6994669c-hvq6s_b7053a68-ce89-48b6-9b5d-1c6dac813a75/manager/0.log" Jan 21 12:11:05 crc kubenswrapper[4824]: I0121 12:11:05.312063 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9q9wj_e38576f0-c423-497a-a30d-60c2e8edd93a/frr/0.log" Jan 21 12:11:05 crc kubenswrapper[4824]: I0121 12:11:05.327462 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9q9wj_e38576f0-c423-497a-a30d-60c2e8edd93a/reloader/0.log" Jan 21 12:11:05 crc kubenswrapper[4824]: I0121 12:11:05.328276 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5fbc59b5b6-hgmf4_944c3123-9063-4731-9581-5b683741edc4/manager/0.log" Jan 21 12:11:05 crc kubenswrapper[4824]: I0121 12:11:05.332999 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9q9wj_e38576f0-c423-497a-a30d-60c2e8edd93a/frr-metrics/0.log" Jan 21 12:11:05 crc kubenswrapper[4824]: I0121 12:11:05.339746 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9q9wj_e38576f0-c423-497a-a30d-60c2e8edd93a/kube-rbac-proxy/0.log" Jan 21 12:11:05 crc kubenswrapper[4824]: I0121 12:11:05.345668 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9q9wj_e38576f0-c423-497a-a30d-60c2e8edd93a/kube-rbac-proxy-frr/0.log" Jan 21 12:11:05 crc kubenswrapper[4824]: I0121 12:11:05.348525 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-77d5c5b54f-mp5ml_2463afb1-0271-4ab3-ad1a-9329ecaeec4d/manager/0.log" Jan 21 12:11:05 crc kubenswrapper[4824]: I0121 12:11:05.352989 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9q9wj_e38576f0-c423-497a-a30d-60c2e8edd93a/cp-frr-files/0.log" Jan 21 12:11:05 crc kubenswrapper[4824]: I0121 12:11:05.359627 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9q9wj_e38576f0-c423-497a-a30d-60c2e8edd93a/cp-reloader/0.log" Jan 21 12:11:05 crc kubenswrapper[4824]: I0121 12:11:05.370672 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9q9wj_e38576f0-c423-497a-a30d-60c2e8edd93a/cp-metrics/0.log" Jan 21 12:11:05 crc kubenswrapper[4824]: I0121 12:11:05.377752 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7df86c4f6c-fdblb_d71f993c-b2c7-47af-884f-4cbe6dcf3c1a/frr-k8s-webhook-server/0.log" Jan 21 12:11:05 crc kubenswrapper[4824]: I0121 12:11:05.410781 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-5c58bcfcff-69lct_714d2131-afc8-4828-840f-e2f0ce727d5d/manager/0.log" Jan 21 12:11:05 crc kubenswrapper[4824]: I0121 12:11:05.424049 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-68bc9bfcb6-hzmp5_81fed5ab-ec3d-46f0-8998-199393bffba0/webhook-server/0.log" Jan 21 12:11:05 crc kubenswrapper[4824]: I0121 12:11:05.731450 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-77c48c7859-dzn7w_57e1fe13-b2c8-4711-9efe-1ebbc7e9d12e/manager/0.log" Jan 21 12:11:05 crc kubenswrapper[4824]: I0121 12:11:05.742041 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-78757b4889-gkwjs_fc911701-7c15-4a20-b89e-709521bdd3a1/manager/0.log" Jan 21 12:11:05 crc kubenswrapper[4824]: I0121 12:11:05.844471 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-767fdc4f47-5mhmd_2e5efa9b-40bc-4860-967d-aa37e1838f81/manager/0.log" Jan 21 12:11:05 crc kubenswrapper[4824]: I0121 12:11:05.853095 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-864f6b75bf-c7vsf_18017a19-1676-4fe9-a97e-ae878f423847/manager/0.log" Jan 21 12:11:05 crc kubenswrapper[4824]: I0121 12:11:05.855698 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-8tsb9_baf5db51-d3ac-4b97-9c5f-afc598d59ab7/speaker/0.log" Jan 21 12:11:05 crc kubenswrapper[4824]: I0121 12:11:05.861531 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-8tsb9_baf5db51-d3ac-4b97-9c5f-afc598d59ab7/kube-rbac-proxy/0.log" Jan 21 12:11:05 crc kubenswrapper[4824]: I0121 12:11:05.915924 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-c87fff755-d5hn9_19b1c150-f5ed-4432-bfc2-717b78eebc7e/manager/0.log" Jan 21 12:11:05 crc kubenswrapper[4824]: I0121 12:11:05.949132 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-cb4666565-7w8p4_8c74d3da-d870-45e9-96b9-70eb2395c977/manager/0.log" Jan 21 12:11:06 crc kubenswrapper[4824]: I0121 12:11:06.016192 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-65849867d6-8ngx6_b38ac60b-1643-44d9-a1fe-5f75129247dc/manager/0.log" Jan 21 12:11:06 crc kubenswrapper[4824]: I0121 12:11:06.024441 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-7fc9b76cf6-tncbx_5429da27-a618-44da-8aeb-12b917af8734/manager/0.log" Jan 21 12:11:06 crc kubenswrapper[4824]: I0121 12:11:06.037536 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-6b68b8b854zxcbl_3fa95966-3147-4e50-8ed8-f702a87992c4/manager/0.log" Jan 21 12:11:06 crc kubenswrapper[4824]: I0121 12:11:06.172216 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-init-dbdbd6855-gkm5b_35334861-8e16-4b16-8cf5-167a049836af/operator/0.log" Jan 21 12:11:07 crc kubenswrapper[4824]: I0121 12:11:07.109619 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-74d84f758c-vjgps_c1128e01-8b7c-4742-b1da-dbf06aa8b4d8/manager/0.log" Jan 21 12:11:07 crc kubenswrapper[4824]: I0121 12:11:07.190804 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-zggxm_78eb1906-d9e6-49f7-a512-ed3cc0a07da7/registry-server/0.log" Jan 21 12:11:07 crc kubenswrapper[4824]: I0121 12:11:07.238773 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-55db956ddc-78z2q_f5639f17-3a5b-4ef4-bd6e-6eeb3274636a/manager/0.log" Jan 21 12:11:07 crc kubenswrapper[4824]: I0121 12:11:07.256473 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-686df47fcb-c8xvs_d8b36106-8f21-4873-890d-c1879ea49068/manager/0.log" Jan 21 12:11:07 crc kubenswrapper[4824]: I0121 12:11:07.277528 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-94lqx_782093ab-751c-4e38-bba3-c5f76308c82e/operator/0.log" Jan 21 12:11:07 crc kubenswrapper[4824]: I0121 12:11:07.296351 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-85dd56d4cc-wlrdt_b4cb80e6-c720-4875-9848-7d9597a354e4/manager/0.log" Jan 21 12:11:07 crc kubenswrapper[4824]: I0121 12:11:07.354529 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-5f8f495fcf-6rz5h_db84b172-59d8-4995-815b-50b492283bee/manager/0.log" Jan 21 12:11:07 crc kubenswrapper[4824]: I0121 12:11:07.361411 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-7cd8bc9dbb-xkp8x_52491e7b-cd8c-403a-942c-312debd6f1c6/manager/0.log" Jan 21 12:11:07 crc kubenswrapper[4824]: I0121 12:11:07.371077 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-64cd966744-9w4ch_0715aa91-2ae5-4e07-8388-c37def4c7743/manager/0.log" Jan 21 12:11:11 crc kubenswrapper[4824]: I0121 12:11:11.368776 4824 generic.go:334] "Generic (PLEG): container finished" podID="d2b727cf-5339-4da5-be88-b46b2bcf5dd1" containerID="2719d0b32f33e24c79ae317ca212bc0f5429a69ed705aafd8970f2d93f3e60b9" exitCode=0 Jan 21 12:11:11 crc kubenswrapper[4824]: I0121 12:11:11.368867 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-n5vql/crc-debug-9n2zc" event={"ID":"d2b727cf-5339-4da5-be88-b46b2bcf5dd1","Type":"ContainerDied","Data":"2719d0b32f33e24c79ae317ca212bc0f5429a69ed705aafd8970f2d93f3e60b9"} Jan 21 12:11:11 crc kubenswrapper[4824]: I0121 12:11:11.981353 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-r65dm_c117483a-fe5c-423b-995e-9c2e7f59566a/control-plane-machine-set-operator/0.log" Jan 21 12:11:11 crc kubenswrapper[4824]: I0121 12:11:11.992643 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-7grz2_fb8bdb2b-bdc7-4433-be10-296fc67dc1c3/kube-rbac-proxy/0.log" Jan 21 12:11:12 crc kubenswrapper[4824]: I0121 12:11:12.000481 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-7grz2_fb8bdb2b-bdc7-4433-be10-296fc67dc1c3/machine-api-operator/0.log" Jan 21 12:11:12 crc kubenswrapper[4824]: I0121 12:11:12.455482 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-n5vql/crc-debug-9n2zc" Jan 21 12:11:12 crc kubenswrapper[4824]: I0121 12:11:12.486294 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-n5vql/crc-debug-9n2zc"] Jan 21 12:11:12 crc kubenswrapper[4824]: I0121 12:11:12.492035 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-n5vql/crc-debug-9n2zc"] Jan 21 12:11:12 crc kubenswrapper[4824]: I0121 12:11:12.541549 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9rbp4\" (UniqueName: \"kubernetes.io/projected/d2b727cf-5339-4da5-be88-b46b2bcf5dd1-kube-api-access-9rbp4\") pod \"d2b727cf-5339-4da5-be88-b46b2bcf5dd1\" (UID: \"d2b727cf-5339-4da5-be88-b46b2bcf5dd1\") " Jan 21 12:11:12 crc kubenswrapper[4824]: I0121 12:11:12.542008 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d2b727cf-5339-4da5-be88-b46b2bcf5dd1-host\") pod \"d2b727cf-5339-4da5-be88-b46b2bcf5dd1\" (UID: \"d2b727cf-5339-4da5-be88-b46b2bcf5dd1\") " Jan 21 12:11:12 crc kubenswrapper[4824]: I0121 12:11:12.542143 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d2b727cf-5339-4da5-be88-b46b2bcf5dd1-host" (OuterVolumeSpecName: "host") pod "d2b727cf-5339-4da5-be88-b46b2bcf5dd1" (UID: "d2b727cf-5339-4da5-be88-b46b2bcf5dd1"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 21 12:11:12 crc kubenswrapper[4824]: I0121 12:11:12.542753 4824 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d2b727cf-5339-4da5-be88-b46b2bcf5dd1-host\") on node \"crc\" DevicePath \"\"" Jan 21 12:11:12 crc kubenswrapper[4824]: I0121 12:11:12.547442 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d2b727cf-5339-4da5-be88-b46b2bcf5dd1-kube-api-access-9rbp4" (OuterVolumeSpecName: "kube-api-access-9rbp4") pod "d2b727cf-5339-4da5-be88-b46b2bcf5dd1" (UID: "d2b727cf-5339-4da5-be88-b46b2bcf5dd1"). InnerVolumeSpecName "kube-api-access-9rbp4". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 12:11:12 crc kubenswrapper[4824]: I0121 12:11:12.644474 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9rbp4\" (UniqueName: \"kubernetes.io/projected/d2b727cf-5339-4da5-be88-b46b2bcf5dd1-kube-api-access-9rbp4\") on node \"crc\" DevicePath \"\"" Jan 21 12:11:13 crc kubenswrapper[4824]: I0121 12:11:13.389059 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2730b37ace706deb236dc24c607e69f539af5cbe65b803a5f377d1ca122f5af3" Jan 21 12:11:13 crc kubenswrapper[4824]: I0121 12:11:13.389151 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-n5vql/crc-debug-9n2zc" Jan 21 12:11:13 crc kubenswrapper[4824]: I0121 12:11:13.620404 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-n5vql/crc-debug-twlgv"] Jan 21 12:11:13 crc kubenswrapper[4824]: E0121 12:11:13.621579 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2b727cf-5339-4da5-be88-b46b2bcf5dd1" containerName="container-00" Jan 21 12:11:13 crc kubenswrapper[4824]: I0121 12:11:13.621676 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2b727cf-5339-4da5-be88-b46b2bcf5dd1" containerName="container-00" Jan 21 12:11:13 crc kubenswrapper[4824]: I0121 12:11:13.621983 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="d2b727cf-5339-4da5-be88-b46b2bcf5dd1" containerName="container-00" Jan 21 12:11:13 crc kubenswrapper[4824]: I0121 12:11:13.623059 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-n5vql/crc-debug-twlgv" Jan 21 12:11:13 crc kubenswrapper[4824]: I0121 12:11:13.667095 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zdz7h\" (UniqueName: \"kubernetes.io/projected/fd685561-3fbc-4a45-85aa-4acfdd089880-kube-api-access-zdz7h\") pod \"crc-debug-twlgv\" (UID: \"fd685561-3fbc-4a45-85aa-4acfdd089880\") " pod="openshift-must-gather-n5vql/crc-debug-twlgv" Jan 21 12:11:13 crc kubenswrapper[4824]: I0121 12:11:13.667170 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/fd685561-3fbc-4a45-85aa-4acfdd089880-host\") pod \"crc-debug-twlgv\" (UID: \"fd685561-3fbc-4a45-85aa-4acfdd089880\") " pod="openshift-must-gather-n5vql/crc-debug-twlgv" Jan 21 12:11:13 crc kubenswrapper[4824]: I0121 12:11:13.768698 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zdz7h\" (UniqueName: \"kubernetes.io/projected/fd685561-3fbc-4a45-85aa-4acfdd089880-kube-api-access-zdz7h\") pod \"crc-debug-twlgv\" (UID: \"fd685561-3fbc-4a45-85aa-4acfdd089880\") " pod="openshift-must-gather-n5vql/crc-debug-twlgv" Jan 21 12:11:13 crc kubenswrapper[4824]: I0121 12:11:13.768766 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/fd685561-3fbc-4a45-85aa-4acfdd089880-host\") pod \"crc-debug-twlgv\" (UID: \"fd685561-3fbc-4a45-85aa-4acfdd089880\") " pod="openshift-must-gather-n5vql/crc-debug-twlgv" Jan 21 12:11:13 crc kubenswrapper[4824]: I0121 12:11:13.768992 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/fd685561-3fbc-4a45-85aa-4acfdd089880-host\") pod \"crc-debug-twlgv\" (UID: \"fd685561-3fbc-4a45-85aa-4acfdd089880\") " pod="openshift-must-gather-n5vql/crc-debug-twlgv" Jan 21 12:11:13 crc kubenswrapper[4824]: I0121 12:11:13.795279 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zdz7h\" (UniqueName: \"kubernetes.io/projected/fd685561-3fbc-4a45-85aa-4acfdd089880-kube-api-access-zdz7h\") pod \"crc-debug-twlgv\" (UID: \"fd685561-3fbc-4a45-85aa-4acfdd089880\") " pod="openshift-must-gather-n5vql/crc-debug-twlgv" Jan 21 12:11:13 crc kubenswrapper[4824]: I0121 12:11:13.942035 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-n5vql/crc-debug-twlgv" Jan 21 12:11:14 crc kubenswrapper[4824]: I0121 12:11:14.058694 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d2b727cf-5339-4da5-be88-b46b2bcf5dd1" path="/var/lib/kubelet/pods/d2b727cf-5339-4da5-be88-b46b2bcf5dd1/volumes" Jan 21 12:11:14 crc kubenswrapper[4824]: I0121 12:11:14.401158 4824 generic.go:334] "Generic (PLEG): container finished" podID="fd685561-3fbc-4a45-85aa-4acfdd089880" containerID="07bf87c3a875d48665a8353210d054b6f5a80d43f1b2712cff93bd33cf233d5b" exitCode=0 Jan 21 12:11:14 crc kubenswrapper[4824]: I0121 12:11:14.401249 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-n5vql/crc-debug-twlgv" event={"ID":"fd685561-3fbc-4a45-85aa-4acfdd089880","Type":"ContainerDied","Data":"07bf87c3a875d48665a8353210d054b6f5a80d43f1b2712cff93bd33cf233d5b"} Jan 21 12:11:14 crc kubenswrapper[4824]: I0121 12:11:14.401336 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-n5vql/crc-debug-twlgv" event={"ID":"fd685561-3fbc-4a45-85aa-4acfdd089880","Type":"ContainerStarted","Data":"d22ae759971e42ba5d4c5d7745735291590008b6acb73204a7c4d5bd657d6073"} Jan 21 12:11:15 crc kubenswrapper[4824]: I0121 12:11:15.057539 4824 scope.go:117] "RemoveContainer" containerID="719b0a497b4be25b481cf2d3456b6c515017599e2aeafd1f30d22bffc35f79d4" Jan 21 12:11:15 crc kubenswrapper[4824]: E0121 12:11:15.058931 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 12:11:15 crc kubenswrapper[4824]: I0121 12:11:15.484262 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-n5vql/crc-debug-twlgv" Jan 21 12:11:15 crc kubenswrapper[4824]: I0121 12:11:15.506115 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/fd685561-3fbc-4a45-85aa-4acfdd089880-host\") pod \"fd685561-3fbc-4a45-85aa-4acfdd089880\" (UID: \"fd685561-3fbc-4a45-85aa-4acfdd089880\") " Jan 21 12:11:15 crc kubenswrapper[4824]: I0121 12:11:15.506193 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zdz7h\" (UniqueName: \"kubernetes.io/projected/fd685561-3fbc-4a45-85aa-4acfdd089880-kube-api-access-zdz7h\") pod \"fd685561-3fbc-4a45-85aa-4acfdd089880\" (UID: \"fd685561-3fbc-4a45-85aa-4acfdd089880\") " Jan 21 12:11:15 crc kubenswrapper[4824]: I0121 12:11:15.509518 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fd685561-3fbc-4a45-85aa-4acfdd089880-host" (OuterVolumeSpecName: "host") pod "fd685561-3fbc-4a45-85aa-4acfdd089880" (UID: "fd685561-3fbc-4a45-85aa-4acfdd089880"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 21 12:11:15 crc kubenswrapper[4824]: I0121 12:11:15.514296 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fd685561-3fbc-4a45-85aa-4acfdd089880-kube-api-access-zdz7h" (OuterVolumeSpecName: "kube-api-access-zdz7h") pod "fd685561-3fbc-4a45-85aa-4acfdd089880" (UID: "fd685561-3fbc-4a45-85aa-4acfdd089880"). InnerVolumeSpecName "kube-api-access-zdz7h". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 12:11:15 crc kubenswrapper[4824]: I0121 12:11:15.608243 4824 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/fd685561-3fbc-4a45-85aa-4acfdd089880-host\") on node \"crc\" DevicePath \"\"" Jan 21 12:11:15 crc kubenswrapper[4824]: I0121 12:11:15.608277 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zdz7h\" (UniqueName: \"kubernetes.io/projected/fd685561-3fbc-4a45-85aa-4acfdd089880-kube-api-access-zdz7h\") on node \"crc\" DevicePath \"\"" Jan 21 12:11:16 crc kubenswrapper[4824]: I0121 12:11:16.199472 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-n5vql/crc-debug-twlgv"] Jan 21 12:11:16 crc kubenswrapper[4824]: I0121 12:11:16.206344 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-n5vql/crc-debug-twlgv"] Jan 21 12:11:16 crc kubenswrapper[4824]: I0121 12:11:16.419497 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d22ae759971e42ba5d4c5d7745735291590008b6acb73204a7c4d5bd657d6073" Jan 21 12:11:16 crc kubenswrapper[4824]: I0121 12:11:16.419565 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-n5vql/crc-debug-twlgv" Jan 21 12:11:17 crc kubenswrapper[4824]: I0121 12:11:17.329762 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-n5vql/crc-debug-vzzhf"] Jan 21 12:11:17 crc kubenswrapper[4824]: E0121 12:11:17.330447 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fd685561-3fbc-4a45-85aa-4acfdd089880" containerName="container-00" Jan 21 12:11:17 crc kubenswrapper[4824]: I0121 12:11:17.330462 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="fd685561-3fbc-4a45-85aa-4acfdd089880" containerName="container-00" Jan 21 12:11:17 crc kubenswrapper[4824]: I0121 12:11:17.330636 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="fd685561-3fbc-4a45-85aa-4acfdd089880" containerName="container-00" Jan 21 12:11:17 crc kubenswrapper[4824]: I0121 12:11:17.331225 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-n5vql/crc-debug-vzzhf" Jan 21 12:11:17 crc kubenswrapper[4824]: I0121 12:11:17.456472 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f96e30dd-f9f8-4f4f-9dee-27b8275b17c4-host\") pod \"crc-debug-vzzhf\" (UID: \"f96e30dd-f9f8-4f4f-9dee-27b8275b17c4\") " pod="openshift-must-gather-n5vql/crc-debug-vzzhf" Jan 21 12:11:17 crc kubenswrapper[4824]: I0121 12:11:17.456580 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7zbzk\" (UniqueName: \"kubernetes.io/projected/f96e30dd-f9f8-4f4f-9dee-27b8275b17c4-kube-api-access-7zbzk\") pod \"crc-debug-vzzhf\" (UID: \"f96e30dd-f9f8-4f4f-9dee-27b8275b17c4\") " pod="openshift-must-gather-n5vql/crc-debug-vzzhf" Jan 21 12:11:17 crc kubenswrapper[4824]: I0121 12:11:17.557748 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f96e30dd-f9f8-4f4f-9dee-27b8275b17c4-host\") pod \"crc-debug-vzzhf\" (UID: \"f96e30dd-f9f8-4f4f-9dee-27b8275b17c4\") " pod="openshift-must-gather-n5vql/crc-debug-vzzhf" Jan 21 12:11:17 crc kubenswrapper[4824]: I0121 12:11:17.557950 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f96e30dd-f9f8-4f4f-9dee-27b8275b17c4-host\") pod \"crc-debug-vzzhf\" (UID: \"f96e30dd-f9f8-4f4f-9dee-27b8275b17c4\") " pod="openshift-must-gather-n5vql/crc-debug-vzzhf" Jan 21 12:11:17 crc kubenswrapper[4824]: I0121 12:11:17.558087 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7zbzk\" (UniqueName: \"kubernetes.io/projected/f96e30dd-f9f8-4f4f-9dee-27b8275b17c4-kube-api-access-7zbzk\") pod \"crc-debug-vzzhf\" (UID: \"f96e30dd-f9f8-4f4f-9dee-27b8275b17c4\") " pod="openshift-must-gather-n5vql/crc-debug-vzzhf" Jan 21 12:11:17 crc kubenswrapper[4824]: I0121 12:11:17.577510 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7zbzk\" (UniqueName: \"kubernetes.io/projected/f96e30dd-f9f8-4f4f-9dee-27b8275b17c4-kube-api-access-7zbzk\") pod \"crc-debug-vzzhf\" (UID: \"f96e30dd-f9f8-4f4f-9dee-27b8275b17c4\") " pod="openshift-must-gather-n5vql/crc-debug-vzzhf" Jan 21 12:11:17 crc kubenswrapper[4824]: I0121 12:11:17.647765 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-n5vql/crc-debug-vzzhf" Jan 21 12:11:18 crc kubenswrapper[4824]: I0121 12:11:18.061941 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fd685561-3fbc-4a45-85aa-4acfdd089880" path="/var/lib/kubelet/pods/fd685561-3fbc-4a45-85aa-4acfdd089880/volumes" Jan 21 12:11:18 crc kubenswrapper[4824]: I0121 12:11:18.450485 4824 generic.go:334] "Generic (PLEG): container finished" podID="f96e30dd-f9f8-4f4f-9dee-27b8275b17c4" containerID="28a96abcbc61a2ef1455e05edd37bad363143f763a99e610aa47efbbc2d11cc9" exitCode=0 Jan 21 12:11:18 crc kubenswrapper[4824]: I0121 12:11:18.450577 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-n5vql/crc-debug-vzzhf" event={"ID":"f96e30dd-f9f8-4f4f-9dee-27b8275b17c4","Type":"ContainerDied","Data":"28a96abcbc61a2ef1455e05edd37bad363143f763a99e610aa47efbbc2d11cc9"} Jan 21 12:11:18 crc kubenswrapper[4824]: I0121 12:11:18.451042 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-n5vql/crc-debug-vzzhf" event={"ID":"f96e30dd-f9f8-4f4f-9dee-27b8275b17c4","Type":"ContainerStarted","Data":"feaebf64732df5c5fc007c0740c2d69db7d3fdf4dcdabb5f6f984ad0740b5eae"} Jan 21 12:11:18 crc kubenswrapper[4824]: I0121 12:11:18.490768 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-n5vql/crc-debug-vzzhf"] Jan 21 12:11:18 crc kubenswrapper[4824]: I0121 12:11:18.498648 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-n5vql/crc-debug-vzzhf"] Jan 21 12:11:18 crc kubenswrapper[4824]: I0121 12:11:18.972175 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-858654f9db-qhlzr_b9405067-2dfa-46f9-9471-fe456b3ca5aa/cert-manager-controller/0.log" Jan 21 12:11:18 crc kubenswrapper[4824]: I0121 12:11:18.986371 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-cf98fcc89-xjwkp_d783bb77-baab-4e1c-99e9-ddba7ef78fc2/cert-manager-cainjector/0.log" Jan 21 12:11:18 crc kubenswrapper[4824]: I0121 12:11:18.996710 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-687f57d79b-r6qj7_b12640f9-bf75-4056-9c20-a75d63a0e177/cert-manager-webhook/0.log" Jan 21 12:11:19 crc kubenswrapper[4824]: I0121 12:11:19.549833 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-n5vql/crc-debug-vzzhf" Jan 21 12:11:19 crc kubenswrapper[4824]: I0121 12:11:19.709423 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f96e30dd-f9f8-4f4f-9dee-27b8275b17c4-host\") pod \"f96e30dd-f9f8-4f4f-9dee-27b8275b17c4\" (UID: \"f96e30dd-f9f8-4f4f-9dee-27b8275b17c4\") " Jan 21 12:11:19 crc kubenswrapper[4824]: I0121 12:11:19.709532 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7zbzk\" (UniqueName: \"kubernetes.io/projected/f96e30dd-f9f8-4f4f-9dee-27b8275b17c4-kube-api-access-7zbzk\") pod \"f96e30dd-f9f8-4f4f-9dee-27b8275b17c4\" (UID: \"f96e30dd-f9f8-4f4f-9dee-27b8275b17c4\") " Jan 21 12:11:19 crc kubenswrapper[4824]: I0121 12:11:19.709546 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f96e30dd-f9f8-4f4f-9dee-27b8275b17c4-host" (OuterVolumeSpecName: "host") pod "f96e30dd-f9f8-4f4f-9dee-27b8275b17c4" (UID: "f96e30dd-f9f8-4f4f-9dee-27b8275b17c4"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 21 12:11:19 crc kubenswrapper[4824]: I0121 12:11:19.711054 4824 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f96e30dd-f9f8-4f4f-9dee-27b8275b17c4-host\") on node \"crc\" DevicePath \"\"" Jan 21 12:11:19 crc kubenswrapper[4824]: I0121 12:11:19.723012 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f96e30dd-f9f8-4f4f-9dee-27b8275b17c4-kube-api-access-7zbzk" (OuterVolumeSpecName: "kube-api-access-7zbzk") pod "f96e30dd-f9f8-4f4f-9dee-27b8275b17c4" (UID: "f96e30dd-f9f8-4f4f-9dee-27b8275b17c4"). InnerVolumeSpecName "kube-api-access-7zbzk". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 12:11:19 crc kubenswrapper[4824]: I0121 12:11:19.813303 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7zbzk\" (UniqueName: \"kubernetes.io/projected/f96e30dd-f9f8-4f4f-9dee-27b8275b17c4-kube-api-access-7zbzk\") on node \"crc\" DevicePath \"\"" Jan 21 12:11:20 crc kubenswrapper[4824]: I0121 12:11:20.058233 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f96e30dd-f9f8-4f4f-9dee-27b8275b17c4" path="/var/lib/kubelet/pods/f96e30dd-f9f8-4f4f-9dee-27b8275b17c4/volumes" Jan 21 12:11:20 crc kubenswrapper[4824]: I0121 12:11:20.473216 4824 scope.go:117] "RemoveContainer" containerID="28a96abcbc61a2ef1455e05edd37bad363143f763a99e610aa47efbbc2d11cc9" Jan 21 12:11:20 crc kubenswrapper[4824]: I0121 12:11:20.473256 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-n5vql/crc-debug-vzzhf" Jan 21 12:11:23 crc kubenswrapper[4824]: I0121 12:11:23.687036 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-7754f76f8b-xjqs5_1575ab9d-016f-403a-990b-3b4c4d53f6f1/nmstate-console-plugin/0.log" Jan 21 12:11:23 crc kubenswrapper[4824]: I0121 12:11:23.707484 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-qjpv9_a01ecff0-9051-4ddb-8c8e-a4f379b6c823/nmstate-handler/0.log" Jan 21 12:11:23 crc kubenswrapper[4824]: I0121 12:11:23.716586 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-54757c584b-w2j6n_30b70686-bf05-4798-b5dc-3c7dc9cee56a/nmstate-metrics/0.log" Jan 21 12:11:23 crc kubenswrapper[4824]: I0121 12:11:23.724266 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-54757c584b-w2j6n_30b70686-bf05-4798-b5dc-3c7dc9cee56a/kube-rbac-proxy/0.log" Jan 21 12:11:23 crc kubenswrapper[4824]: I0121 12:11:23.734280 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-646758c888-zmc4k_df60ca53-0573-4a6a-9235-86e0a1c17f91/nmstate-operator/0.log" Jan 21 12:11:23 crc kubenswrapper[4824]: I0121 12:11:23.745680 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-8474b5b9d8-h6kkh_edf54909-b53e-44de-90bc-355e23bbc767/nmstate-webhook/0.log" Jan 21 12:11:27 crc kubenswrapper[4824]: I0121 12:11:27.049942 4824 scope.go:117] "RemoveContainer" containerID="719b0a497b4be25b481cf2d3456b6c515017599e2aeafd1f30d22bffc35f79d4" Jan 21 12:11:27 crc kubenswrapper[4824]: E0121 12:11:27.050754 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 12:11:33 crc kubenswrapper[4824]: I0121 12:11:33.414382 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-6968d8fdc4-bbv7p_d30695b8-6f6a-48f0-88ed-9181484634b8/controller/0.log" Jan 21 12:11:33 crc kubenswrapper[4824]: I0121 12:11:33.420870 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-6968d8fdc4-bbv7p_d30695b8-6f6a-48f0-88ed-9181484634b8/kube-rbac-proxy/0.log" Jan 21 12:11:33 crc kubenswrapper[4824]: I0121 12:11:33.435584 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9q9wj_e38576f0-c423-497a-a30d-60c2e8edd93a/controller/0.log" Jan 21 12:11:34 crc kubenswrapper[4824]: I0121 12:11:34.846277 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9q9wj_e38576f0-c423-497a-a30d-60c2e8edd93a/frr/0.log" Jan 21 12:11:34 crc kubenswrapper[4824]: I0121 12:11:34.858697 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9q9wj_e38576f0-c423-497a-a30d-60c2e8edd93a/reloader/0.log" Jan 21 12:11:34 crc kubenswrapper[4824]: I0121 12:11:34.863192 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9q9wj_e38576f0-c423-497a-a30d-60c2e8edd93a/frr-metrics/0.log" Jan 21 12:11:34 crc kubenswrapper[4824]: I0121 12:11:34.871832 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9q9wj_e38576f0-c423-497a-a30d-60c2e8edd93a/kube-rbac-proxy/0.log" Jan 21 12:11:34 crc kubenswrapper[4824]: I0121 12:11:34.878861 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9q9wj_e38576f0-c423-497a-a30d-60c2e8edd93a/kube-rbac-proxy-frr/0.log" Jan 21 12:11:34 crc kubenswrapper[4824]: I0121 12:11:34.885819 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9q9wj_e38576f0-c423-497a-a30d-60c2e8edd93a/cp-frr-files/0.log" Jan 21 12:11:34 crc kubenswrapper[4824]: I0121 12:11:34.892994 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9q9wj_e38576f0-c423-497a-a30d-60c2e8edd93a/cp-reloader/0.log" Jan 21 12:11:34 crc kubenswrapper[4824]: I0121 12:11:34.900115 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9q9wj_e38576f0-c423-497a-a30d-60c2e8edd93a/cp-metrics/0.log" Jan 21 12:11:34 crc kubenswrapper[4824]: I0121 12:11:34.909707 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7df86c4f6c-fdblb_d71f993c-b2c7-47af-884f-4cbe6dcf3c1a/frr-k8s-webhook-server/0.log" Jan 21 12:11:34 crc kubenswrapper[4824]: I0121 12:11:34.934678 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-5c58bcfcff-69lct_714d2131-afc8-4828-840f-e2f0ce727d5d/manager/0.log" Jan 21 12:11:34 crc kubenswrapper[4824]: I0121 12:11:34.943524 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-68bc9bfcb6-hzmp5_81fed5ab-ec3d-46f0-8998-199393bffba0/webhook-server/0.log" Jan 21 12:11:35 crc kubenswrapper[4824]: I0121 12:11:35.312284 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-8tsb9_baf5db51-d3ac-4b97-9c5f-afc598d59ab7/speaker/0.log" Jan 21 12:11:35 crc kubenswrapper[4824]: I0121 12:11:35.321390 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-8tsb9_baf5db51-d3ac-4b97-9c5f-afc598d59ab7/kube-rbac-proxy/0.log" Jan 21 12:11:38 crc kubenswrapper[4824]: I0121 12:11:38.320775 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcwqkxq_d9005d95-8246-435d-aaea-5c60d58175c6/extract/0.log" Jan 21 12:11:38 crc kubenswrapper[4824]: I0121 12:11:38.328337 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcwqkxq_d9005d95-8246-435d-aaea-5c60d58175c6/util/0.log" Jan 21 12:11:38 crc kubenswrapper[4824]: I0121 12:11:38.335782 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcwqkxq_d9005d95-8246-435d-aaea-5c60d58175c6/pull/0.log" Jan 21 12:11:38 crc kubenswrapper[4824]: I0121 12:11:38.343532 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713g7gj4_d2dd714a-8d35-49a1-b6c2-7bec802eed41/extract/0.log" Jan 21 12:11:38 crc kubenswrapper[4824]: I0121 12:11:38.349821 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713g7gj4_d2dd714a-8d35-49a1-b6c2-7bec802eed41/util/0.log" Jan 21 12:11:38 crc kubenswrapper[4824]: I0121 12:11:38.355868 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713g7gj4_d2dd714a-8d35-49a1-b6c2-7bec802eed41/pull/0.log" Jan 21 12:11:38 crc kubenswrapper[4824]: I0121 12:11:38.781238 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-rgl7w_c321fe0f-f966-4865-a619-6f5f26e9bb50/registry-server/0.log" Jan 21 12:11:38 crc kubenswrapper[4824]: I0121 12:11:38.788528 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-rgl7w_c321fe0f-f966-4865-a619-6f5f26e9bb50/extract-utilities/0.log" Jan 21 12:11:38 crc kubenswrapper[4824]: I0121 12:11:38.793306 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-rgl7w_c321fe0f-f966-4865-a619-6f5f26e9bb50/extract-content/0.log" Jan 21 12:11:39 crc kubenswrapper[4824]: I0121 12:11:39.318548 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-m4srg_53e1053f-f40b-4afc-ad9c-b07d46fc1f94/registry-server/0.log" Jan 21 12:11:39 crc kubenswrapper[4824]: I0121 12:11:39.323196 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-m4srg_53e1053f-f40b-4afc-ad9c-b07d46fc1f94/extract-utilities/0.log" Jan 21 12:11:39 crc kubenswrapper[4824]: I0121 12:11:39.329719 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-m4srg_53e1053f-f40b-4afc-ad9c-b07d46fc1f94/extract-content/0.log" Jan 21 12:11:39 crc kubenswrapper[4824]: I0121 12:11:39.342113 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-xcxs9_b4c29122-d9b2-4f85-8b29-c0e7d888667f/marketplace-operator/0.log" Jan 21 12:11:39 crc kubenswrapper[4824]: I0121 12:11:39.493118 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-k2t4f_46423d25-afc7-4a50-ba21-5f7c7b7012f2/registry-server/0.log" Jan 21 12:11:39 crc kubenswrapper[4824]: I0121 12:11:39.497301 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-k2t4f_46423d25-afc7-4a50-ba21-5f7c7b7012f2/extract-utilities/0.log" Jan 21 12:11:39 crc kubenswrapper[4824]: I0121 12:11:39.502769 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-k2t4f_46423d25-afc7-4a50-ba21-5f7c7b7012f2/extract-content/0.log" Jan 21 12:11:40 crc kubenswrapper[4824]: I0121 12:11:40.049080 4824 scope.go:117] "RemoveContainer" containerID="719b0a497b4be25b481cf2d3456b6c515017599e2aeafd1f30d22bffc35f79d4" Jan 21 12:11:40 crc kubenswrapper[4824]: E0121 12:11:40.049752 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 12:11:40 crc kubenswrapper[4824]: I0121 12:11:40.064615 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-cmbbh_3447df50-807b-487c-9acf-2a2250780bdc/registry-server/0.log" Jan 21 12:11:40 crc kubenswrapper[4824]: I0121 12:11:40.071905 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-cmbbh_3447df50-807b-487c-9acf-2a2250780bdc/extract-utilities/0.log" Jan 21 12:11:40 crc kubenswrapper[4824]: I0121 12:11:40.077211 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-cmbbh_3447df50-807b-487c-9acf-2a2250780bdc/extract-content/0.log" Jan 21 12:11:53 crc kubenswrapper[4824]: I0121 12:11:53.049300 4824 scope.go:117] "RemoveContainer" containerID="719b0a497b4be25b481cf2d3456b6c515017599e2aeafd1f30d22bffc35f79d4" Jan 21 12:11:53 crc kubenswrapper[4824]: E0121 12:11:53.051144 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 12:12:04 crc kubenswrapper[4824]: I0121 12:12:04.049759 4824 scope.go:117] "RemoveContainer" containerID="719b0a497b4be25b481cf2d3456b6c515017599e2aeafd1f30d22bffc35f79d4" Jan 21 12:12:04 crc kubenswrapper[4824]: E0121 12:12:04.050854 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 12:12:16 crc kubenswrapper[4824]: I0121 12:12:16.050206 4824 scope.go:117] "RemoveContainer" containerID="719b0a497b4be25b481cf2d3456b6c515017599e2aeafd1f30d22bffc35f79d4" Jan 21 12:12:16 crc kubenswrapper[4824]: E0121 12:12:16.051091 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 12:12:29 crc kubenswrapper[4824]: I0121 12:12:29.049593 4824 scope.go:117] "RemoveContainer" containerID="719b0a497b4be25b481cf2d3456b6c515017599e2aeafd1f30d22bffc35f79d4" Jan 21 12:12:29 crc kubenswrapper[4824]: E0121 12:12:29.050330 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 12:12:37 crc kubenswrapper[4824]: I0121 12:12:37.942266 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-6968d8fdc4-bbv7p_d30695b8-6f6a-48f0-88ed-9181484634b8/controller/0.log" Jan 21 12:12:37 crc kubenswrapper[4824]: I0121 12:12:37.949894 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-6968d8fdc4-bbv7p_d30695b8-6f6a-48f0-88ed-9181484634b8/kube-rbac-proxy/0.log" Jan 21 12:12:37 crc kubenswrapper[4824]: I0121 12:12:37.977796 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9q9wj_e38576f0-c423-497a-a30d-60c2e8edd93a/controller/0.log" Jan 21 12:12:37 crc kubenswrapper[4824]: I0121 12:12:37.999266 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-858654f9db-qhlzr_b9405067-2dfa-46f9-9471-fe456b3ca5aa/cert-manager-controller/0.log" Jan 21 12:12:38 crc kubenswrapper[4824]: I0121 12:12:38.017230 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-cf98fcc89-xjwkp_d783bb77-baab-4e1c-99e9-ddba7ef78fc2/cert-manager-cainjector/0.log" Jan 21 12:12:38 crc kubenswrapper[4824]: I0121 12:12:38.034784 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-687f57d79b-r6qj7_b12640f9-bf75-4056-9c20-a75d63a0e177/cert-manager-webhook/0.log" Jan 21 12:12:38 crc kubenswrapper[4824]: I0121 12:12:38.929005 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7ddb5c749-bvptq_8ed6bb2c-2770-4785-bf38-9b6fd0bd8472/manager/0.log" Jan 21 12:12:38 crc kubenswrapper[4824]: I0121 12:12:38.981668 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-9b68f5989-24r54_a25733d4-a5f8-4c97-a897-8d96e637c253/manager/0.log" Jan 21 12:12:38 crc kubenswrapper[4824]: I0121 12:12:38.997901 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_db4b9aacd6f51b5aff2d66474158631eb1e58fb4d1358e39c8aaaca1bdn7cxr_72ab26cb-149e-4825-9587-3acfa66a368f/extract/0.log" Jan 21 12:12:39 crc kubenswrapper[4824]: I0121 12:12:39.011190 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_db4b9aacd6f51b5aff2d66474158631eb1e58fb4d1358e39c8aaaca1bdn7cxr_72ab26cb-149e-4825-9587-3acfa66a368f/util/0.log" Jan 21 12:12:39 crc kubenswrapper[4824]: I0121 12:12:39.024025 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_db4b9aacd6f51b5aff2d66474158631eb1e58fb4d1358e39c8aaaca1bdn7cxr_72ab26cb-149e-4825-9587-3acfa66a368f/pull/0.log" Jan 21 12:12:39 crc kubenswrapper[4824]: I0121 12:12:39.035187 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-9f958b845-wtfj4_14dd1f12-3943-42e4-be1b-c9e37e49d21b/manager/0.log" Jan 21 12:12:39 crc kubenswrapper[4824]: I0121 12:12:39.147922 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-c6994669c-hvq6s_b7053a68-ce89-48b6-9b5d-1c6dac813a75/manager/0.log" Jan 21 12:12:39 crc kubenswrapper[4824]: I0121 12:12:39.235050 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5fbc59b5b6-hgmf4_944c3123-9063-4731-9581-5b683741edc4/manager/0.log" Jan 21 12:12:39 crc kubenswrapper[4824]: I0121 12:12:39.244558 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-77d5c5b54f-mp5ml_2463afb1-0271-4ab3-ad1a-9329ecaeec4d/manager/0.log" Jan 21 12:12:39 crc kubenswrapper[4824]: I0121 12:12:39.563479 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9q9wj_e38576f0-c423-497a-a30d-60c2e8edd93a/frr/0.log" Jan 21 12:12:39 crc kubenswrapper[4824]: I0121 12:12:39.574073 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9q9wj_e38576f0-c423-497a-a30d-60c2e8edd93a/reloader/0.log" Jan 21 12:12:39 crc kubenswrapper[4824]: I0121 12:12:39.579345 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9q9wj_e38576f0-c423-497a-a30d-60c2e8edd93a/frr-metrics/0.log" Jan 21 12:12:39 crc kubenswrapper[4824]: I0121 12:12:39.588829 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9q9wj_e38576f0-c423-497a-a30d-60c2e8edd93a/kube-rbac-proxy/0.log" Jan 21 12:12:39 crc kubenswrapper[4824]: I0121 12:12:39.595141 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9q9wj_e38576f0-c423-497a-a30d-60c2e8edd93a/kube-rbac-proxy-frr/0.log" Jan 21 12:12:39 crc kubenswrapper[4824]: I0121 12:12:39.601802 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9q9wj_e38576f0-c423-497a-a30d-60c2e8edd93a/cp-frr-files/0.log" Jan 21 12:12:39 crc kubenswrapper[4824]: I0121 12:12:39.608900 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9q9wj_e38576f0-c423-497a-a30d-60c2e8edd93a/cp-reloader/0.log" Jan 21 12:12:39 crc kubenswrapper[4824]: I0121 12:12:39.616063 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9q9wj_e38576f0-c423-497a-a30d-60c2e8edd93a/cp-metrics/0.log" Jan 21 12:12:39 crc kubenswrapper[4824]: I0121 12:12:39.624051 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7df86c4f6c-fdblb_d71f993c-b2c7-47af-884f-4cbe6dcf3c1a/frr-k8s-webhook-server/0.log" Jan 21 12:12:39 crc kubenswrapper[4824]: I0121 12:12:39.626261 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-77c48c7859-dzn7w_57e1fe13-b2c8-4711-9efe-1ebbc7e9d12e/manager/0.log" Jan 21 12:12:39 crc kubenswrapper[4824]: I0121 12:12:39.635421 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-78757b4889-gkwjs_fc911701-7c15-4a20-b89e-709521bdd3a1/manager/0.log" Jan 21 12:12:39 crc kubenswrapper[4824]: I0121 12:12:39.652276 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-5c58bcfcff-69lct_714d2131-afc8-4828-840f-e2f0ce727d5d/manager/0.log" Jan 21 12:12:39 crc kubenswrapper[4824]: I0121 12:12:39.660732 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-68bc9bfcb6-hzmp5_81fed5ab-ec3d-46f0-8998-199393bffba0/webhook-server/0.log" Jan 21 12:12:39 crc kubenswrapper[4824]: I0121 12:12:39.719452 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-767fdc4f47-5mhmd_2e5efa9b-40bc-4860-967d-aa37e1838f81/manager/0.log" Jan 21 12:12:39 crc kubenswrapper[4824]: I0121 12:12:39.731144 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-864f6b75bf-c7vsf_18017a19-1676-4fe9-a97e-ae878f423847/manager/0.log" Jan 21 12:12:39 crc kubenswrapper[4824]: I0121 12:12:39.771684 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-c87fff755-d5hn9_19b1c150-f5ed-4432-bfc2-717b78eebc7e/manager/0.log" Jan 21 12:12:39 crc kubenswrapper[4824]: I0121 12:12:39.829244 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-cb4666565-7w8p4_8c74d3da-d870-45e9-96b9-70eb2395c977/manager/0.log" Jan 21 12:12:39 crc kubenswrapper[4824]: I0121 12:12:39.949745 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-65849867d6-8ngx6_b38ac60b-1643-44d9-a1fe-5f75129247dc/manager/0.log" Jan 21 12:12:39 crc kubenswrapper[4824]: I0121 12:12:39.961674 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-7fc9b76cf6-tncbx_5429da27-a618-44da-8aeb-12b917af8734/manager/0.log" Jan 21 12:12:39 crc kubenswrapper[4824]: I0121 12:12:39.987431 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-6b68b8b854zxcbl_3fa95966-3147-4e50-8ed8-f702a87992c4/manager/0.log" Jan 21 12:12:40 crc kubenswrapper[4824]: I0121 12:12:40.152774 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-8tsb9_baf5db51-d3ac-4b97-9c5f-afc598d59ab7/speaker/0.log" Jan 21 12:12:40 crc kubenswrapper[4824]: I0121 12:12:40.163085 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-8tsb9_baf5db51-d3ac-4b97-9c5f-afc598d59ab7/kube-rbac-proxy/0.log" Jan 21 12:12:40 crc kubenswrapper[4824]: I0121 12:12:40.201718 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-init-dbdbd6855-gkm5b_35334861-8e16-4b16-8cf5-167a049836af/operator/0.log" Jan 21 12:12:41 crc kubenswrapper[4824]: I0121 12:12:41.050375 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-858654f9db-qhlzr_b9405067-2dfa-46f9-9471-fe456b3ca5aa/cert-manager-controller/0.log" Jan 21 12:12:41 crc kubenswrapper[4824]: I0121 12:12:41.065354 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-cf98fcc89-xjwkp_d783bb77-baab-4e1c-99e9-ddba7ef78fc2/cert-manager-cainjector/0.log" Jan 21 12:12:41 crc kubenswrapper[4824]: I0121 12:12:41.072590 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-687f57d79b-r6qj7_b12640f9-bf75-4056-9c20-a75d63a0e177/cert-manager-webhook/0.log" Jan 21 12:12:41 crc kubenswrapper[4824]: I0121 12:12:41.353728 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-74d84f758c-vjgps_c1128e01-8b7c-4742-b1da-dbf06aa8b4d8/manager/0.log" Jan 21 12:12:41 crc kubenswrapper[4824]: I0121 12:12:41.362654 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-zggxm_78eb1906-d9e6-49f7-a512-ed3cc0a07da7/registry-server/0.log" Jan 21 12:12:41 crc kubenswrapper[4824]: I0121 12:12:41.407209 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-55db956ddc-78z2q_f5639f17-3a5b-4ef4-bd6e-6eeb3274636a/manager/0.log" Jan 21 12:12:41 crc kubenswrapper[4824]: I0121 12:12:41.435510 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-686df47fcb-c8xvs_d8b36106-8f21-4873-890d-c1879ea49068/manager/0.log" Jan 21 12:12:41 crc kubenswrapper[4824]: I0121 12:12:41.453700 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-94lqx_782093ab-751c-4e38-bba3-c5f76308c82e/operator/0.log" Jan 21 12:12:41 crc kubenswrapper[4824]: I0121 12:12:41.473920 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-85dd56d4cc-wlrdt_b4cb80e6-c720-4875-9848-7d9597a354e4/manager/0.log" Jan 21 12:12:41 crc kubenswrapper[4824]: I0121 12:12:41.538914 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-5f8f495fcf-6rz5h_db84b172-59d8-4995-815b-50b492283bee/manager/0.log" Jan 21 12:12:41 crc kubenswrapper[4824]: I0121 12:12:41.549701 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-7cd8bc9dbb-xkp8x_52491e7b-cd8c-403a-942c-312debd6f1c6/manager/0.log" Jan 21 12:12:41 crc kubenswrapper[4824]: I0121 12:12:41.558341 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-64cd966744-9w4ch_0715aa91-2ae5-4e07-8388-c37def4c7743/manager/0.log" Jan 21 12:12:41 crc kubenswrapper[4824]: I0121 12:12:41.706589 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-r65dm_c117483a-fe5c-423b-995e-9c2e7f59566a/control-plane-machine-set-operator/0.log" Jan 21 12:12:41 crc kubenswrapper[4824]: I0121 12:12:41.715792 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-7grz2_fb8bdb2b-bdc7-4433-be10-296fc67dc1c3/kube-rbac-proxy/0.log" Jan 21 12:12:41 crc kubenswrapper[4824]: I0121 12:12:41.727399 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-7grz2_fb8bdb2b-bdc7-4433-be10-296fc67dc1c3/machine-api-operator/0.log" Jan 21 12:12:42 crc kubenswrapper[4824]: I0121 12:12:42.049261 4824 scope.go:117] "RemoveContainer" containerID="719b0a497b4be25b481cf2d3456b6c515017599e2aeafd1f30d22bffc35f79d4" Jan 21 12:12:42 crc kubenswrapper[4824]: E0121 12:12:42.049570 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 12:12:42 crc kubenswrapper[4824]: I0121 12:12:42.413669 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7ddb5c749-bvptq_8ed6bb2c-2770-4785-bf38-9b6fd0bd8472/manager/0.log" Jan 21 12:12:42 crc kubenswrapper[4824]: I0121 12:12:42.454811 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-9b68f5989-24r54_a25733d4-a5f8-4c97-a897-8d96e637c253/manager/0.log" Jan 21 12:12:42 crc kubenswrapper[4824]: I0121 12:12:42.466339 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_db4b9aacd6f51b5aff2d66474158631eb1e58fb4d1358e39c8aaaca1bdn7cxr_72ab26cb-149e-4825-9587-3acfa66a368f/extract/0.log" Jan 21 12:12:42 crc kubenswrapper[4824]: I0121 12:12:42.471416 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_db4b9aacd6f51b5aff2d66474158631eb1e58fb4d1358e39c8aaaca1bdn7cxr_72ab26cb-149e-4825-9587-3acfa66a368f/util/0.log" Jan 21 12:12:42 crc kubenswrapper[4824]: I0121 12:12:42.480094 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_db4b9aacd6f51b5aff2d66474158631eb1e58fb4d1358e39c8aaaca1bdn7cxr_72ab26cb-149e-4825-9587-3acfa66a368f/pull/0.log" Jan 21 12:12:42 crc kubenswrapper[4824]: I0121 12:12:42.490477 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-9f958b845-wtfj4_14dd1f12-3943-42e4-be1b-c9e37e49d21b/manager/0.log" Jan 21 12:12:42 crc kubenswrapper[4824]: I0121 12:12:42.559884 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-c6994669c-hvq6s_b7053a68-ce89-48b6-9b5d-1c6dac813a75/manager/0.log" Jan 21 12:12:42 crc kubenswrapper[4824]: I0121 12:12:42.637522 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5fbc59b5b6-hgmf4_944c3123-9063-4731-9581-5b683741edc4/manager/0.log" Jan 21 12:12:42 crc kubenswrapper[4824]: I0121 12:12:42.654103 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-77d5c5b54f-mp5ml_2463afb1-0271-4ab3-ad1a-9329ecaeec4d/manager/0.log" Jan 21 12:12:42 crc kubenswrapper[4824]: I0121 12:12:42.803688 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-7754f76f8b-xjqs5_1575ab9d-016f-403a-990b-3b4c4d53f6f1/nmstate-console-plugin/0.log" Jan 21 12:12:42 crc kubenswrapper[4824]: I0121 12:12:42.822439 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-qjpv9_a01ecff0-9051-4ddb-8c8e-a4f379b6c823/nmstate-handler/0.log" Jan 21 12:12:42 crc kubenswrapper[4824]: I0121 12:12:42.831287 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-54757c584b-w2j6n_30b70686-bf05-4798-b5dc-3c7dc9cee56a/nmstate-metrics/0.log" Jan 21 12:12:42 crc kubenswrapper[4824]: I0121 12:12:42.838407 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-54757c584b-w2j6n_30b70686-bf05-4798-b5dc-3c7dc9cee56a/kube-rbac-proxy/0.log" Jan 21 12:12:42 crc kubenswrapper[4824]: I0121 12:12:42.859324 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-646758c888-zmc4k_df60ca53-0573-4a6a-9235-86e0a1c17f91/nmstate-operator/0.log" Jan 21 12:12:42 crc kubenswrapper[4824]: I0121 12:12:42.869038 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-8474b5b9d8-h6kkh_edf54909-b53e-44de-90bc-355e23bbc767/nmstate-webhook/0.log" Jan 21 12:12:42 crc kubenswrapper[4824]: I0121 12:12:42.949553 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-77c48c7859-dzn7w_57e1fe13-b2c8-4711-9efe-1ebbc7e9d12e/manager/0.log" Jan 21 12:12:42 crc kubenswrapper[4824]: I0121 12:12:42.960860 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-78757b4889-gkwjs_fc911701-7c15-4a20-b89e-709521bdd3a1/manager/0.log" Jan 21 12:12:43 crc kubenswrapper[4824]: I0121 12:12:43.026354 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-767fdc4f47-5mhmd_2e5efa9b-40bc-4860-967d-aa37e1838f81/manager/0.log" Jan 21 12:12:43 crc kubenswrapper[4824]: I0121 12:12:43.036389 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-864f6b75bf-c7vsf_18017a19-1676-4fe9-a97e-ae878f423847/manager/0.log" Jan 21 12:12:43 crc kubenswrapper[4824]: I0121 12:12:43.066486 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-c87fff755-d5hn9_19b1c150-f5ed-4432-bfc2-717b78eebc7e/manager/0.log" Jan 21 12:12:43 crc kubenswrapper[4824]: I0121 12:12:43.116159 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-cb4666565-7w8p4_8c74d3da-d870-45e9-96b9-70eb2395c977/manager/0.log" Jan 21 12:12:43 crc kubenswrapper[4824]: I0121 12:12:43.192998 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-65849867d6-8ngx6_b38ac60b-1643-44d9-a1fe-5f75129247dc/manager/0.log" Jan 21 12:12:43 crc kubenswrapper[4824]: I0121 12:12:43.205177 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-7fc9b76cf6-tncbx_5429da27-a618-44da-8aeb-12b917af8734/manager/0.log" Jan 21 12:12:43 crc kubenswrapper[4824]: I0121 12:12:43.221212 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-6b68b8b854zxcbl_3fa95966-3147-4e50-8ed8-f702a87992c4/manager/0.log" Jan 21 12:12:43 crc kubenswrapper[4824]: I0121 12:12:43.379845 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-init-dbdbd6855-gkm5b_35334861-8e16-4b16-8cf5-167a049836af/operator/0.log" Jan 21 12:12:44 crc kubenswrapper[4824]: I0121 12:12:44.700728 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-74d84f758c-vjgps_c1128e01-8b7c-4742-b1da-dbf06aa8b4d8/manager/0.log" Jan 21 12:12:44 crc kubenswrapper[4824]: I0121 12:12:44.714464 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-zggxm_78eb1906-d9e6-49f7-a512-ed3cc0a07da7/registry-server/0.log" Jan 21 12:12:44 crc kubenswrapper[4824]: I0121 12:12:44.770602 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-55db956ddc-78z2q_f5639f17-3a5b-4ef4-bd6e-6eeb3274636a/manager/0.log" Jan 21 12:12:44 crc kubenswrapper[4824]: I0121 12:12:44.792141 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-686df47fcb-c8xvs_d8b36106-8f21-4873-890d-c1879ea49068/manager/0.log" Jan 21 12:12:44 crc kubenswrapper[4824]: I0121 12:12:44.821878 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-94lqx_782093ab-751c-4e38-bba3-c5f76308c82e/operator/0.log" Jan 21 12:12:44 crc kubenswrapper[4824]: I0121 12:12:44.848202 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-85dd56d4cc-wlrdt_b4cb80e6-c720-4875-9848-7d9597a354e4/manager/0.log" Jan 21 12:12:44 crc kubenswrapper[4824]: I0121 12:12:44.912827 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-5f8f495fcf-6rz5h_db84b172-59d8-4995-815b-50b492283bee/manager/0.log" Jan 21 12:12:44 crc kubenswrapper[4824]: I0121 12:12:44.922344 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-7cd8bc9dbb-xkp8x_52491e7b-cd8c-403a-942c-312debd6f1c6/manager/0.log" Jan 21 12:12:44 crc kubenswrapper[4824]: I0121 12:12:44.933588 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-64cd966744-9w4ch_0715aa91-2ae5-4e07-8388-c37def4c7743/manager/0.log" Jan 21 12:12:46 crc kubenswrapper[4824]: I0121 12:12:46.345904 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-additional-cni-plugins-gjtnh_3a8d2cac-2d0f-4894-9bc0-2abd81699c47/kube-multus-additional-cni-plugins/0.log" Jan 21 12:12:46 crc kubenswrapper[4824]: I0121 12:12:46.353029 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-additional-cni-plugins-gjtnh_3a8d2cac-2d0f-4894-9bc0-2abd81699c47/egress-router-binary-copy/0.log" Jan 21 12:12:46 crc kubenswrapper[4824]: I0121 12:12:46.359544 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-additional-cni-plugins-gjtnh_3a8d2cac-2d0f-4894-9bc0-2abd81699c47/cni-plugins/0.log" Jan 21 12:12:46 crc kubenswrapper[4824]: I0121 12:12:46.365359 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-additional-cni-plugins-gjtnh_3a8d2cac-2d0f-4894-9bc0-2abd81699c47/bond-cni-plugin/0.log" Jan 21 12:12:46 crc kubenswrapper[4824]: I0121 12:12:46.371022 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-additional-cni-plugins-gjtnh_3a8d2cac-2d0f-4894-9bc0-2abd81699c47/routeoverride-cni/0.log" Jan 21 12:12:46 crc kubenswrapper[4824]: I0121 12:12:46.376541 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-additional-cni-plugins-gjtnh_3a8d2cac-2d0f-4894-9bc0-2abd81699c47/whereabouts-cni-bincopy/0.log" Jan 21 12:12:46 crc kubenswrapper[4824]: I0121 12:12:46.381813 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-additional-cni-plugins-gjtnh_3a8d2cac-2d0f-4894-9bc0-2abd81699c47/whereabouts-cni/0.log" Jan 21 12:12:46 crc kubenswrapper[4824]: I0121 12:12:46.412004 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-admission-controller-857f4d67dd-27d2q_7352d717-b2a3-4cd2-9e35-3ae2fe2d7b34/multus-admission-controller/0.log" Jan 21 12:12:46 crc kubenswrapper[4824]: I0121 12:12:46.415708 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-admission-controller-857f4d67dd-27d2q_7352d717-b2a3-4cd2-9e35-3ae2fe2d7b34/kube-rbac-proxy/0.log" Jan 21 12:12:46 crc kubenswrapper[4824]: I0121 12:12:46.456032 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-x94nb_6b9588d1-5876-4c2c-8de5-e5c0ad46f04e/kube-multus/2.log" Jan 21 12:12:46 crc kubenswrapper[4824]: I0121 12:12:46.518411 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-x94nb_6b9588d1-5876-4c2c-8de5-e5c0ad46f04e/kube-multus/3.log" Jan 21 12:12:46 crc kubenswrapper[4824]: I0121 12:12:46.548755 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_network-metrics-daemon-9rmnw_8f9bd12b-22ee-44ed-a8b9-c100d043f691/network-metrics-daemon/0.log" Jan 21 12:12:46 crc kubenswrapper[4824]: I0121 12:12:46.553537 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_network-metrics-daemon-9rmnw_8f9bd12b-22ee-44ed-a8b9-c100d043f691/kube-rbac-proxy/0.log" Jan 21 12:12:53 crc kubenswrapper[4824]: I0121 12:12:53.049347 4824 scope.go:117] "RemoveContainer" containerID="719b0a497b4be25b481cf2d3456b6c515017599e2aeafd1f30d22bffc35f79d4" Jan 21 12:12:53 crc kubenswrapper[4824]: E0121 12:12:53.050046 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 12:13:04 crc kubenswrapper[4824]: I0121 12:13:04.049529 4824 scope.go:117] "RemoveContainer" containerID="719b0a497b4be25b481cf2d3456b6c515017599e2aeafd1f30d22bffc35f79d4" Jan 21 12:13:04 crc kubenswrapper[4824]: E0121 12:13:04.050505 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 12:13:19 crc kubenswrapper[4824]: I0121 12:13:19.051249 4824 scope.go:117] "RemoveContainer" containerID="719b0a497b4be25b481cf2d3456b6c515017599e2aeafd1f30d22bffc35f79d4" Jan 21 12:13:19 crc kubenswrapper[4824]: E0121 12:13:19.052240 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 12:13:32 crc kubenswrapper[4824]: I0121 12:13:32.049826 4824 scope.go:117] "RemoveContainer" containerID="719b0a497b4be25b481cf2d3456b6c515017599e2aeafd1f30d22bffc35f79d4" Jan 21 12:13:32 crc kubenswrapper[4824]: E0121 12:13:32.051092 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 12:13:45 crc kubenswrapper[4824]: I0121 12:13:45.051374 4824 scope.go:117] "RemoveContainer" containerID="719b0a497b4be25b481cf2d3456b6c515017599e2aeafd1f30d22bffc35f79d4" Jan 21 12:13:45 crc kubenswrapper[4824]: E0121 12:13:45.052398 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 12:13:56 crc kubenswrapper[4824]: I0121 12:13:56.049738 4824 scope.go:117] "RemoveContainer" containerID="719b0a497b4be25b481cf2d3456b6c515017599e2aeafd1f30d22bffc35f79d4" Jan 21 12:13:56 crc kubenswrapper[4824]: E0121 12:13:56.050594 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 12:14:08 crc kubenswrapper[4824]: I0121 12:14:08.059986 4824 scope.go:117] "RemoveContainer" containerID="719b0a497b4be25b481cf2d3456b6c515017599e2aeafd1f30d22bffc35f79d4" Jan 21 12:14:08 crc kubenswrapper[4824]: E0121 12:14:08.060898 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 12:14:20 crc kubenswrapper[4824]: I0121 12:14:20.050323 4824 scope.go:117] "RemoveContainer" containerID="719b0a497b4be25b481cf2d3456b6c515017599e2aeafd1f30d22bffc35f79d4" Jan 21 12:14:20 crc kubenswrapper[4824]: E0121 12:14:20.051171 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 12:14:33 crc kubenswrapper[4824]: I0121 12:14:33.050255 4824 scope.go:117] "RemoveContainer" containerID="719b0a497b4be25b481cf2d3456b6c515017599e2aeafd1f30d22bffc35f79d4" Jan 21 12:14:33 crc kubenswrapper[4824]: E0121 12:14:33.051137 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 12:14:33 crc kubenswrapper[4824]: I0121 12:14:33.760157 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-8ctsm"] Jan 21 12:14:33 crc kubenswrapper[4824]: E0121 12:14:33.760866 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f96e30dd-f9f8-4f4f-9dee-27b8275b17c4" containerName="container-00" Jan 21 12:14:33 crc kubenswrapper[4824]: I0121 12:14:33.760978 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="f96e30dd-f9f8-4f4f-9dee-27b8275b17c4" containerName="container-00" Jan 21 12:14:33 crc kubenswrapper[4824]: I0121 12:14:33.761249 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="f96e30dd-f9f8-4f4f-9dee-27b8275b17c4" containerName="container-00" Jan 21 12:14:33 crc kubenswrapper[4824]: I0121 12:14:33.762763 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8ctsm" Jan 21 12:14:33 crc kubenswrapper[4824]: I0121 12:14:33.768191 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-8ctsm"] Jan 21 12:14:33 crc kubenswrapper[4824]: I0121 12:14:33.815845 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jgk5l\" (UniqueName: \"kubernetes.io/projected/38d18f54-9d85-4302-81d2-f678960544f3-kube-api-access-jgk5l\") pod \"certified-operators-8ctsm\" (UID: \"38d18f54-9d85-4302-81d2-f678960544f3\") " pod="openshift-marketplace/certified-operators-8ctsm" Jan 21 12:14:33 crc kubenswrapper[4824]: I0121 12:14:33.815895 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/38d18f54-9d85-4302-81d2-f678960544f3-utilities\") pod \"certified-operators-8ctsm\" (UID: \"38d18f54-9d85-4302-81d2-f678960544f3\") " pod="openshift-marketplace/certified-operators-8ctsm" Jan 21 12:14:33 crc kubenswrapper[4824]: I0121 12:14:33.815945 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/38d18f54-9d85-4302-81d2-f678960544f3-catalog-content\") pod \"certified-operators-8ctsm\" (UID: \"38d18f54-9d85-4302-81d2-f678960544f3\") " pod="openshift-marketplace/certified-operators-8ctsm" Jan 21 12:14:33 crc kubenswrapper[4824]: I0121 12:14:33.918640 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/38d18f54-9d85-4302-81d2-f678960544f3-utilities\") pod \"certified-operators-8ctsm\" (UID: \"38d18f54-9d85-4302-81d2-f678960544f3\") " pod="openshift-marketplace/certified-operators-8ctsm" Jan 21 12:14:33 crc kubenswrapper[4824]: I0121 12:14:33.919177 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/38d18f54-9d85-4302-81d2-f678960544f3-utilities\") pod \"certified-operators-8ctsm\" (UID: \"38d18f54-9d85-4302-81d2-f678960544f3\") " pod="openshift-marketplace/certified-operators-8ctsm" Jan 21 12:14:33 crc kubenswrapper[4824]: I0121 12:14:33.919357 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/38d18f54-9d85-4302-81d2-f678960544f3-catalog-content\") pod \"certified-operators-8ctsm\" (UID: \"38d18f54-9d85-4302-81d2-f678960544f3\") " pod="openshift-marketplace/certified-operators-8ctsm" Jan 21 12:14:33 crc kubenswrapper[4824]: I0121 12:14:33.919653 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/38d18f54-9d85-4302-81d2-f678960544f3-catalog-content\") pod \"certified-operators-8ctsm\" (UID: \"38d18f54-9d85-4302-81d2-f678960544f3\") " pod="openshift-marketplace/certified-operators-8ctsm" Jan 21 12:14:33 crc kubenswrapper[4824]: I0121 12:14:33.919869 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jgk5l\" (UniqueName: \"kubernetes.io/projected/38d18f54-9d85-4302-81d2-f678960544f3-kube-api-access-jgk5l\") pod \"certified-operators-8ctsm\" (UID: \"38d18f54-9d85-4302-81d2-f678960544f3\") " pod="openshift-marketplace/certified-operators-8ctsm" Jan 21 12:14:33 crc kubenswrapper[4824]: I0121 12:14:33.944008 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jgk5l\" (UniqueName: \"kubernetes.io/projected/38d18f54-9d85-4302-81d2-f678960544f3-kube-api-access-jgk5l\") pod \"certified-operators-8ctsm\" (UID: \"38d18f54-9d85-4302-81d2-f678960544f3\") " pod="openshift-marketplace/certified-operators-8ctsm" Jan 21 12:14:34 crc kubenswrapper[4824]: I0121 12:14:34.094488 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8ctsm" Jan 21 12:14:34 crc kubenswrapper[4824]: I0121 12:14:34.524029 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-8ctsm"] Jan 21 12:14:35 crc kubenswrapper[4824]: I0121 12:14:35.170470 4824 generic.go:334] "Generic (PLEG): container finished" podID="38d18f54-9d85-4302-81d2-f678960544f3" containerID="bec54bd98045422889493aa646023d39be4c285605ed55b79ced274c1c282494" exitCode=0 Jan 21 12:14:35 crc kubenswrapper[4824]: I0121 12:14:35.170593 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8ctsm" event={"ID":"38d18f54-9d85-4302-81d2-f678960544f3","Type":"ContainerDied","Data":"bec54bd98045422889493aa646023d39be4c285605ed55b79ced274c1c282494"} Jan 21 12:14:35 crc kubenswrapper[4824]: I0121 12:14:35.170890 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8ctsm" event={"ID":"38d18f54-9d85-4302-81d2-f678960544f3","Type":"ContainerStarted","Data":"85ab2a5e6d6c105b66b6695758ecd564e9e8e6bb0a07a420c6427da89d4eadb4"} Jan 21 12:14:35 crc kubenswrapper[4824]: I0121 12:14:35.172622 4824 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 21 12:14:36 crc kubenswrapper[4824]: I0121 12:14:36.183149 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8ctsm" event={"ID":"38d18f54-9d85-4302-81d2-f678960544f3","Type":"ContainerStarted","Data":"07ffb2ef8629c66b4c1d39d559821a8a3f13f7f33f616a0ff9801ffeb015525c"} Jan 21 12:14:37 crc kubenswrapper[4824]: I0121 12:14:37.193433 4824 generic.go:334] "Generic (PLEG): container finished" podID="38d18f54-9d85-4302-81d2-f678960544f3" containerID="07ffb2ef8629c66b4c1d39d559821a8a3f13f7f33f616a0ff9801ffeb015525c" exitCode=0 Jan 21 12:14:37 crc kubenswrapper[4824]: I0121 12:14:37.193557 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8ctsm" event={"ID":"38d18f54-9d85-4302-81d2-f678960544f3","Type":"ContainerDied","Data":"07ffb2ef8629c66b4c1d39d559821a8a3f13f7f33f616a0ff9801ffeb015525c"} Jan 21 12:14:38 crc kubenswrapper[4824]: I0121 12:14:38.207222 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8ctsm" event={"ID":"38d18f54-9d85-4302-81d2-f678960544f3","Type":"ContainerStarted","Data":"a242a14ddfe52636385e86585559c300a4ea0af91a4b278a3d9f0a7a69f550a2"} Jan 21 12:14:44 crc kubenswrapper[4824]: I0121 12:14:44.094614 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-8ctsm" Jan 21 12:14:44 crc kubenswrapper[4824]: I0121 12:14:44.095441 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-8ctsm" Jan 21 12:14:44 crc kubenswrapper[4824]: I0121 12:14:44.139498 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-8ctsm" Jan 21 12:14:44 crc kubenswrapper[4824]: I0121 12:14:44.160071 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-8ctsm" podStartSLOduration=8.633838012 podStartE2EDuration="11.160034758s" podCreationTimestamp="2026-01-21 12:14:33 +0000 UTC" firstStartedPulling="2026-01-21 12:14:35.172243026 +0000 UTC m=+3817.465272318" lastFinishedPulling="2026-01-21 12:14:37.698439782 +0000 UTC m=+3819.991469064" observedRunningTime="2026-01-21 12:14:38.231483261 +0000 UTC m=+3820.524512553" watchObservedRunningTime="2026-01-21 12:14:44.160034758 +0000 UTC m=+3826.453064050" Jan 21 12:14:44 crc kubenswrapper[4824]: I0121 12:14:44.337822 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-8ctsm" Jan 21 12:14:44 crc kubenswrapper[4824]: I0121 12:14:44.385906 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-8ctsm"] Jan 21 12:14:46 crc kubenswrapper[4824]: I0121 12:14:46.050684 4824 scope.go:117] "RemoveContainer" containerID="719b0a497b4be25b481cf2d3456b6c515017599e2aeafd1f30d22bffc35f79d4" Jan 21 12:14:46 crc kubenswrapper[4824]: E0121 12:14:46.051433 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 12:14:46 crc kubenswrapper[4824]: I0121 12:14:46.326069 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-8ctsm" podUID="38d18f54-9d85-4302-81d2-f678960544f3" containerName="registry-server" containerID="cri-o://a242a14ddfe52636385e86585559c300a4ea0af91a4b278a3d9f0a7a69f550a2" gracePeriod=2 Jan 21 12:14:46 crc kubenswrapper[4824]: I0121 12:14:46.722293 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8ctsm" Jan 21 12:14:46 crc kubenswrapper[4824]: I0121 12:14:46.766684 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/38d18f54-9d85-4302-81d2-f678960544f3-utilities\") pod \"38d18f54-9d85-4302-81d2-f678960544f3\" (UID: \"38d18f54-9d85-4302-81d2-f678960544f3\") " Jan 21 12:14:46 crc kubenswrapper[4824]: I0121 12:14:46.766748 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/38d18f54-9d85-4302-81d2-f678960544f3-catalog-content\") pod \"38d18f54-9d85-4302-81d2-f678960544f3\" (UID: \"38d18f54-9d85-4302-81d2-f678960544f3\") " Jan 21 12:14:46 crc kubenswrapper[4824]: I0121 12:14:46.767410 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/38d18f54-9d85-4302-81d2-f678960544f3-utilities" (OuterVolumeSpecName: "utilities") pod "38d18f54-9d85-4302-81d2-f678960544f3" (UID: "38d18f54-9d85-4302-81d2-f678960544f3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 12:14:46 crc kubenswrapper[4824]: I0121 12:14:46.805462 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/38d18f54-9d85-4302-81d2-f678960544f3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "38d18f54-9d85-4302-81d2-f678960544f3" (UID: "38d18f54-9d85-4302-81d2-f678960544f3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 12:14:46 crc kubenswrapper[4824]: I0121 12:14:46.868232 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jgk5l\" (UniqueName: \"kubernetes.io/projected/38d18f54-9d85-4302-81d2-f678960544f3-kube-api-access-jgk5l\") pod \"38d18f54-9d85-4302-81d2-f678960544f3\" (UID: \"38d18f54-9d85-4302-81d2-f678960544f3\") " Jan 21 12:14:46 crc kubenswrapper[4824]: I0121 12:14:46.869360 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/38d18f54-9d85-4302-81d2-f678960544f3-utilities\") on node \"crc\" DevicePath \"\"" Jan 21 12:14:46 crc kubenswrapper[4824]: I0121 12:14:46.869391 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/38d18f54-9d85-4302-81d2-f678960544f3-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 21 12:14:46 crc kubenswrapper[4824]: I0121 12:14:46.873699 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/38d18f54-9d85-4302-81d2-f678960544f3-kube-api-access-jgk5l" (OuterVolumeSpecName: "kube-api-access-jgk5l") pod "38d18f54-9d85-4302-81d2-f678960544f3" (UID: "38d18f54-9d85-4302-81d2-f678960544f3"). InnerVolumeSpecName "kube-api-access-jgk5l". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 12:14:46 crc kubenswrapper[4824]: I0121 12:14:46.971598 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jgk5l\" (UniqueName: \"kubernetes.io/projected/38d18f54-9d85-4302-81d2-f678960544f3-kube-api-access-jgk5l\") on node \"crc\" DevicePath \"\"" Jan 21 12:14:47 crc kubenswrapper[4824]: I0121 12:14:47.336848 4824 generic.go:334] "Generic (PLEG): container finished" podID="38d18f54-9d85-4302-81d2-f678960544f3" containerID="a242a14ddfe52636385e86585559c300a4ea0af91a4b278a3d9f0a7a69f550a2" exitCode=0 Jan 21 12:14:47 crc kubenswrapper[4824]: I0121 12:14:47.337092 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8ctsm" event={"ID":"38d18f54-9d85-4302-81d2-f678960544f3","Type":"ContainerDied","Data":"a242a14ddfe52636385e86585559c300a4ea0af91a4b278a3d9f0a7a69f550a2"} Jan 21 12:14:47 crc kubenswrapper[4824]: I0121 12:14:47.337856 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8ctsm" event={"ID":"38d18f54-9d85-4302-81d2-f678960544f3","Type":"ContainerDied","Data":"85ab2a5e6d6c105b66b6695758ecd564e9e8e6bb0a07a420c6427da89d4eadb4"} Jan 21 12:14:47 crc kubenswrapper[4824]: I0121 12:14:47.337918 4824 scope.go:117] "RemoveContainer" containerID="a242a14ddfe52636385e86585559c300a4ea0af91a4b278a3d9f0a7a69f550a2" Jan 21 12:14:47 crc kubenswrapper[4824]: I0121 12:14:47.337228 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8ctsm" Jan 21 12:14:47 crc kubenswrapper[4824]: I0121 12:14:47.360207 4824 scope.go:117] "RemoveContainer" containerID="07ffb2ef8629c66b4c1d39d559821a8a3f13f7f33f616a0ff9801ffeb015525c" Jan 21 12:14:47 crc kubenswrapper[4824]: I0121 12:14:47.379892 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-8ctsm"] Jan 21 12:14:47 crc kubenswrapper[4824]: I0121 12:14:47.386377 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-8ctsm"] Jan 21 12:14:47 crc kubenswrapper[4824]: I0121 12:14:47.400838 4824 scope.go:117] "RemoveContainer" containerID="bec54bd98045422889493aa646023d39be4c285605ed55b79ced274c1c282494" Jan 21 12:14:47 crc kubenswrapper[4824]: I0121 12:14:47.422674 4824 scope.go:117] "RemoveContainer" containerID="a242a14ddfe52636385e86585559c300a4ea0af91a4b278a3d9f0a7a69f550a2" Jan 21 12:14:47 crc kubenswrapper[4824]: E0121 12:14:47.423243 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a242a14ddfe52636385e86585559c300a4ea0af91a4b278a3d9f0a7a69f550a2\": container with ID starting with a242a14ddfe52636385e86585559c300a4ea0af91a4b278a3d9f0a7a69f550a2 not found: ID does not exist" containerID="a242a14ddfe52636385e86585559c300a4ea0af91a4b278a3d9f0a7a69f550a2" Jan 21 12:14:47 crc kubenswrapper[4824]: I0121 12:14:47.423279 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a242a14ddfe52636385e86585559c300a4ea0af91a4b278a3d9f0a7a69f550a2"} err="failed to get container status \"a242a14ddfe52636385e86585559c300a4ea0af91a4b278a3d9f0a7a69f550a2\": rpc error: code = NotFound desc = could not find container \"a242a14ddfe52636385e86585559c300a4ea0af91a4b278a3d9f0a7a69f550a2\": container with ID starting with a242a14ddfe52636385e86585559c300a4ea0af91a4b278a3d9f0a7a69f550a2 not found: ID does not exist" Jan 21 12:14:47 crc kubenswrapper[4824]: I0121 12:14:47.423304 4824 scope.go:117] "RemoveContainer" containerID="07ffb2ef8629c66b4c1d39d559821a8a3f13f7f33f616a0ff9801ffeb015525c" Jan 21 12:14:47 crc kubenswrapper[4824]: E0121 12:14:47.423760 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"07ffb2ef8629c66b4c1d39d559821a8a3f13f7f33f616a0ff9801ffeb015525c\": container with ID starting with 07ffb2ef8629c66b4c1d39d559821a8a3f13f7f33f616a0ff9801ffeb015525c not found: ID does not exist" containerID="07ffb2ef8629c66b4c1d39d559821a8a3f13f7f33f616a0ff9801ffeb015525c" Jan 21 12:14:47 crc kubenswrapper[4824]: I0121 12:14:47.423862 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"07ffb2ef8629c66b4c1d39d559821a8a3f13f7f33f616a0ff9801ffeb015525c"} err="failed to get container status \"07ffb2ef8629c66b4c1d39d559821a8a3f13f7f33f616a0ff9801ffeb015525c\": rpc error: code = NotFound desc = could not find container \"07ffb2ef8629c66b4c1d39d559821a8a3f13f7f33f616a0ff9801ffeb015525c\": container with ID starting with 07ffb2ef8629c66b4c1d39d559821a8a3f13f7f33f616a0ff9801ffeb015525c not found: ID does not exist" Jan 21 12:14:47 crc kubenswrapper[4824]: I0121 12:14:47.423950 4824 scope.go:117] "RemoveContainer" containerID="bec54bd98045422889493aa646023d39be4c285605ed55b79ced274c1c282494" Jan 21 12:14:47 crc kubenswrapper[4824]: E0121 12:14:47.424548 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bec54bd98045422889493aa646023d39be4c285605ed55b79ced274c1c282494\": container with ID starting with bec54bd98045422889493aa646023d39be4c285605ed55b79ced274c1c282494 not found: ID does not exist" containerID="bec54bd98045422889493aa646023d39be4c285605ed55b79ced274c1c282494" Jan 21 12:14:47 crc kubenswrapper[4824]: I0121 12:14:47.424581 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bec54bd98045422889493aa646023d39be4c285605ed55b79ced274c1c282494"} err="failed to get container status \"bec54bd98045422889493aa646023d39be4c285605ed55b79ced274c1c282494\": rpc error: code = NotFound desc = could not find container \"bec54bd98045422889493aa646023d39be4c285605ed55b79ced274c1c282494\": container with ID starting with bec54bd98045422889493aa646023d39be4c285605ed55b79ced274c1c282494 not found: ID does not exist" Jan 21 12:14:48 crc kubenswrapper[4824]: I0121 12:14:48.059341 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="38d18f54-9d85-4302-81d2-f678960544f3" path="/var/lib/kubelet/pods/38d18f54-9d85-4302-81d2-f678960544f3/volumes" Jan 21 12:14:58 crc kubenswrapper[4824]: I0121 12:14:58.192012 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-lbnms"] Jan 21 12:14:58 crc kubenswrapper[4824]: E0121 12:14:58.192893 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="38d18f54-9d85-4302-81d2-f678960544f3" containerName="extract-utilities" Jan 21 12:14:58 crc kubenswrapper[4824]: I0121 12:14:58.192908 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="38d18f54-9d85-4302-81d2-f678960544f3" containerName="extract-utilities" Jan 21 12:14:58 crc kubenswrapper[4824]: E0121 12:14:58.192946 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="38d18f54-9d85-4302-81d2-f678960544f3" containerName="extract-content" Jan 21 12:14:58 crc kubenswrapper[4824]: I0121 12:14:58.192974 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="38d18f54-9d85-4302-81d2-f678960544f3" containerName="extract-content" Jan 21 12:14:58 crc kubenswrapper[4824]: E0121 12:14:58.192986 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="38d18f54-9d85-4302-81d2-f678960544f3" containerName="registry-server" Jan 21 12:14:58 crc kubenswrapper[4824]: I0121 12:14:58.192993 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="38d18f54-9d85-4302-81d2-f678960544f3" containerName="registry-server" Jan 21 12:14:58 crc kubenswrapper[4824]: I0121 12:14:58.193207 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="38d18f54-9d85-4302-81d2-f678960544f3" containerName="registry-server" Jan 21 12:14:58 crc kubenswrapper[4824]: I0121 12:14:58.194461 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lbnms" Jan 21 12:14:58 crc kubenswrapper[4824]: I0121 12:14:58.200259 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-lbnms"] Jan 21 12:14:58 crc kubenswrapper[4824]: I0121 12:14:58.370407 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dmcqm\" (UniqueName: \"kubernetes.io/projected/5c25db90-0383-4d60-bc0d-e5fdd37540cd-kube-api-access-dmcqm\") pod \"redhat-operators-lbnms\" (UID: \"5c25db90-0383-4d60-bc0d-e5fdd37540cd\") " pod="openshift-marketplace/redhat-operators-lbnms" Jan 21 12:14:58 crc kubenswrapper[4824]: I0121 12:14:58.370476 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5c25db90-0383-4d60-bc0d-e5fdd37540cd-catalog-content\") pod \"redhat-operators-lbnms\" (UID: \"5c25db90-0383-4d60-bc0d-e5fdd37540cd\") " pod="openshift-marketplace/redhat-operators-lbnms" Jan 21 12:14:58 crc kubenswrapper[4824]: I0121 12:14:58.370526 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5c25db90-0383-4d60-bc0d-e5fdd37540cd-utilities\") pod \"redhat-operators-lbnms\" (UID: \"5c25db90-0383-4d60-bc0d-e5fdd37540cd\") " pod="openshift-marketplace/redhat-operators-lbnms" Jan 21 12:14:58 crc kubenswrapper[4824]: I0121 12:14:58.472926 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dmcqm\" (UniqueName: \"kubernetes.io/projected/5c25db90-0383-4d60-bc0d-e5fdd37540cd-kube-api-access-dmcqm\") pod \"redhat-operators-lbnms\" (UID: \"5c25db90-0383-4d60-bc0d-e5fdd37540cd\") " pod="openshift-marketplace/redhat-operators-lbnms" Jan 21 12:14:58 crc kubenswrapper[4824]: I0121 12:14:58.473009 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5c25db90-0383-4d60-bc0d-e5fdd37540cd-catalog-content\") pod \"redhat-operators-lbnms\" (UID: \"5c25db90-0383-4d60-bc0d-e5fdd37540cd\") " pod="openshift-marketplace/redhat-operators-lbnms" Jan 21 12:14:58 crc kubenswrapper[4824]: I0121 12:14:58.473057 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5c25db90-0383-4d60-bc0d-e5fdd37540cd-utilities\") pod \"redhat-operators-lbnms\" (UID: \"5c25db90-0383-4d60-bc0d-e5fdd37540cd\") " pod="openshift-marketplace/redhat-operators-lbnms" Jan 21 12:14:58 crc kubenswrapper[4824]: I0121 12:14:58.473566 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5c25db90-0383-4d60-bc0d-e5fdd37540cd-utilities\") pod \"redhat-operators-lbnms\" (UID: \"5c25db90-0383-4d60-bc0d-e5fdd37540cd\") " pod="openshift-marketplace/redhat-operators-lbnms" Jan 21 12:14:58 crc kubenswrapper[4824]: I0121 12:14:58.473801 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5c25db90-0383-4d60-bc0d-e5fdd37540cd-catalog-content\") pod \"redhat-operators-lbnms\" (UID: \"5c25db90-0383-4d60-bc0d-e5fdd37540cd\") " pod="openshift-marketplace/redhat-operators-lbnms" Jan 21 12:14:58 crc kubenswrapper[4824]: I0121 12:14:58.492384 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dmcqm\" (UniqueName: \"kubernetes.io/projected/5c25db90-0383-4d60-bc0d-e5fdd37540cd-kube-api-access-dmcqm\") pod \"redhat-operators-lbnms\" (UID: \"5c25db90-0383-4d60-bc0d-e5fdd37540cd\") " pod="openshift-marketplace/redhat-operators-lbnms" Jan 21 12:14:58 crc kubenswrapper[4824]: I0121 12:14:58.522084 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lbnms" Jan 21 12:14:58 crc kubenswrapper[4824]: I0121 12:14:58.986997 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-lbnms"] Jan 21 12:14:59 crc kubenswrapper[4824]: I0121 12:14:59.050893 4824 scope.go:117] "RemoveContainer" containerID="719b0a497b4be25b481cf2d3456b6c515017599e2aeafd1f30d22bffc35f79d4" Jan 21 12:14:59 crc kubenswrapper[4824]: E0121 12:14:59.051235 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 12:14:59 crc kubenswrapper[4824]: I0121 12:14:59.456361 4824 generic.go:334] "Generic (PLEG): container finished" podID="5c25db90-0383-4d60-bc0d-e5fdd37540cd" containerID="fa8c13273e3cbd8fb43a1da76e11231e79bf4dae4568aef9ad7f1840d60b4170" exitCode=0 Jan 21 12:14:59 crc kubenswrapper[4824]: I0121 12:14:59.456426 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lbnms" event={"ID":"5c25db90-0383-4d60-bc0d-e5fdd37540cd","Type":"ContainerDied","Data":"fa8c13273e3cbd8fb43a1da76e11231e79bf4dae4568aef9ad7f1840d60b4170"} Jan 21 12:14:59 crc kubenswrapper[4824]: I0121 12:14:59.456471 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lbnms" event={"ID":"5c25db90-0383-4d60-bc0d-e5fdd37540cd","Type":"ContainerStarted","Data":"ba64d2b0b833265f8588d182cbd463930b1169951658d8b2c78ac3122d41d2dc"} Jan 21 12:15:00 crc kubenswrapper[4824]: I0121 12:15:00.174522 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29483295-m2m69"] Jan 21 12:15:00 crc kubenswrapper[4824]: I0121 12:15:00.176928 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29483295-m2m69" Jan 21 12:15:00 crc kubenswrapper[4824]: I0121 12:15:00.184569 4824 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Jan 21 12:15:00 crc kubenswrapper[4824]: I0121 12:15:00.184713 4824 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Jan 21 12:15:00 crc kubenswrapper[4824]: I0121 12:15:00.184952 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29483295-m2m69"] Jan 21 12:15:00 crc kubenswrapper[4824]: I0121 12:15:00.217502 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/81672912-3e15-467d-9c3c-8a6a3b297d28-secret-volume\") pod \"collect-profiles-29483295-m2m69\" (UID: \"81672912-3e15-467d-9c3c-8a6a3b297d28\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29483295-m2m69" Jan 21 12:15:00 crc kubenswrapper[4824]: I0121 12:15:00.217574 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hv4dm\" (UniqueName: \"kubernetes.io/projected/81672912-3e15-467d-9c3c-8a6a3b297d28-kube-api-access-hv4dm\") pod \"collect-profiles-29483295-m2m69\" (UID: \"81672912-3e15-467d-9c3c-8a6a3b297d28\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29483295-m2m69" Jan 21 12:15:00 crc kubenswrapper[4824]: I0121 12:15:00.217620 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/81672912-3e15-467d-9c3c-8a6a3b297d28-config-volume\") pod \"collect-profiles-29483295-m2m69\" (UID: \"81672912-3e15-467d-9c3c-8a6a3b297d28\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29483295-m2m69" Jan 21 12:15:00 crc kubenswrapper[4824]: I0121 12:15:00.320313 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/81672912-3e15-467d-9c3c-8a6a3b297d28-secret-volume\") pod \"collect-profiles-29483295-m2m69\" (UID: \"81672912-3e15-467d-9c3c-8a6a3b297d28\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29483295-m2m69" Jan 21 12:15:00 crc kubenswrapper[4824]: I0121 12:15:00.320398 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hv4dm\" (UniqueName: \"kubernetes.io/projected/81672912-3e15-467d-9c3c-8a6a3b297d28-kube-api-access-hv4dm\") pod \"collect-profiles-29483295-m2m69\" (UID: \"81672912-3e15-467d-9c3c-8a6a3b297d28\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29483295-m2m69" Jan 21 12:15:00 crc kubenswrapper[4824]: I0121 12:15:00.320431 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/81672912-3e15-467d-9c3c-8a6a3b297d28-config-volume\") pod \"collect-profiles-29483295-m2m69\" (UID: \"81672912-3e15-467d-9c3c-8a6a3b297d28\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29483295-m2m69" Jan 21 12:15:00 crc kubenswrapper[4824]: I0121 12:15:00.321400 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/81672912-3e15-467d-9c3c-8a6a3b297d28-config-volume\") pod \"collect-profiles-29483295-m2m69\" (UID: \"81672912-3e15-467d-9c3c-8a6a3b297d28\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29483295-m2m69" Jan 21 12:15:00 crc kubenswrapper[4824]: I0121 12:15:00.327612 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/81672912-3e15-467d-9c3c-8a6a3b297d28-secret-volume\") pod \"collect-profiles-29483295-m2m69\" (UID: \"81672912-3e15-467d-9c3c-8a6a3b297d28\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29483295-m2m69" Jan 21 12:15:00 crc kubenswrapper[4824]: I0121 12:15:00.334951 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hv4dm\" (UniqueName: \"kubernetes.io/projected/81672912-3e15-467d-9c3c-8a6a3b297d28-kube-api-access-hv4dm\") pod \"collect-profiles-29483295-m2m69\" (UID: \"81672912-3e15-467d-9c3c-8a6a3b297d28\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29483295-m2m69" Jan 21 12:15:00 crc kubenswrapper[4824]: I0121 12:15:00.496622 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29483295-m2m69" Jan 21 12:15:00 crc kubenswrapper[4824]: I0121 12:15:00.970906 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29483295-m2m69"] Jan 21 12:15:01 crc kubenswrapper[4824]: I0121 12:15:01.476475 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lbnms" event={"ID":"5c25db90-0383-4d60-bc0d-e5fdd37540cd","Type":"ContainerStarted","Data":"b676ca21de63116716e6fd6fea1edb453448e2299b189264664e1f1e06956444"} Jan 21 12:15:01 crc kubenswrapper[4824]: I0121 12:15:01.479319 4824 generic.go:334] "Generic (PLEG): container finished" podID="81672912-3e15-467d-9c3c-8a6a3b297d28" containerID="f4ebcbc8adfa426ef5d7d6d30a337211dc204ce7b6a3c0e1a84921e00e5e354f" exitCode=0 Jan 21 12:15:01 crc kubenswrapper[4824]: I0121 12:15:01.479354 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29483295-m2m69" event={"ID":"81672912-3e15-467d-9c3c-8a6a3b297d28","Type":"ContainerDied","Data":"f4ebcbc8adfa426ef5d7d6d30a337211dc204ce7b6a3c0e1a84921e00e5e354f"} Jan 21 12:15:01 crc kubenswrapper[4824]: I0121 12:15:01.479372 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29483295-m2m69" event={"ID":"81672912-3e15-467d-9c3c-8a6a3b297d28","Type":"ContainerStarted","Data":"31b5621ec04585ca9ce7815dbcbeb50db0208149c8f301f56e27d700b8ba7e23"} Jan 21 12:15:02 crc kubenswrapper[4824]: I0121 12:15:02.779120 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29483295-m2m69" Jan 21 12:15:02 crc kubenswrapper[4824]: I0121 12:15:02.973049 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/81672912-3e15-467d-9c3c-8a6a3b297d28-config-volume\") pod \"81672912-3e15-467d-9c3c-8a6a3b297d28\" (UID: \"81672912-3e15-467d-9c3c-8a6a3b297d28\") " Jan 21 12:15:02 crc kubenswrapper[4824]: I0121 12:15:02.973175 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hv4dm\" (UniqueName: \"kubernetes.io/projected/81672912-3e15-467d-9c3c-8a6a3b297d28-kube-api-access-hv4dm\") pod \"81672912-3e15-467d-9c3c-8a6a3b297d28\" (UID: \"81672912-3e15-467d-9c3c-8a6a3b297d28\") " Jan 21 12:15:02 crc kubenswrapper[4824]: I0121 12:15:02.973260 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/81672912-3e15-467d-9c3c-8a6a3b297d28-secret-volume\") pod \"81672912-3e15-467d-9c3c-8a6a3b297d28\" (UID: \"81672912-3e15-467d-9c3c-8a6a3b297d28\") " Jan 21 12:15:02 crc kubenswrapper[4824]: I0121 12:15:02.974087 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/81672912-3e15-467d-9c3c-8a6a3b297d28-config-volume" (OuterVolumeSpecName: "config-volume") pod "81672912-3e15-467d-9c3c-8a6a3b297d28" (UID: "81672912-3e15-467d-9c3c-8a6a3b297d28"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 21 12:15:02 crc kubenswrapper[4824]: I0121 12:15:02.985073 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/81672912-3e15-467d-9c3c-8a6a3b297d28-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "81672912-3e15-467d-9c3c-8a6a3b297d28" (UID: "81672912-3e15-467d-9c3c-8a6a3b297d28"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 21 12:15:02 crc kubenswrapper[4824]: I0121 12:15:02.989840 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/81672912-3e15-467d-9c3c-8a6a3b297d28-kube-api-access-hv4dm" (OuterVolumeSpecName: "kube-api-access-hv4dm") pod "81672912-3e15-467d-9c3c-8a6a3b297d28" (UID: "81672912-3e15-467d-9c3c-8a6a3b297d28"). InnerVolumeSpecName "kube-api-access-hv4dm". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 12:15:03 crc kubenswrapper[4824]: I0121 12:15:03.076176 4824 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/81672912-3e15-467d-9c3c-8a6a3b297d28-config-volume\") on node \"crc\" DevicePath \"\"" Jan 21 12:15:03 crc kubenswrapper[4824]: I0121 12:15:03.076215 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hv4dm\" (UniqueName: \"kubernetes.io/projected/81672912-3e15-467d-9c3c-8a6a3b297d28-kube-api-access-hv4dm\") on node \"crc\" DevicePath \"\"" Jan 21 12:15:03 crc kubenswrapper[4824]: I0121 12:15:03.076229 4824 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/81672912-3e15-467d-9c3c-8a6a3b297d28-secret-volume\") on node \"crc\" DevicePath \"\"" Jan 21 12:15:03 crc kubenswrapper[4824]: I0121 12:15:03.500988 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29483295-m2m69" event={"ID":"81672912-3e15-467d-9c3c-8a6a3b297d28","Type":"ContainerDied","Data":"31b5621ec04585ca9ce7815dbcbeb50db0208149c8f301f56e27d700b8ba7e23"} Jan 21 12:15:03 crc kubenswrapper[4824]: I0121 12:15:03.501043 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29483295-m2m69" Jan 21 12:15:03 crc kubenswrapper[4824]: I0121 12:15:03.501049 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="31b5621ec04585ca9ce7815dbcbeb50db0208149c8f301f56e27d700b8ba7e23" Jan 21 12:15:03 crc kubenswrapper[4824]: I0121 12:15:03.862786 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29483250-mrvrz"] Jan 21 12:15:03 crc kubenswrapper[4824]: I0121 12:15:03.867847 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29483250-mrvrz"] Jan 21 12:15:04 crc kubenswrapper[4824]: I0121 12:15:04.060637 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4c02bf47-f856-44f3-af2c-cc96c45471c1" path="/var/lib/kubelet/pods/4c02bf47-f856-44f3-af2c-cc96c45471c1/volumes" Jan 21 12:15:04 crc kubenswrapper[4824]: I0121 12:15:04.521516 4824 generic.go:334] "Generic (PLEG): container finished" podID="5c25db90-0383-4d60-bc0d-e5fdd37540cd" containerID="b676ca21de63116716e6fd6fea1edb453448e2299b189264664e1f1e06956444" exitCode=0 Jan 21 12:15:04 crc kubenswrapper[4824]: I0121 12:15:04.521568 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lbnms" event={"ID":"5c25db90-0383-4d60-bc0d-e5fdd37540cd","Type":"ContainerDied","Data":"b676ca21de63116716e6fd6fea1edb453448e2299b189264664e1f1e06956444"} Jan 21 12:15:05 crc kubenswrapper[4824]: I0121 12:15:05.532815 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lbnms" event={"ID":"5c25db90-0383-4d60-bc0d-e5fdd37540cd","Type":"ContainerStarted","Data":"c97ddcc101915df9eb9d9296ea49ddd25ce2ec7550fe80e68e52db457f337a25"} Jan 21 12:15:05 crc kubenswrapper[4824]: I0121 12:15:05.570627 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-lbnms" podStartSLOduration=1.997598907 podStartE2EDuration="7.570605504s" podCreationTimestamp="2026-01-21 12:14:58 +0000 UTC" firstStartedPulling="2026-01-21 12:14:59.458580143 +0000 UTC m=+3841.751609435" lastFinishedPulling="2026-01-21 12:15:05.03158674 +0000 UTC m=+3847.324616032" observedRunningTime="2026-01-21 12:15:05.56509791 +0000 UTC m=+3847.858127202" watchObservedRunningTime="2026-01-21 12:15:05.570605504 +0000 UTC m=+3847.863634796" Jan 21 12:15:08 crc kubenswrapper[4824]: I0121 12:15:08.523804 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-lbnms" Jan 21 12:15:08 crc kubenswrapper[4824]: I0121 12:15:08.524556 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-lbnms" Jan 21 12:15:09 crc kubenswrapper[4824]: I0121 12:15:09.563451 4824 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-lbnms" podUID="5c25db90-0383-4d60-bc0d-e5fdd37540cd" containerName="registry-server" probeResult="failure" output=< Jan 21 12:15:09 crc kubenswrapper[4824]: timeout: failed to connect service ":50051" within 1s Jan 21 12:15:09 crc kubenswrapper[4824]: > Jan 21 12:15:10 crc kubenswrapper[4824]: I0121 12:15:10.050544 4824 scope.go:117] "RemoveContainer" containerID="719b0a497b4be25b481cf2d3456b6c515017599e2aeafd1f30d22bffc35f79d4" Jan 21 12:15:10 crc kubenswrapper[4824]: E0121 12:15:10.052047 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 12:15:18 crc kubenswrapper[4824]: I0121 12:15:18.570099 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-lbnms" Jan 21 12:15:18 crc kubenswrapper[4824]: I0121 12:15:18.624662 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-lbnms" Jan 21 12:15:18 crc kubenswrapper[4824]: I0121 12:15:18.806820 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-lbnms"] Jan 21 12:15:19 crc kubenswrapper[4824]: I0121 12:15:19.679257 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-lbnms" podUID="5c25db90-0383-4d60-bc0d-e5fdd37540cd" containerName="registry-server" containerID="cri-o://c97ddcc101915df9eb9d9296ea49ddd25ce2ec7550fe80e68e52db457f337a25" gracePeriod=2 Jan 21 12:15:20 crc kubenswrapper[4824]: I0121 12:15:20.128177 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lbnms" Jan 21 12:15:20 crc kubenswrapper[4824]: I0121 12:15:20.187197 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5c25db90-0383-4d60-bc0d-e5fdd37540cd-utilities\") pod \"5c25db90-0383-4d60-bc0d-e5fdd37540cd\" (UID: \"5c25db90-0383-4d60-bc0d-e5fdd37540cd\") " Jan 21 12:15:20 crc kubenswrapper[4824]: I0121 12:15:20.187306 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dmcqm\" (UniqueName: \"kubernetes.io/projected/5c25db90-0383-4d60-bc0d-e5fdd37540cd-kube-api-access-dmcqm\") pod \"5c25db90-0383-4d60-bc0d-e5fdd37540cd\" (UID: \"5c25db90-0383-4d60-bc0d-e5fdd37540cd\") " Jan 21 12:15:20 crc kubenswrapper[4824]: I0121 12:15:20.187489 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5c25db90-0383-4d60-bc0d-e5fdd37540cd-catalog-content\") pod \"5c25db90-0383-4d60-bc0d-e5fdd37540cd\" (UID: \"5c25db90-0383-4d60-bc0d-e5fdd37540cd\") " Jan 21 12:15:20 crc kubenswrapper[4824]: I0121 12:15:20.188727 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5c25db90-0383-4d60-bc0d-e5fdd37540cd-utilities" (OuterVolumeSpecName: "utilities") pod "5c25db90-0383-4d60-bc0d-e5fdd37540cd" (UID: "5c25db90-0383-4d60-bc0d-e5fdd37540cd"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 12:15:20 crc kubenswrapper[4824]: I0121 12:15:20.195881 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5c25db90-0383-4d60-bc0d-e5fdd37540cd-kube-api-access-dmcqm" (OuterVolumeSpecName: "kube-api-access-dmcqm") pod "5c25db90-0383-4d60-bc0d-e5fdd37540cd" (UID: "5c25db90-0383-4d60-bc0d-e5fdd37540cd"). InnerVolumeSpecName "kube-api-access-dmcqm". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 12:15:20 crc kubenswrapper[4824]: I0121 12:15:20.289153 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5c25db90-0383-4d60-bc0d-e5fdd37540cd-utilities\") on node \"crc\" DevicePath \"\"" Jan 21 12:15:20 crc kubenswrapper[4824]: I0121 12:15:20.289184 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dmcqm\" (UniqueName: \"kubernetes.io/projected/5c25db90-0383-4d60-bc0d-e5fdd37540cd-kube-api-access-dmcqm\") on node \"crc\" DevicePath \"\"" Jan 21 12:15:20 crc kubenswrapper[4824]: I0121 12:15:20.289762 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5c25db90-0383-4d60-bc0d-e5fdd37540cd-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5c25db90-0383-4d60-bc0d-e5fdd37540cd" (UID: "5c25db90-0383-4d60-bc0d-e5fdd37540cd"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 12:15:20 crc kubenswrapper[4824]: I0121 12:15:20.390419 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5c25db90-0383-4d60-bc0d-e5fdd37540cd-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 21 12:15:20 crc kubenswrapper[4824]: I0121 12:15:20.690397 4824 generic.go:334] "Generic (PLEG): container finished" podID="5c25db90-0383-4d60-bc0d-e5fdd37540cd" containerID="c97ddcc101915df9eb9d9296ea49ddd25ce2ec7550fe80e68e52db457f337a25" exitCode=0 Jan 21 12:15:20 crc kubenswrapper[4824]: I0121 12:15:20.690450 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lbnms" event={"ID":"5c25db90-0383-4d60-bc0d-e5fdd37540cd","Type":"ContainerDied","Data":"c97ddcc101915df9eb9d9296ea49ddd25ce2ec7550fe80e68e52db457f337a25"} Jan 21 12:15:20 crc kubenswrapper[4824]: I0121 12:15:20.690497 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lbnms" event={"ID":"5c25db90-0383-4d60-bc0d-e5fdd37540cd","Type":"ContainerDied","Data":"ba64d2b0b833265f8588d182cbd463930b1169951658d8b2c78ac3122d41d2dc"} Jan 21 12:15:20 crc kubenswrapper[4824]: I0121 12:15:20.690520 4824 scope.go:117] "RemoveContainer" containerID="c97ddcc101915df9eb9d9296ea49ddd25ce2ec7550fe80e68e52db457f337a25" Jan 21 12:15:20 crc kubenswrapper[4824]: I0121 12:15:20.690687 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lbnms" Jan 21 12:15:20 crc kubenswrapper[4824]: I0121 12:15:20.715127 4824 scope.go:117] "RemoveContainer" containerID="b676ca21de63116716e6fd6fea1edb453448e2299b189264664e1f1e06956444" Jan 21 12:15:20 crc kubenswrapper[4824]: I0121 12:15:20.730603 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-lbnms"] Jan 21 12:15:20 crc kubenswrapper[4824]: I0121 12:15:20.739010 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-lbnms"] Jan 21 12:15:20 crc kubenswrapper[4824]: I0121 12:15:20.752852 4824 scope.go:117] "RemoveContainer" containerID="fa8c13273e3cbd8fb43a1da76e11231e79bf4dae4568aef9ad7f1840d60b4170" Jan 21 12:15:20 crc kubenswrapper[4824]: I0121 12:15:20.775291 4824 scope.go:117] "RemoveContainer" containerID="c97ddcc101915df9eb9d9296ea49ddd25ce2ec7550fe80e68e52db457f337a25" Jan 21 12:15:20 crc kubenswrapper[4824]: E0121 12:15:20.775663 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c97ddcc101915df9eb9d9296ea49ddd25ce2ec7550fe80e68e52db457f337a25\": container with ID starting with c97ddcc101915df9eb9d9296ea49ddd25ce2ec7550fe80e68e52db457f337a25 not found: ID does not exist" containerID="c97ddcc101915df9eb9d9296ea49ddd25ce2ec7550fe80e68e52db457f337a25" Jan 21 12:15:20 crc kubenswrapper[4824]: I0121 12:15:20.775699 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c97ddcc101915df9eb9d9296ea49ddd25ce2ec7550fe80e68e52db457f337a25"} err="failed to get container status \"c97ddcc101915df9eb9d9296ea49ddd25ce2ec7550fe80e68e52db457f337a25\": rpc error: code = NotFound desc = could not find container \"c97ddcc101915df9eb9d9296ea49ddd25ce2ec7550fe80e68e52db457f337a25\": container with ID starting with c97ddcc101915df9eb9d9296ea49ddd25ce2ec7550fe80e68e52db457f337a25 not found: ID does not exist" Jan 21 12:15:20 crc kubenswrapper[4824]: I0121 12:15:20.775725 4824 scope.go:117] "RemoveContainer" containerID="b676ca21de63116716e6fd6fea1edb453448e2299b189264664e1f1e06956444" Jan 21 12:15:20 crc kubenswrapper[4824]: E0121 12:15:20.776011 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b676ca21de63116716e6fd6fea1edb453448e2299b189264664e1f1e06956444\": container with ID starting with b676ca21de63116716e6fd6fea1edb453448e2299b189264664e1f1e06956444 not found: ID does not exist" containerID="b676ca21de63116716e6fd6fea1edb453448e2299b189264664e1f1e06956444" Jan 21 12:15:20 crc kubenswrapper[4824]: I0121 12:15:20.776037 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b676ca21de63116716e6fd6fea1edb453448e2299b189264664e1f1e06956444"} err="failed to get container status \"b676ca21de63116716e6fd6fea1edb453448e2299b189264664e1f1e06956444\": rpc error: code = NotFound desc = could not find container \"b676ca21de63116716e6fd6fea1edb453448e2299b189264664e1f1e06956444\": container with ID starting with b676ca21de63116716e6fd6fea1edb453448e2299b189264664e1f1e06956444 not found: ID does not exist" Jan 21 12:15:20 crc kubenswrapper[4824]: I0121 12:15:20.776053 4824 scope.go:117] "RemoveContainer" containerID="fa8c13273e3cbd8fb43a1da76e11231e79bf4dae4568aef9ad7f1840d60b4170" Jan 21 12:15:20 crc kubenswrapper[4824]: E0121 12:15:20.776348 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fa8c13273e3cbd8fb43a1da76e11231e79bf4dae4568aef9ad7f1840d60b4170\": container with ID starting with fa8c13273e3cbd8fb43a1da76e11231e79bf4dae4568aef9ad7f1840d60b4170 not found: ID does not exist" containerID="fa8c13273e3cbd8fb43a1da76e11231e79bf4dae4568aef9ad7f1840d60b4170" Jan 21 12:15:20 crc kubenswrapper[4824]: I0121 12:15:20.776369 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fa8c13273e3cbd8fb43a1da76e11231e79bf4dae4568aef9ad7f1840d60b4170"} err="failed to get container status \"fa8c13273e3cbd8fb43a1da76e11231e79bf4dae4568aef9ad7f1840d60b4170\": rpc error: code = NotFound desc = could not find container \"fa8c13273e3cbd8fb43a1da76e11231e79bf4dae4568aef9ad7f1840d60b4170\": container with ID starting with fa8c13273e3cbd8fb43a1da76e11231e79bf4dae4568aef9ad7f1840d60b4170 not found: ID does not exist" Jan 21 12:15:22 crc kubenswrapper[4824]: I0121 12:15:22.059477 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5c25db90-0383-4d60-bc0d-e5fdd37540cd" path="/var/lib/kubelet/pods/5c25db90-0383-4d60-bc0d-e5fdd37540cd/volumes" Jan 21 12:15:24 crc kubenswrapper[4824]: I0121 12:15:24.049166 4824 scope.go:117] "RemoveContainer" containerID="719b0a497b4be25b481cf2d3456b6c515017599e2aeafd1f30d22bffc35f79d4" Jan 21 12:15:24 crc kubenswrapper[4824]: E0121 12:15:24.049887 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 12:15:38 crc kubenswrapper[4824]: I0121 12:15:38.056095 4824 scope.go:117] "RemoveContainer" containerID="719b0a497b4be25b481cf2d3456b6c515017599e2aeafd1f30d22bffc35f79d4" Jan 21 12:15:38 crc kubenswrapper[4824]: E0121 12:15:38.057184 4824 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zqjg9_openshift-machine-config-operator(33f3d922-4ffe-409b-a49a-d88c85898260)\"" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" Jan 21 12:15:53 crc kubenswrapper[4824]: I0121 12:15:53.050054 4824 scope.go:117] "RemoveContainer" containerID="719b0a497b4be25b481cf2d3456b6c515017599e2aeafd1f30d22bffc35f79d4" Jan 21 12:15:53 crc kubenswrapper[4824]: I0121 12:15:53.993229 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" event={"ID":"33f3d922-4ffe-409b-a49a-d88c85898260","Type":"ContainerStarted","Data":"d9d0949df611c52cc67603e5984be93524636cdbd8451c97adc3963dfdbe83d5"} Jan 21 12:15:55 crc kubenswrapper[4824]: I0121 12:15:55.970994 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/swift-proxy-79f4c948b7-4hz4p" podUID="01d4ce4d-3728-4f4c-9fbf-ce58793b6b7d" containerName="proxy-server" probeResult="failure" output="HTTP probe failed with statuscode: 502" Jan 21 12:16:01 crc kubenswrapper[4824]: I0121 12:16:01.696948 4824 scope.go:117] "RemoveContainer" containerID="dd264d8d4b02d26df3192da84e03b87f54958338166fb00a9d9562f0fbc98220" Jan 21 12:17:01 crc kubenswrapper[4824]: I0121 12:17:01.779861 4824 scope.go:117] "RemoveContainer" containerID="2719d0b32f33e24c79ae317ca212bc0f5429a69ed705aafd8970f2d93f3e60b9" Jan 21 12:18:01 crc kubenswrapper[4824]: I0121 12:18:01.841664 4824 scope.go:117] "RemoveContainer" containerID="07bf87c3a875d48665a8353210d054b6f5a80d43f1b2712cff93bd33cf233d5b" Jan 21 12:18:15 crc kubenswrapper[4824]: I0121 12:18:15.749356 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-prbq4"] Jan 21 12:18:15 crc kubenswrapper[4824]: E0121 12:18:15.750256 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c25db90-0383-4d60-bc0d-e5fdd37540cd" containerName="extract-content" Jan 21 12:18:15 crc kubenswrapper[4824]: I0121 12:18:15.750273 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c25db90-0383-4d60-bc0d-e5fdd37540cd" containerName="extract-content" Jan 21 12:18:15 crc kubenswrapper[4824]: E0121 12:18:15.750291 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c25db90-0383-4d60-bc0d-e5fdd37540cd" containerName="registry-server" Jan 21 12:18:15 crc kubenswrapper[4824]: I0121 12:18:15.750297 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c25db90-0383-4d60-bc0d-e5fdd37540cd" containerName="registry-server" Jan 21 12:18:15 crc kubenswrapper[4824]: E0121 12:18:15.750308 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="81672912-3e15-467d-9c3c-8a6a3b297d28" containerName="collect-profiles" Jan 21 12:18:15 crc kubenswrapper[4824]: I0121 12:18:15.750315 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="81672912-3e15-467d-9c3c-8a6a3b297d28" containerName="collect-profiles" Jan 21 12:18:15 crc kubenswrapper[4824]: E0121 12:18:15.750345 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c25db90-0383-4d60-bc0d-e5fdd37540cd" containerName="extract-utilities" Jan 21 12:18:15 crc kubenswrapper[4824]: I0121 12:18:15.750351 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c25db90-0383-4d60-bc0d-e5fdd37540cd" containerName="extract-utilities" Jan 21 12:18:15 crc kubenswrapper[4824]: I0121 12:18:15.750569 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="81672912-3e15-467d-9c3c-8a6a3b297d28" containerName="collect-profiles" Jan 21 12:18:15 crc kubenswrapper[4824]: I0121 12:18:15.750584 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="5c25db90-0383-4d60-bc0d-e5fdd37540cd" containerName="registry-server" Jan 21 12:18:15 crc kubenswrapper[4824]: I0121 12:18:15.751851 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-prbq4" Jan 21 12:18:15 crc kubenswrapper[4824]: I0121 12:18:15.757247 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-prbq4"] Jan 21 12:18:15 crc kubenswrapper[4824]: I0121 12:18:15.922177 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d7d81231-dda8-4cc9-b7a3-26e26fd4c3b1-utilities\") pod \"community-operators-prbq4\" (UID: \"d7d81231-dda8-4cc9-b7a3-26e26fd4c3b1\") " pod="openshift-marketplace/community-operators-prbq4" Jan 21 12:18:15 crc kubenswrapper[4824]: I0121 12:18:15.922324 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d7d81231-dda8-4cc9-b7a3-26e26fd4c3b1-catalog-content\") pod \"community-operators-prbq4\" (UID: \"d7d81231-dda8-4cc9-b7a3-26e26fd4c3b1\") " pod="openshift-marketplace/community-operators-prbq4" Jan 21 12:18:15 crc kubenswrapper[4824]: I0121 12:18:15.922893 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7nmcb\" (UniqueName: \"kubernetes.io/projected/d7d81231-dda8-4cc9-b7a3-26e26fd4c3b1-kube-api-access-7nmcb\") pod \"community-operators-prbq4\" (UID: \"d7d81231-dda8-4cc9-b7a3-26e26fd4c3b1\") " pod="openshift-marketplace/community-operators-prbq4" Jan 21 12:18:16 crc kubenswrapper[4824]: I0121 12:18:16.025712 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d7d81231-dda8-4cc9-b7a3-26e26fd4c3b1-utilities\") pod \"community-operators-prbq4\" (UID: \"d7d81231-dda8-4cc9-b7a3-26e26fd4c3b1\") " pod="openshift-marketplace/community-operators-prbq4" Jan 21 12:18:16 crc kubenswrapper[4824]: I0121 12:18:16.025775 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d7d81231-dda8-4cc9-b7a3-26e26fd4c3b1-catalog-content\") pod \"community-operators-prbq4\" (UID: \"d7d81231-dda8-4cc9-b7a3-26e26fd4c3b1\") " pod="openshift-marketplace/community-operators-prbq4" Jan 21 12:18:16 crc kubenswrapper[4824]: I0121 12:18:16.025858 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7nmcb\" (UniqueName: \"kubernetes.io/projected/d7d81231-dda8-4cc9-b7a3-26e26fd4c3b1-kube-api-access-7nmcb\") pod \"community-operators-prbq4\" (UID: \"d7d81231-dda8-4cc9-b7a3-26e26fd4c3b1\") " pod="openshift-marketplace/community-operators-prbq4" Jan 21 12:18:16 crc kubenswrapper[4824]: I0121 12:18:16.026193 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d7d81231-dda8-4cc9-b7a3-26e26fd4c3b1-utilities\") pod \"community-operators-prbq4\" (UID: \"d7d81231-dda8-4cc9-b7a3-26e26fd4c3b1\") " pod="openshift-marketplace/community-operators-prbq4" Jan 21 12:18:16 crc kubenswrapper[4824]: I0121 12:18:16.026420 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d7d81231-dda8-4cc9-b7a3-26e26fd4c3b1-catalog-content\") pod \"community-operators-prbq4\" (UID: \"d7d81231-dda8-4cc9-b7a3-26e26fd4c3b1\") " pod="openshift-marketplace/community-operators-prbq4" Jan 21 12:18:16 crc kubenswrapper[4824]: I0121 12:18:16.049239 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7nmcb\" (UniqueName: \"kubernetes.io/projected/d7d81231-dda8-4cc9-b7a3-26e26fd4c3b1-kube-api-access-7nmcb\") pod \"community-operators-prbq4\" (UID: \"d7d81231-dda8-4cc9-b7a3-26e26fd4c3b1\") " pod="openshift-marketplace/community-operators-prbq4" Jan 21 12:18:16 crc kubenswrapper[4824]: I0121 12:18:16.065747 4824 patch_prober.go:28] interesting pod/machine-config-daemon-zqjg9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 21 12:18:16 crc kubenswrapper[4824]: I0121 12:18:16.065805 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 21 12:18:16 crc kubenswrapper[4824]: I0121 12:18:16.075841 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-prbq4" Jan 21 12:18:16 crc kubenswrapper[4824]: I0121 12:18:16.552257 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-prbq4"] Jan 21 12:18:17 crc kubenswrapper[4824]: I0121 12:18:17.335772 4824 generic.go:334] "Generic (PLEG): container finished" podID="d7d81231-dda8-4cc9-b7a3-26e26fd4c3b1" containerID="1d151a2c4798cdf4c00e18f1fd6ec9e76400d5de8c6c07713b44c9afa46983b5" exitCode=0 Jan 21 12:18:17 crc kubenswrapper[4824]: I0121 12:18:17.335885 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-prbq4" event={"ID":"d7d81231-dda8-4cc9-b7a3-26e26fd4c3b1","Type":"ContainerDied","Data":"1d151a2c4798cdf4c00e18f1fd6ec9e76400d5de8c6c07713b44c9afa46983b5"} Jan 21 12:18:17 crc kubenswrapper[4824]: I0121 12:18:17.336117 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-prbq4" event={"ID":"d7d81231-dda8-4cc9-b7a3-26e26fd4c3b1","Type":"ContainerStarted","Data":"b9354b8fbffbcb7440fe899058d36ff7c2aea9e5cfda3cc98ecfdcc6a57082cf"} Jan 21 12:18:18 crc kubenswrapper[4824]: I0121 12:18:18.348736 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-prbq4" event={"ID":"d7d81231-dda8-4cc9-b7a3-26e26fd4c3b1","Type":"ContainerStarted","Data":"3310bd7f7ea302d3f0ffa6be853c1cff8bd444d8c9bd56219e474af7edb4e819"} Jan 21 12:18:19 crc kubenswrapper[4824]: I0121 12:18:19.358562 4824 generic.go:334] "Generic (PLEG): container finished" podID="d7d81231-dda8-4cc9-b7a3-26e26fd4c3b1" containerID="3310bd7f7ea302d3f0ffa6be853c1cff8bd444d8c9bd56219e474af7edb4e819" exitCode=0 Jan 21 12:18:19 crc kubenswrapper[4824]: I0121 12:18:19.358917 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-prbq4" event={"ID":"d7d81231-dda8-4cc9-b7a3-26e26fd4c3b1","Type":"ContainerDied","Data":"3310bd7f7ea302d3f0ffa6be853c1cff8bd444d8c9bd56219e474af7edb4e819"} Jan 21 12:18:20 crc kubenswrapper[4824]: I0121 12:18:20.369827 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-prbq4" event={"ID":"d7d81231-dda8-4cc9-b7a3-26e26fd4c3b1","Type":"ContainerStarted","Data":"b37424659149b4e06e34d4081bac6d51440c828aa5e5da63656c7fff13943c40"} Jan 21 12:18:20 crc kubenswrapper[4824]: I0121 12:18:20.386129 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-prbq4" podStartSLOduration=2.7814604149999997 podStartE2EDuration="5.386115407s" podCreationTimestamp="2026-01-21 12:18:15 +0000 UTC" firstStartedPulling="2026-01-21 12:18:17.337829693 +0000 UTC m=+4039.630858975" lastFinishedPulling="2026-01-21 12:18:19.942484675 +0000 UTC m=+4042.235513967" observedRunningTime="2026-01-21 12:18:20.383575799 +0000 UTC m=+4042.676605092" watchObservedRunningTime="2026-01-21 12:18:20.386115407 +0000 UTC m=+4042.679144699" Jan 21 12:18:26 crc kubenswrapper[4824]: I0121 12:18:26.076866 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-prbq4" Jan 21 12:18:26 crc kubenswrapper[4824]: I0121 12:18:26.077581 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-prbq4" Jan 21 12:18:26 crc kubenswrapper[4824]: I0121 12:18:26.121754 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-prbq4" Jan 21 12:18:26 crc kubenswrapper[4824]: I0121 12:18:26.458092 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-prbq4" Jan 21 12:18:26 crc kubenswrapper[4824]: I0121 12:18:26.496563 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-prbq4"] Jan 21 12:18:28 crc kubenswrapper[4824]: I0121 12:18:28.434175 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-prbq4" podUID="d7d81231-dda8-4cc9-b7a3-26e26fd4c3b1" containerName="registry-server" containerID="cri-o://b37424659149b4e06e34d4081bac6d51440c828aa5e5da63656c7fff13943c40" gracePeriod=2 Jan 21 12:18:29 crc kubenswrapper[4824]: I0121 12:18:29.355811 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-prbq4" Jan 21 12:18:29 crc kubenswrapper[4824]: I0121 12:18:29.444890 4824 generic.go:334] "Generic (PLEG): container finished" podID="d7d81231-dda8-4cc9-b7a3-26e26fd4c3b1" containerID="b37424659149b4e06e34d4081bac6d51440c828aa5e5da63656c7fff13943c40" exitCode=0 Jan 21 12:18:29 crc kubenswrapper[4824]: I0121 12:18:29.444943 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-prbq4" event={"ID":"d7d81231-dda8-4cc9-b7a3-26e26fd4c3b1","Type":"ContainerDied","Data":"b37424659149b4e06e34d4081bac6d51440c828aa5e5da63656c7fff13943c40"} Jan 21 12:18:29 crc kubenswrapper[4824]: I0121 12:18:29.444989 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-prbq4" event={"ID":"d7d81231-dda8-4cc9-b7a3-26e26fd4c3b1","Type":"ContainerDied","Data":"b9354b8fbffbcb7440fe899058d36ff7c2aea9e5cfda3cc98ecfdcc6a57082cf"} Jan 21 12:18:29 crc kubenswrapper[4824]: I0121 12:18:29.445010 4824 scope.go:117] "RemoveContainer" containerID="b37424659149b4e06e34d4081bac6d51440c828aa5e5da63656c7fff13943c40" Jan 21 12:18:29 crc kubenswrapper[4824]: I0121 12:18:29.445025 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-prbq4" Jan 21 12:18:29 crc kubenswrapper[4824]: I0121 12:18:29.475539 4824 scope.go:117] "RemoveContainer" containerID="3310bd7f7ea302d3f0ffa6be853c1cff8bd444d8c9bd56219e474af7edb4e819" Jan 21 12:18:29 crc kubenswrapper[4824]: I0121 12:18:29.500739 4824 scope.go:117] "RemoveContainer" containerID="1d151a2c4798cdf4c00e18f1fd6ec9e76400d5de8c6c07713b44c9afa46983b5" Jan 21 12:18:29 crc kubenswrapper[4824]: I0121 12:18:29.510472 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d7d81231-dda8-4cc9-b7a3-26e26fd4c3b1-catalog-content\") pod \"d7d81231-dda8-4cc9-b7a3-26e26fd4c3b1\" (UID: \"d7d81231-dda8-4cc9-b7a3-26e26fd4c3b1\") " Jan 21 12:18:29 crc kubenswrapper[4824]: I0121 12:18:29.510777 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7nmcb\" (UniqueName: \"kubernetes.io/projected/d7d81231-dda8-4cc9-b7a3-26e26fd4c3b1-kube-api-access-7nmcb\") pod \"d7d81231-dda8-4cc9-b7a3-26e26fd4c3b1\" (UID: \"d7d81231-dda8-4cc9-b7a3-26e26fd4c3b1\") " Jan 21 12:18:29 crc kubenswrapper[4824]: I0121 12:18:29.510989 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d7d81231-dda8-4cc9-b7a3-26e26fd4c3b1-utilities\") pod \"d7d81231-dda8-4cc9-b7a3-26e26fd4c3b1\" (UID: \"d7d81231-dda8-4cc9-b7a3-26e26fd4c3b1\") " Jan 21 12:18:29 crc kubenswrapper[4824]: I0121 12:18:29.511598 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d7d81231-dda8-4cc9-b7a3-26e26fd4c3b1-utilities" (OuterVolumeSpecName: "utilities") pod "d7d81231-dda8-4cc9-b7a3-26e26fd4c3b1" (UID: "d7d81231-dda8-4cc9-b7a3-26e26fd4c3b1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 12:18:29 crc kubenswrapper[4824]: I0121 12:18:29.511786 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d7d81231-dda8-4cc9-b7a3-26e26fd4c3b1-utilities\") on node \"crc\" DevicePath \"\"" Jan 21 12:18:29 crc kubenswrapper[4824]: I0121 12:18:29.519236 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d7d81231-dda8-4cc9-b7a3-26e26fd4c3b1-kube-api-access-7nmcb" (OuterVolumeSpecName: "kube-api-access-7nmcb") pod "d7d81231-dda8-4cc9-b7a3-26e26fd4c3b1" (UID: "d7d81231-dda8-4cc9-b7a3-26e26fd4c3b1"). InnerVolumeSpecName "kube-api-access-7nmcb". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 12:18:29 crc kubenswrapper[4824]: I0121 12:18:29.556858 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d7d81231-dda8-4cc9-b7a3-26e26fd4c3b1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d7d81231-dda8-4cc9-b7a3-26e26fd4c3b1" (UID: "d7d81231-dda8-4cc9-b7a3-26e26fd4c3b1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 12:18:29 crc kubenswrapper[4824]: I0121 12:18:29.568090 4824 scope.go:117] "RemoveContainer" containerID="b37424659149b4e06e34d4081bac6d51440c828aa5e5da63656c7fff13943c40" Jan 21 12:18:29 crc kubenswrapper[4824]: E0121 12:18:29.568511 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b37424659149b4e06e34d4081bac6d51440c828aa5e5da63656c7fff13943c40\": container with ID starting with b37424659149b4e06e34d4081bac6d51440c828aa5e5da63656c7fff13943c40 not found: ID does not exist" containerID="b37424659149b4e06e34d4081bac6d51440c828aa5e5da63656c7fff13943c40" Jan 21 12:18:29 crc kubenswrapper[4824]: I0121 12:18:29.568550 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b37424659149b4e06e34d4081bac6d51440c828aa5e5da63656c7fff13943c40"} err="failed to get container status \"b37424659149b4e06e34d4081bac6d51440c828aa5e5da63656c7fff13943c40\": rpc error: code = NotFound desc = could not find container \"b37424659149b4e06e34d4081bac6d51440c828aa5e5da63656c7fff13943c40\": container with ID starting with b37424659149b4e06e34d4081bac6d51440c828aa5e5da63656c7fff13943c40 not found: ID does not exist" Jan 21 12:18:29 crc kubenswrapper[4824]: I0121 12:18:29.568574 4824 scope.go:117] "RemoveContainer" containerID="3310bd7f7ea302d3f0ffa6be853c1cff8bd444d8c9bd56219e474af7edb4e819" Jan 21 12:18:29 crc kubenswrapper[4824]: E0121 12:18:29.568945 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3310bd7f7ea302d3f0ffa6be853c1cff8bd444d8c9bd56219e474af7edb4e819\": container with ID starting with 3310bd7f7ea302d3f0ffa6be853c1cff8bd444d8c9bd56219e474af7edb4e819 not found: ID does not exist" containerID="3310bd7f7ea302d3f0ffa6be853c1cff8bd444d8c9bd56219e474af7edb4e819" Jan 21 12:18:29 crc kubenswrapper[4824]: I0121 12:18:29.569009 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3310bd7f7ea302d3f0ffa6be853c1cff8bd444d8c9bd56219e474af7edb4e819"} err="failed to get container status \"3310bd7f7ea302d3f0ffa6be853c1cff8bd444d8c9bd56219e474af7edb4e819\": rpc error: code = NotFound desc = could not find container \"3310bd7f7ea302d3f0ffa6be853c1cff8bd444d8c9bd56219e474af7edb4e819\": container with ID starting with 3310bd7f7ea302d3f0ffa6be853c1cff8bd444d8c9bd56219e474af7edb4e819 not found: ID does not exist" Jan 21 12:18:29 crc kubenswrapper[4824]: I0121 12:18:29.569048 4824 scope.go:117] "RemoveContainer" containerID="1d151a2c4798cdf4c00e18f1fd6ec9e76400d5de8c6c07713b44c9afa46983b5" Jan 21 12:18:29 crc kubenswrapper[4824]: E0121 12:18:29.569358 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1d151a2c4798cdf4c00e18f1fd6ec9e76400d5de8c6c07713b44c9afa46983b5\": container with ID starting with 1d151a2c4798cdf4c00e18f1fd6ec9e76400d5de8c6c07713b44c9afa46983b5 not found: ID does not exist" containerID="1d151a2c4798cdf4c00e18f1fd6ec9e76400d5de8c6c07713b44c9afa46983b5" Jan 21 12:18:29 crc kubenswrapper[4824]: I0121 12:18:29.569390 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1d151a2c4798cdf4c00e18f1fd6ec9e76400d5de8c6c07713b44c9afa46983b5"} err="failed to get container status \"1d151a2c4798cdf4c00e18f1fd6ec9e76400d5de8c6c07713b44c9afa46983b5\": rpc error: code = NotFound desc = could not find container \"1d151a2c4798cdf4c00e18f1fd6ec9e76400d5de8c6c07713b44c9afa46983b5\": container with ID starting with 1d151a2c4798cdf4c00e18f1fd6ec9e76400d5de8c6c07713b44c9afa46983b5 not found: ID does not exist" Jan 21 12:18:29 crc kubenswrapper[4824]: I0121 12:18:29.614042 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d7d81231-dda8-4cc9-b7a3-26e26fd4c3b1-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 21 12:18:29 crc kubenswrapper[4824]: I0121 12:18:29.614076 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7nmcb\" (UniqueName: \"kubernetes.io/projected/d7d81231-dda8-4cc9-b7a3-26e26fd4c3b1-kube-api-access-7nmcb\") on node \"crc\" DevicePath \"\"" Jan 21 12:18:29 crc kubenswrapper[4824]: I0121 12:18:29.780621 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-prbq4"] Jan 21 12:18:29 crc kubenswrapper[4824]: I0121 12:18:29.793542 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-prbq4"] Jan 21 12:18:30 crc kubenswrapper[4824]: I0121 12:18:30.066191 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d7d81231-dda8-4cc9-b7a3-26e26fd4c3b1" path="/var/lib/kubelet/pods/d7d81231-dda8-4cc9-b7a3-26e26fd4c3b1/volumes" Jan 21 12:18:46 crc kubenswrapper[4824]: I0121 12:18:46.065133 4824 patch_prober.go:28] interesting pod/machine-config-daemon-zqjg9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 21 12:18:46 crc kubenswrapper[4824]: I0121 12:18:46.066047 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 21 12:18:51 crc kubenswrapper[4824]: I0121 12:18:51.634651 4824 generic.go:334] "Generic (PLEG): container finished" podID="e1c80411-6f38-43b2-8656-639978632741" containerID="d0b75fbd713ed74fa81844f4c6ccb4b25946f8456765e9861e694df4f9f52b15" exitCode=0 Jan 21 12:18:51 crc kubenswrapper[4824]: I0121 12:18:51.634741 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-n5vql/must-gather-jkhmz" event={"ID":"e1c80411-6f38-43b2-8656-639978632741","Type":"ContainerDied","Data":"d0b75fbd713ed74fa81844f4c6ccb4b25946f8456765e9861e694df4f9f52b15"} Jan 21 12:18:51 crc kubenswrapper[4824]: I0121 12:18:51.635618 4824 scope.go:117] "RemoveContainer" containerID="d0b75fbd713ed74fa81844f4c6ccb4b25946f8456765e9861e694df4f9f52b15" Jan 21 12:18:51 crc kubenswrapper[4824]: I0121 12:18:51.694737 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-n5vql_must-gather-jkhmz_e1c80411-6f38-43b2-8656-639978632741/gather/0.log" Jan 21 12:19:00 crc kubenswrapper[4824]: I0121 12:19:00.325166 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-n5vql/must-gather-jkhmz"] Jan 21 12:19:00 crc kubenswrapper[4824]: I0121 12:19:00.326124 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-n5vql/must-gather-jkhmz" podUID="e1c80411-6f38-43b2-8656-639978632741" containerName="copy" containerID="cri-o://0937d1d47c67d983a6ee92637b82dba42aee39048f14d63afc09202819ad4750" gracePeriod=2 Jan 21 12:19:00 crc kubenswrapper[4824]: I0121 12:19:00.331113 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-n5vql/must-gather-jkhmz"] Jan 21 12:19:00 crc kubenswrapper[4824]: I0121 12:19:00.720514 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-n5vql_must-gather-jkhmz_e1c80411-6f38-43b2-8656-639978632741/copy/0.log" Jan 21 12:19:00 crc kubenswrapper[4824]: I0121 12:19:00.721105 4824 generic.go:334] "Generic (PLEG): container finished" podID="e1c80411-6f38-43b2-8656-639978632741" containerID="0937d1d47c67d983a6ee92637b82dba42aee39048f14d63afc09202819ad4750" exitCode=143 Jan 21 12:19:00 crc kubenswrapper[4824]: I0121 12:19:00.721157 4824 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="65ed37028631bc528541d931c3708c861369ab00cb45d0d01989f2e346501aa0" Jan 21 12:19:00 crc kubenswrapper[4824]: I0121 12:19:00.728518 4824 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-n5vql_must-gather-jkhmz_e1c80411-6f38-43b2-8656-639978632741/copy/0.log" Jan 21 12:19:00 crc kubenswrapper[4824]: I0121 12:19:00.728927 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-n5vql/must-gather-jkhmz" Jan 21 12:19:00 crc kubenswrapper[4824]: I0121 12:19:00.914473 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fdgjw\" (UniqueName: \"kubernetes.io/projected/e1c80411-6f38-43b2-8656-639978632741-kube-api-access-fdgjw\") pod \"e1c80411-6f38-43b2-8656-639978632741\" (UID: \"e1c80411-6f38-43b2-8656-639978632741\") " Jan 21 12:19:00 crc kubenswrapper[4824]: I0121 12:19:00.914833 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/e1c80411-6f38-43b2-8656-639978632741-must-gather-output\") pod \"e1c80411-6f38-43b2-8656-639978632741\" (UID: \"e1c80411-6f38-43b2-8656-639978632741\") " Jan 21 12:19:00 crc kubenswrapper[4824]: I0121 12:19:00.927122 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e1c80411-6f38-43b2-8656-639978632741-kube-api-access-fdgjw" (OuterVolumeSpecName: "kube-api-access-fdgjw") pod "e1c80411-6f38-43b2-8656-639978632741" (UID: "e1c80411-6f38-43b2-8656-639978632741"). InnerVolumeSpecName "kube-api-access-fdgjw". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 12:19:01 crc kubenswrapper[4824]: I0121 12:19:01.016525 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fdgjw\" (UniqueName: \"kubernetes.io/projected/e1c80411-6f38-43b2-8656-639978632741-kube-api-access-fdgjw\") on node \"crc\" DevicePath \"\"" Jan 21 12:19:01 crc kubenswrapper[4824]: I0121 12:19:01.096332 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e1c80411-6f38-43b2-8656-639978632741-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "e1c80411-6f38-43b2-8656-639978632741" (UID: "e1c80411-6f38-43b2-8656-639978632741"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 12:19:01 crc kubenswrapper[4824]: I0121 12:19:01.117645 4824 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/e1c80411-6f38-43b2-8656-639978632741-must-gather-output\") on node \"crc\" DevicePath \"\"" Jan 21 12:19:01 crc kubenswrapper[4824]: I0121 12:19:01.730155 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-n5vql/must-gather-jkhmz" Jan 21 12:19:02 crc kubenswrapper[4824]: I0121 12:19:02.064331 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e1c80411-6f38-43b2-8656-639978632741" path="/var/lib/kubelet/pods/e1c80411-6f38-43b2-8656-639978632741/volumes" Jan 21 12:19:02 crc kubenswrapper[4824]: I0121 12:19:02.237846 4824 scope.go:117] "RemoveContainer" containerID="0937d1d47c67d983a6ee92637b82dba42aee39048f14d63afc09202819ad4750" Jan 21 12:19:02 crc kubenswrapper[4824]: I0121 12:19:02.277373 4824 scope.go:117] "RemoveContainer" containerID="d0b75fbd713ed74fa81844f4c6ccb4b25946f8456765e9861e694df4f9f52b15" Jan 21 12:19:16 crc kubenswrapper[4824]: I0121 12:19:16.065249 4824 patch_prober.go:28] interesting pod/machine-config-daemon-zqjg9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 21 12:19:16 crc kubenswrapper[4824]: I0121 12:19:16.065881 4824 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 21 12:19:16 crc kubenswrapper[4824]: I0121 12:19:16.065931 4824 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" Jan 21 12:19:16 crc kubenswrapper[4824]: I0121 12:19:16.066396 4824 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d9d0949df611c52cc67603e5984be93524636cdbd8451c97adc3963dfdbe83d5"} pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 21 12:19:16 crc kubenswrapper[4824]: I0121 12:19:16.066450 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" podUID="33f3d922-4ffe-409b-a49a-d88c85898260" containerName="machine-config-daemon" containerID="cri-o://d9d0949df611c52cc67603e5984be93524636cdbd8451c97adc3963dfdbe83d5" gracePeriod=600 Jan 21 12:19:16 crc kubenswrapper[4824]: I0121 12:19:16.858202 4824 generic.go:334] "Generic (PLEG): container finished" podID="33f3d922-4ffe-409b-a49a-d88c85898260" containerID="d9d0949df611c52cc67603e5984be93524636cdbd8451c97adc3963dfdbe83d5" exitCode=0 Jan 21 12:19:16 crc kubenswrapper[4824]: I0121 12:19:16.858278 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" event={"ID":"33f3d922-4ffe-409b-a49a-d88c85898260","Type":"ContainerDied","Data":"d9d0949df611c52cc67603e5984be93524636cdbd8451c97adc3963dfdbe83d5"} Jan 21 12:19:16 crc kubenswrapper[4824]: I0121 12:19:16.859009 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zqjg9" event={"ID":"33f3d922-4ffe-409b-a49a-d88c85898260","Type":"ContainerStarted","Data":"111c22954de19114358173702fa304b6c5ccb7d35d262c1f76a3e710a19b138b"} Jan 21 12:19:16 crc kubenswrapper[4824]: I0121 12:19:16.859044 4824 scope.go:117] "RemoveContainer" containerID="719b0a497b4be25b481cf2d3456b6c515017599e2aeafd1f30d22bffc35f79d4" Jan 21 12:19:56 crc kubenswrapper[4824]: I0121 12:19:56.556867 4824 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-5449q"] Jan 21 12:19:56 crc kubenswrapper[4824]: E0121 12:19:56.557837 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7d81231-dda8-4cc9-b7a3-26e26fd4c3b1" containerName="extract-content" Jan 21 12:19:56 crc kubenswrapper[4824]: I0121 12:19:56.557851 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7d81231-dda8-4cc9-b7a3-26e26fd4c3b1" containerName="extract-content" Jan 21 12:19:56 crc kubenswrapper[4824]: E0121 12:19:56.557866 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7d81231-dda8-4cc9-b7a3-26e26fd4c3b1" containerName="registry-server" Jan 21 12:19:56 crc kubenswrapper[4824]: I0121 12:19:56.557871 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7d81231-dda8-4cc9-b7a3-26e26fd4c3b1" containerName="registry-server" Jan 21 12:19:56 crc kubenswrapper[4824]: E0121 12:19:56.557885 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1c80411-6f38-43b2-8656-639978632741" containerName="copy" Jan 21 12:19:56 crc kubenswrapper[4824]: I0121 12:19:56.557891 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1c80411-6f38-43b2-8656-639978632741" containerName="copy" Jan 21 12:19:56 crc kubenswrapper[4824]: E0121 12:19:56.557909 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7d81231-dda8-4cc9-b7a3-26e26fd4c3b1" containerName="extract-utilities" Jan 21 12:19:56 crc kubenswrapper[4824]: I0121 12:19:56.557915 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7d81231-dda8-4cc9-b7a3-26e26fd4c3b1" containerName="extract-utilities" Jan 21 12:19:56 crc kubenswrapper[4824]: E0121 12:19:56.557927 4824 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1c80411-6f38-43b2-8656-639978632741" containerName="gather" Jan 21 12:19:56 crc kubenswrapper[4824]: I0121 12:19:56.557932 4824 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1c80411-6f38-43b2-8656-639978632741" containerName="gather" Jan 21 12:19:56 crc kubenswrapper[4824]: I0121 12:19:56.558113 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="d7d81231-dda8-4cc9-b7a3-26e26fd4c3b1" containerName="registry-server" Jan 21 12:19:56 crc kubenswrapper[4824]: I0121 12:19:56.558130 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="e1c80411-6f38-43b2-8656-639978632741" containerName="copy" Jan 21 12:19:56 crc kubenswrapper[4824]: I0121 12:19:56.558149 4824 memory_manager.go:354] "RemoveStaleState removing state" podUID="e1c80411-6f38-43b2-8656-639978632741" containerName="gather" Jan 21 12:19:56 crc kubenswrapper[4824]: I0121 12:19:56.559438 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5449q" Jan 21 12:19:56 crc kubenswrapper[4824]: I0121 12:19:56.566773 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-5449q"] Jan 21 12:19:56 crc kubenswrapper[4824]: I0121 12:19:56.623837 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-976n5\" (UniqueName: \"kubernetes.io/projected/10d70708-c370-4167-be39-ef73bf918cce-kube-api-access-976n5\") pod \"redhat-marketplace-5449q\" (UID: \"10d70708-c370-4167-be39-ef73bf918cce\") " pod="openshift-marketplace/redhat-marketplace-5449q" Jan 21 12:19:56 crc kubenswrapper[4824]: I0121 12:19:56.624032 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/10d70708-c370-4167-be39-ef73bf918cce-utilities\") pod \"redhat-marketplace-5449q\" (UID: \"10d70708-c370-4167-be39-ef73bf918cce\") " pod="openshift-marketplace/redhat-marketplace-5449q" Jan 21 12:19:56 crc kubenswrapper[4824]: I0121 12:19:56.624207 4824 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/10d70708-c370-4167-be39-ef73bf918cce-catalog-content\") pod \"redhat-marketplace-5449q\" (UID: \"10d70708-c370-4167-be39-ef73bf918cce\") " pod="openshift-marketplace/redhat-marketplace-5449q" Jan 21 12:19:56 crc kubenswrapper[4824]: I0121 12:19:56.725786 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/10d70708-c370-4167-be39-ef73bf918cce-catalog-content\") pod \"redhat-marketplace-5449q\" (UID: \"10d70708-c370-4167-be39-ef73bf918cce\") " pod="openshift-marketplace/redhat-marketplace-5449q" Jan 21 12:19:56 crc kubenswrapper[4824]: I0121 12:19:56.726034 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-976n5\" (UniqueName: \"kubernetes.io/projected/10d70708-c370-4167-be39-ef73bf918cce-kube-api-access-976n5\") pod \"redhat-marketplace-5449q\" (UID: \"10d70708-c370-4167-be39-ef73bf918cce\") " pod="openshift-marketplace/redhat-marketplace-5449q" Jan 21 12:19:56 crc kubenswrapper[4824]: I0121 12:19:56.726117 4824 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/10d70708-c370-4167-be39-ef73bf918cce-utilities\") pod \"redhat-marketplace-5449q\" (UID: \"10d70708-c370-4167-be39-ef73bf918cce\") " pod="openshift-marketplace/redhat-marketplace-5449q" Jan 21 12:19:56 crc kubenswrapper[4824]: I0121 12:19:56.726392 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/10d70708-c370-4167-be39-ef73bf918cce-catalog-content\") pod \"redhat-marketplace-5449q\" (UID: \"10d70708-c370-4167-be39-ef73bf918cce\") " pod="openshift-marketplace/redhat-marketplace-5449q" Jan 21 12:19:56 crc kubenswrapper[4824]: I0121 12:19:56.726526 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/10d70708-c370-4167-be39-ef73bf918cce-utilities\") pod \"redhat-marketplace-5449q\" (UID: \"10d70708-c370-4167-be39-ef73bf918cce\") " pod="openshift-marketplace/redhat-marketplace-5449q" Jan 21 12:19:56 crc kubenswrapper[4824]: I0121 12:19:56.744061 4824 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-976n5\" (UniqueName: \"kubernetes.io/projected/10d70708-c370-4167-be39-ef73bf918cce-kube-api-access-976n5\") pod \"redhat-marketplace-5449q\" (UID: \"10d70708-c370-4167-be39-ef73bf918cce\") " pod="openshift-marketplace/redhat-marketplace-5449q" Jan 21 12:19:56 crc kubenswrapper[4824]: I0121 12:19:56.880477 4824 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5449q" Jan 21 12:19:57 crc kubenswrapper[4824]: I0121 12:19:57.340525 4824 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-5449q"] Jan 21 12:19:58 crc kubenswrapper[4824]: I0121 12:19:58.199779 4824 generic.go:334] "Generic (PLEG): container finished" podID="10d70708-c370-4167-be39-ef73bf918cce" containerID="254b8a867e9def0ba8da696a4037c52adb453700307bef8b7147a0ad56a1901d" exitCode=0 Jan 21 12:19:58 crc kubenswrapper[4824]: I0121 12:19:58.199826 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5449q" event={"ID":"10d70708-c370-4167-be39-ef73bf918cce","Type":"ContainerDied","Data":"254b8a867e9def0ba8da696a4037c52adb453700307bef8b7147a0ad56a1901d"} Jan 21 12:19:58 crc kubenswrapper[4824]: I0121 12:19:58.200197 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5449q" event={"ID":"10d70708-c370-4167-be39-ef73bf918cce","Type":"ContainerStarted","Data":"aa80bc3d83a5ca946ba8cf648ddeffd27293e1a25cfe3b133babcd1feb735afd"} Jan 21 12:19:58 crc kubenswrapper[4824]: I0121 12:19:58.202014 4824 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 21 12:19:59 crc kubenswrapper[4824]: I0121 12:19:59.211913 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5449q" event={"ID":"10d70708-c370-4167-be39-ef73bf918cce","Type":"ContainerStarted","Data":"9e2d51f0f4d6327d14565f1a4aa07332426a336e1bb1ce82b93bf99682269221"} Jan 21 12:20:00 crc kubenswrapper[4824]: I0121 12:20:00.223174 4824 generic.go:334] "Generic (PLEG): container finished" podID="10d70708-c370-4167-be39-ef73bf918cce" containerID="9e2d51f0f4d6327d14565f1a4aa07332426a336e1bb1ce82b93bf99682269221" exitCode=0 Jan 21 12:20:00 crc kubenswrapper[4824]: I0121 12:20:00.223219 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5449q" event={"ID":"10d70708-c370-4167-be39-ef73bf918cce","Type":"ContainerDied","Data":"9e2d51f0f4d6327d14565f1a4aa07332426a336e1bb1ce82b93bf99682269221"} Jan 21 12:20:01 crc kubenswrapper[4824]: I0121 12:20:01.234702 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5449q" event={"ID":"10d70708-c370-4167-be39-ef73bf918cce","Type":"ContainerStarted","Data":"8dc43b54a61b6be7c14210c38f108e23d9ae9d43d18b060b45efb69c4651279d"} Jan 21 12:20:01 crc kubenswrapper[4824]: I0121 12:20:01.263617 4824 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-5449q" podStartSLOduration=2.7529481369999997 podStartE2EDuration="5.263599853s" podCreationTimestamp="2026-01-21 12:19:56 +0000 UTC" firstStartedPulling="2026-01-21 12:19:58.201734575 +0000 UTC m=+4140.494763867" lastFinishedPulling="2026-01-21 12:20:00.71238629 +0000 UTC m=+4143.005415583" observedRunningTime="2026-01-21 12:20:01.258015533 +0000 UTC m=+4143.551044825" watchObservedRunningTime="2026-01-21 12:20:01.263599853 +0000 UTC m=+4143.556629145" Jan 21 12:20:06 crc kubenswrapper[4824]: I0121 12:20:06.882010 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-5449q" Jan 21 12:20:06 crc kubenswrapper[4824]: I0121 12:20:06.882760 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-5449q" Jan 21 12:20:06 crc kubenswrapper[4824]: I0121 12:20:06.919066 4824 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-5449q" Jan 21 12:20:07 crc kubenswrapper[4824]: I0121 12:20:07.331697 4824 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-5449q" Jan 21 12:20:07 crc kubenswrapper[4824]: I0121 12:20:07.373562 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-5449q"] Jan 21 12:20:09 crc kubenswrapper[4824]: I0121 12:20:09.311011 4824 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-5449q" podUID="10d70708-c370-4167-be39-ef73bf918cce" containerName="registry-server" containerID="cri-o://8dc43b54a61b6be7c14210c38f108e23d9ae9d43d18b060b45efb69c4651279d" gracePeriod=2 Jan 21 12:20:09 crc kubenswrapper[4824]: I0121 12:20:09.766013 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5449q" Jan 21 12:20:09 crc kubenswrapper[4824]: I0121 12:20:09.795368 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/10d70708-c370-4167-be39-ef73bf918cce-utilities\") pod \"10d70708-c370-4167-be39-ef73bf918cce\" (UID: \"10d70708-c370-4167-be39-ef73bf918cce\") " Jan 21 12:20:09 crc kubenswrapper[4824]: I0121 12:20:09.796161 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/10d70708-c370-4167-be39-ef73bf918cce-utilities" (OuterVolumeSpecName: "utilities") pod "10d70708-c370-4167-be39-ef73bf918cce" (UID: "10d70708-c370-4167-be39-ef73bf918cce"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 12:20:09 crc kubenswrapper[4824]: I0121 12:20:09.796620 4824 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/10d70708-c370-4167-be39-ef73bf918cce-utilities\") on node \"crc\" DevicePath \"\"" Jan 21 12:20:09 crc kubenswrapper[4824]: I0121 12:20:09.898935 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/10d70708-c370-4167-be39-ef73bf918cce-catalog-content\") pod \"10d70708-c370-4167-be39-ef73bf918cce\" (UID: \"10d70708-c370-4167-be39-ef73bf918cce\") " Jan 21 12:20:09 crc kubenswrapper[4824]: I0121 12:20:09.899033 4824 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-976n5\" (UniqueName: \"kubernetes.io/projected/10d70708-c370-4167-be39-ef73bf918cce-kube-api-access-976n5\") pod \"10d70708-c370-4167-be39-ef73bf918cce\" (UID: \"10d70708-c370-4167-be39-ef73bf918cce\") " Jan 21 12:20:09 crc kubenswrapper[4824]: I0121 12:20:09.906309 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/10d70708-c370-4167-be39-ef73bf918cce-kube-api-access-976n5" (OuterVolumeSpecName: "kube-api-access-976n5") pod "10d70708-c370-4167-be39-ef73bf918cce" (UID: "10d70708-c370-4167-be39-ef73bf918cce"). InnerVolumeSpecName "kube-api-access-976n5". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 21 12:20:09 crc kubenswrapper[4824]: I0121 12:20:09.921991 4824 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/10d70708-c370-4167-be39-ef73bf918cce-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "10d70708-c370-4167-be39-ef73bf918cce" (UID: "10d70708-c370-4167-be39-ef73bf918cce"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 21 12:20:10 crc kubenswrapper[4824]: I0121 12:20:10.001829 4824 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/10d70708-c370-4167-be39-ef73bf918cce-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 21 12:20:10 crc kubenswrapper[4824]: I0121 12:20:10.001877 4824 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-976n5\" (UniqueName: \"kubernetes.io/projected/10d70708-c370-4167-be39-ef73bf918cce-kube-api-access-976n5\") on node \"crc\" DevicePath \"\"" Jan 21 12:20:10 crc kubenswrapper[4824]: I0121 12:20:10.320205 4824 generic.go:334] "Generic (PLEG): container finished" podID="10d70708-c370-4167-be39-ef73bf918cce" containerID="8dc43b54a61b6be7c14210c38f108e23d9ae9d43d18b060b45efb69c4651279d" exitCode=0 Jan 21 12:20:10 crc kubenswrapper[4824]: I0121 12:20:10.320252 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5449q" event={"ID":"10d70708-c370-4167-be39-ef73bf918cce","Type":"ContainerDied","Data":"8dc43b54a61b6be7c14210c38f108e23d9ae9d43d18b060b45efb69c4651279d"} Jan 21 12:20:10 crc kubenswrapper[4824]: I0121 12:20:10.320259 4824 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5449q" Jan 21 12:20:10 crc kubenswrapper[4824]: I0121 12:20:10.320286 4824 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5449q" event={"ID":"10d70708-c370-4167-be39-ef73bf918cce","Type":"ContainerDied","Data":"aa80bc3d83a5ca946ba8cf648ddeffd27293e1a25cfe3b133babcd1feb735afd"} Jan 21 12:20:10 crc kubenswrapper[4824]: I0121 12:20:10.320308 4824 scope.go:117] "RemoveContainer" containerID="8dc43b54a61b6be7c14210c38f108e23d9ae9d43d18b060b45efb69c4651279d" Jan 21 12:20:10 crc kubenswrapper[4824]: I0121 12:20:10.338392 4824 scope.go:117] "RemoveContainer" containerID="9e2d51f0f4d6327d14565f1a4aa07332426a336e1bb1ce82b93bf99682269221" Jan 21 12:20:10 crc kubenswrapper[4824]: I0121 12:20:10.339429 4824 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-5449q"] Jan 21 12:20:10 crc kubenswrapper[4824]: I0121 12:20:10.344694 4824 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-5449q"] Jan 21 12:20:10 crc kubenswrapper[4824]: I0121 12:20:10.357437 4824 scope.go:117] "RemoveContainer" containerID="254b8a867e9def0ba8da696a4037c52adb453700307bef8b7147a0ad56a1901d" Jan 21 12:20:10 crc kubenswrapper[4824]: I0121 12:20:10.410465 4824 scope.go:117] "RemoveContainer" containerID="8dc43b54a61b6be7c14210c38f108e23d9ae9d43d18b060b45efb69c4651279d" Jan 21 12:20:10 crc kubenswrapper[4824]: E0121 12:20:10.410976 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8dc43b54a61b6be7c14210c38f108e23d9ae9d43d18b060b45efb69c4651279d\": container with ID starting with 8dc43b54a61b6be7c14210c38f108e23d9ae9d43d18b060b45efb69c4651279d not found: ID does not exist" containerID="8dc43b54a61b6be7c14210c38f108e23d9ae9d43d18b060b45efb69c4651279d" Jan 21 12:20:10 crc kubenswrapper[4824]: I0121 12:20:10.411032 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8dc43b54a61b6be7c14210c38f108e23d9ae9d43d18b060b45efb69c4651279d"} err="failed to get container status \"8dc43b54a61b6be7c14210c38f108e23d9ae9d43d18b060b45efb69c4651279d\": rpc error: code = NotFound desc = could not find container \"8dc43b54a61b6be7c14210c38f108e23d9ae9d43d18b060b45efb69c4651279d\": container with ID starting with 8dc43b54a61b6be7c14210c38f108e23d9ae9d43d18b060b45efb69c4651279d not found: ID does not exist" Jan 21 12:20:10 crc kubenswrapper[4824]: I0121 12:20:10.411065 4824 scope.go:117] "RemoveContainer" containerID="9e2d51f0f4d6327d14565f1a4aa07332426a336e1bb1ce82b93bf99682269221" Jan 21 12:20:10 crc kubenswrapper[4824]: E0121 12:20:10.411494 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9e2d51f0f4d6327d14565f1a4aa07332426a336e1bb1ce82b93bf99682269221\": container with ID starting with 9e2d51f0f4d6327d14565f1a4aa07332426a336e1bb1ce82b93bf99682269221 not found: ID does not exist" containerID="9e2d51f0f4d6327d14565f1a4aa07332426a336e1bb1ce82b93bf99682269221" Jan 21 12:20:10 crc kubenswrapper[4824]: I0121 12:20:10.411536 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9e2d51f0f4d6327d14565f1a4aa07332426a336e1bb1ce82b93bf99682269221"} err="failed to get container status \"9e2d51f0f4d6327d14565f1a4aa07332426a336e1bb1ce82b93bf99682269221\": rpc error: code = NotFound desc = could not find container \"9e2d51f0f4d6327d14565f1a4aa07332426a336e1bb1ce82b93bf99682269221\": container with ID starting with 9e2d51f0f4d6327d14565f1a4aa07332426a336e1bb1ce82b93bf99682269221 not found: ID does not exist" Jan 21 12:20:10 crc kubenswrapper[4824]: I0121 12:20:10.411578 4824 scope.go:117] "RemoveContainer" containerID="254b8a867e9def0ba8da696a4037c52adb453700307bef8b7147a0ad56a1901d" Jan 21 12:20:10 crc kubenswrapper[4824]: E0121 12:20:10.411850 4824 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"254b8a867e9def0ba8da696a4037c52adb453700307bef8b7147a0ad56a1901d\": container with ID starting with 254b8a867e9def0ba8da696a4037c52adb453700307bef8b7147a0ad56a1901d not found: ID does not exist" containerID="254b8a867e9def0ba8da696a4037c52adb453700307bef8b7147a0ad56a1901d" Jan 21 12:20:10 crc kubenswrapper[4824]: I0121 12:20:10.411884 4824 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"254b8a867e9def0ba8da696a4037c52adb453700307bef8b7147a0ad56a1901d"} err="failed to get container status \"254b8a867e9def0ba8da696a4037c52adb453700307bef8b7147a0ad56a1901d\": rpc error: code = NotFound desc = could not find container \"254b8a867e9def0ba8da696a4037c52adb453700307bef8b7147a0ad56a1901d\": container with ID starting with 254b8a867e9def0ba8da696a4037c52adb453700307bef8b7147a0ad56a1901d not found: ID does not exist" Jan 21 12:20:12 crc kubenswrapper[4824]: I0121 12:20:12.060949 4824 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="10d70708-c370-4167-be39-ef73bf918cce" path="/var/lib/kubelet/pods/10d70708-c370-4167-be39-ef73bf918cce/volumes"